id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1637593266
Add account name on home page and transaction view Currently, the account name is only visible on the account list. This task should create a component that gets the account's name from the account list, and if not found, it shows an unknown label. left is design, right is current PR implementation notice that: 1- "Account 1" font size is bigger in right. 2- "fuel...asd" font size is bigger in right 3- "ETH 0.00" font size is smaller in right 4- spacing to the bottom after "ETH 0.00" is smaller in right 5- eye icon is a bit smaller in right
gharchive/issue
2023-03-23T13:54:18
2025-04-01T04:32:33.991975
{ "authors": [ "LuizAsFight", "luizstacio" ], "repo": "FuelLabs/fuels-wallet", "url": "https://github.com/FuelLabs/fuels-wallet/issues/632", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1239340955
Parse all modules in a project before beginning the type-checking stage Currently, we parse and type-check each module in a sway project one at a time in a rough depth-first order. We should separate the concerns of parsing and and type-checking into separate stages both for code clarity and to unlock solutions to many issues related to inter-module definitions and the order in which we type-check items. From a quick scan, I believe this is a blocker (or at least a first-stage) for addressing the following: #201 #409 #870 #1267 #1557. @emilyaherbert's recent idea w.r.t. introducing a new "collection" stage prior to type-checking (no issue yet). Potential Solution The existing sway_parse::Program type more closely matches what we currently conceive of as a parsed "module" in Sway. We should rename what is currently referred to as program to module. I.e. sway_parse::program becomes sway_parse::module, sway_parse::Program becomes sway_parse::Module. Re-introduce a new sway_parse::Program type that is a small wrapper around a tree of sway_parse::Modules. Add a high-level function the sway_parse lib.rs called parse_project that begins by calling parse_file for the root module, then parses loads and parses all submodules based on the parent module's parsed dependencies. Refactor the type_check step in terms of the new sway_parse::Program type. More specifically, refactor the import_new_file function to import_new_dependency where rather than loading and parsing the file, we look up the pre-parsed module in the module tree. Marking as P: Critical as this is a blocker for #409 (critical) as well as a number of other high priority issues. I'm not sure that the logic for hunting down files for sub-modules is something that belongs in the parser. The parser is, in my mind at least, just something that converts a block of text to an AST (where that AST represents the structure of the original text as literally as possible). The old parser already had far too much logic built into it which is why the old AST still exists and we now have another AST layer on top of it. So I'm wary of heading in that direction again. I think it's important to keep the stages of the compilation pipeline as simple and well-defined as possible. I think the general idea of what you're proposing is right though, just that it belongs in sway-core. Consider, for instance, that if we ever decide to add rust-like macros then it won't be possible to know about sub-modules until after macro expansion. Thanks for the input @canndrew, that makes a lot of sense. Revising the "Proposed Solution" above with this in mind, I'm thinking we: Keep step 1. Change steps 2. and 3. to introduce the new parsed Program type and the parse_project function into sway-core rather than into sway-parse. Keep step 4. Yep :+1: Currently, we parse and type-check each module in a sway project one at a time in a rough depth-first order. We should separate the concerns of parsing and and type-checking into separate stages both for code clarity and to unlock solutions to many issues related to inter-module definitions and the order in which we type-check items. I 100% agree with this. But, at a high level, I think that this proposal is missing a key step will create a powerful mechanism in the compiler that we can use to solve a bunch of different issues. At a high-level I propose this design: step 1) the parsing phase. Parsing of all modules, separately. The goal of this step is to transform simple tokens to an internal AST representation. This could be the Expression type. step 2) the "collection" phase. This is a new phase that we would need to implement, and what it would do is this: It would gather high level information about the declarations inside of each module, without type checking the contents of these declarations. The result of this phase would be some sort of "collection context" that contains key information about struct definitions, enum definitions, function definitions, trait definitions, impl definitions and more. step 3) the type checking phase. During this phase, the "collection context" would be passed around during type checking, such that all declarations (enum/struct/etc) have the ability to have knowledge about every other declaration. Importantly, we would need to create a new data structure to allow us to do this without dummy function bodies. Similar to the type engine, I propose a "declaration engine" that allows declarations of the type Ref that refer to declarations that will be type checked out of order. The benefits of this design are this: step 2 introduces a powerful mechanism that will allow us to grow the compiler in the future, adding additional layers of inference, static analysis, and optimizations it allows us to have knowledge of all declarations, from the perspective of any one module. This would allow us to provide awesome benefits to users. For example, when using Rust in VSCode, VSCode will offer import suggestions for types as you are coding. For any one new type that you use, VSCode + Rust might have n different suggestions, all of which are modules that are yet to be imported to the module you are currently working on. There is even an auto-import feature that I personally use constantly. This change opens the doors for the compiler to support recursive functions, natively, reducing code size We could solve each of these issues inherently, without any temporary fixes or use of dummy functions: https://github.com/FuelLabs/sway/issues/65 https://github.com/FuelLabs/sway/issues/76 https://github.com/FuelLabs/sway/issues/201 https://github.com/FuelLabs/sway/issues/409 https://github.com/FuelLabs/sway/issues/738 https://github.com/FuelLabs/sway/issues/862 https://github.com/FuelLabs/sway/issues/870 https://github.com/FuelLabs/sway/issues/1267 https://github.com/FuelLabs/sway/issues/1527 https://github.com/FuelLabs/sway/issues/1548 https://github.com/FuelLabs/sway/issues/1555 https://github.com/FuelLabs/sway/issues/1557 To be clear @mitchmindtree is that the "recent idea w.r.t. introducing a new "collection" stage prior to type-checking" that I mentioned previously, would encapsulate the changes that you are hoping to see. By introducing the idea of a "collection" stage as a mechanism, we would be able to apply that mechanism to a variety of issues. @emilyaherbert I'm very excited about your proposal for a collection phase between parsing and type-checking! I think it makes a lot of sense, and will not only assist with monomorphizing, but also help to resolve the order in which declarations should get type-checked across modules more generally, and probably a bunch of other useful pre-type-check-preparation stuff as you mention :) Would you mind opening a dedicated issue for the new collection phase with all of these motivations and related issues? While I understand there's some overlap with this issue, my original intent for this issue was to be a small, actionable, easily reviewable, first-step toward these larger refactorings that we want to make. I still think it might be worth addressing the separation of parsing and type-checking as described in this issue first, and that doing so would make the addition of a new collection phase in between just a little easier/smaller. I am proposing that we perform a fundamental shift in how the compiler thinks of 1) ast node ordering, and 2) how the compiler performs its internal tasks. Right now, the ordering of the AST is critically important for how the compiler performs its internal tasks. In my opinion, this is an anti-pattern for a compiler. The order of evaluation of the AST nodes should not be a factor of importance inside the compiler. What I am proposing is that we achieve a fundamental shift in mindset, changing the compiler such that evaluation order of AST nodes has no effect. After this change, we could theoretically randomize the order of evaluation and the output should be identical. Are you suggesting we move the parsing of dep statements earlier in the compiler so we have access to all of them? Yes. This proposal suggests that we parse all the Sway files involved in a project first. Then perform the "collection" phase on all files, creating a "collection context." The "collection context" would be used to perform type checking, IR stuff, code gen, etc, such that it would make the concept of evaluating the AST in the "right order" obsolete. We currently use node_dependencies to order our "collection" for type checking. Are you suggesting that would become uneccessary? And if so, how? Yes and no. node_dependencies would be replaced with the "collection context". We already have this via the Root namespace. We could suggest auto-imports via that namespace currently. Sure, but this is a difficult task when the enum that you need to import, for example, is to be evaluated at a later point in the internal AST ordering. By removing the need to internally order the AST, this task becomes easier. I think there may be some issues here that are misrepresented, or that I'm understanding incorrectly. At a high-level, there are several issues that I believe to be intractable without this proposal. There are also issues that could be patched or avoided given the current system. But, they must all be addressed individually, each requiring their own level of involvement with creating a new mechanism within the compiler (additional checks, dummy functions, etc). In additional to tackling the intractable issues, this proposal seeks to eliminate a wide swatch of other issues at the fundamental level, without the need to address them individually. Additionally, I believe that "special-casing" particular fixes and check to be an anti-pattern, and this proposal eliminates much of the need for "special-casing." Let me go through the issues individually: https://github.com/FuelLabs/sway/issues/65 is related to codegen, not dependency resolution. We don't have the notion of a function stack. This proposal would allow us to perform type checking of recursive functions without the need for dummy functions, and would make the idea of a "function engine" (discussed on slack) significantly more simple. i.e. imagine that the "collection phase" collects: 1) function signatures for all functions, 2) information about all traits, 3) signatures for methods for traits, 4) information about which traits are implemented for which types, across all Sway files, 5) the signatures for all of those methods from all of those traits for all of those types. All before type checking. This is information that we fundamentally do no have in the compiler during type checking as it stands currently. https://github.com/FuelLabs/sway/issues/76 used to be implemented but had a bug so was reverted, and would still require a manual check. I'm not sure how this change would alter that. This proposal shifts this mindset. There is no need for a check for recursive dependencies. Recursive dependencies are fully supported, out of the box. https://github.com/FuelLabs/sway/issues/201 we would still need a manual check right? We can do this currently, no? There is no need for a manual check. Recursive use is supported natively. https://github.com/FuelLabs/sway/issues/409 This one would be solved for sure by moving the parsing of dep statements earlier so we could build a dependency map. I think that's the main value prop here. This proposal encompasses this change. The "collection context" could contain information about the dependency graph, making the concept of "explicit ordering" obsolete. https://github.com/FuelLabs/sway/issues/738 -- could you elaborate on how this would impact the tracking issue for future IR? Are you suggesting we would stop inlining functions as a part of this? If so, it should be noted that that's actually also due to codegen difficulties. Yes, we could stop inlining functions as a result of this proposal. As I mentioned above, this proposal would help us construct a "function engine" meaning we could stop inlining functions, without the need for introducing dummy functions. Also I will note that this is not just a codegen issue https://github.com/FuelLabs/sway/issues/1557. Before codegen even occurs, type checking is currently unequipped for non-inlining of functions. https://github.com/FuelLabs/sway/issues/862 I'm still not clear on what the original complaint is in this issue, so I can't comment. It is ambiguous to me what makes our current notion of unification/reference-based equality in the type system not "true". The current method of monomorphization creates one copy per call to each function or method, leading to wasted compute and unneeded code size. This PR is tracking changing this to creating one copy per type usage pattern, and will need to go in after we have stopped inlining functions. https://github.com/FuelLabs/sway/issues/870 is in node_dependencies and would still be an issue requiring a manual check. This proposal removes the need for a manual check. Out of order impl would be supported natively. https://github.com/FuelLabs/sway/issues/970 is just something that hasn't had type checking implemented for it yet, not sure how a parsing change like this would impact that. https://github.com/FuelLabs/sway/issues/1159 same as above https://github.com/FuelLabs/sway/issues/1162 same as above https://github.com/FuelLabs/sway/issues/1163 same as above. https://github.com/FuelLabs/sway/issues/1163 same as above. Fundamentally, this proposal is not just a parsing change. It is a complete shift in how the compiler thinks of dep/decl/impl/etc ordering, and eliminates the need for specific ordering at all. During implementation of this shift, parsing will be affected. Given this example: struct Foo<T> { x: T } impl<T> Foo<T> where T: Double + Triple { fn math(self) -> Foo<T> { let a = self.x.double(); let b = self.x.triple(); Foo { x: a + b } } } trait Double<T> { fn double(self) -> T; } trait Triple<T> { fn triple(self) -> T; } impl Double<u64> for u64 { fn double(self) -> u64 { self + self } } impl Triple<u64> for u64 { fn triple(self) -> u64 { self + self + self } } During type checking of the impl<T> Foo<T> where T: Double + Triple block, the compiler would have an understand of all the types in the entire project that implement Double and Triple, and would be able to use that information. The change might not be so apparent with this example. But imporantly, mutually recursive impl's would be inherently supported. Function from abi not recognized when called from an instantiated ContractCaller #1261 This is actually related to the special casing done around the constant evaluation for abi casts -- we need to know more information at compile time than with typical method applications, and there's a bug in that constant evaluation during the type checking phase. I don't know what specific additional information that you have in mind, but the "collection context" could help us gather that information. https://github.com/FuelLabs/sway/issues/1267 is related to adding a monomorphization step, not dep parsing, and is blocked on the implementation of type inference. This proposal would make the idea of a "function engine" and delayed monomorphization significantly easier. i.e. imagine that the "collection context" collects information about which types use each generic function. If this were the case, we could perform monomorphization separately, and could save compute by avoiding non-unique monomorphization. Rust does something similar: https://doc.rust-lang.org/nightly/nightly-rustc/rustc_monomorphize/collector/index.html https://github.com/FuelLabs/sway/issues/1298 is just a dead_code_analysis link that's missing. This definitely could be patched with the current system. However it would still be affected by this proposal. https://github.com/FuelLabs/sway/issues/1311 is strictly a type checking bug This proposal means that this check could happen before type checking, if we wanted to do that. Potentially saving compute. https://github.com/FuelLabs/sway/issues/1325 is a feature proposal and could be implemented with the way things are right now. Currently, type checking is conflated with dependency resolution, meaning that ripping out type inference will be a significant amount of work. This proposal lays the ground work for adding this feature in a more simple way. https://github.com/FuelLabs/sway/issues/1491 is an error message change and could be solved with Function from abi not recognized when called from an instantiated ContractCaller #1261 This proposal would make this more simple. https://github.com/FuelLabs/sway/issues/1527 once again inlining is not a collections/parsing limitation and was rather done for ease of codegen. We can remove the inline everything approach. Inlining functions is currently limited by type checking https://github.com/FuelLabs/sway/issues/1557. Type checking does not currently have any mechanism for handling non-inlined functions. We chatted about this on slack, but I believe the best solution to this is the concept of a "function engine", and the concept of a function engine is made significantly easier with this proposal. Without this proposal, the function engine would rely heavily on a mechanism that accounts for functions or methods that may or may not exist. But with this proposal, during type checking when the function engine is being populated and used, it would have the knowledge of the existence of all functions and methods in the project. https://github.com/FuelLabs/sway/issues/1548 is related to the fact that we don't name inner methods as top-level declarations and therefore they are not ordered. This would still require implementation with the new approach as far as I can see. Yes, but it would be made easier with this proposal. https://github.com/FuelLabs/sway/issues/1555 is related to the same inlining thing. This would be made easier with this proposal. https://github.com/FuelLabs/sway/issues/1557 is the inlining thing. Inlining would be significantly easier with this proposal. https://github.com/FuelLabs/sway/issues/1584 is pretty unrelated in general, it's missing check for a recursive enum decl. This proposal makes this issues solvable, without the need for special-casing. Currently, it would be impossible to preemptively know that this enum declaration contains a recursive element, without doing it inside type checking. I'm reading these issues and thinking that https://github.com/FuelLabs/sway/issues/1557 is the core issue that you're suggesting this collecting phase would solve, but actually that's a codegen/type checking thing and has no relation to the fact that modules are parsed in a project after type checking begins. Could you elaborate on how this could impact https://github.com/FuelLabs/sway/issues/1557? This is not the core issue that this proposal solves. This proposal introduces a "collection context" mechanism that has many applications. One application is that it would eliminate the concept of "dep/impl/decl ordering" from the compiler. Another application is that it would ease efforts towards function inlining and towards reducing non-unique monomorphization. Another application could be improving type inference. While I understand there's some overlap with this issue, my original intent for this issue was to be a small, actionable, easily reviewable, first-step toward these larger refactorings that we want to make. I still think it might be worth addressing the separation of parsing and type-checking as described in this issue first, and that doing so would make the addition of a new collection phase in between just a little easier/smaller. SGTM @mitchmindtree !
gharchive/issue
2022-05-18T02:08:00
2025-04-01T04:32:34.035248
{ "authors": [ "canndrew", "emilyaherbert", "mitchmindtree" ], "repo": "FuelLabs/sway", "url": "https://github.com/FuelLabs/sway/issues/1578", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1249881611
Use a context instead of a global in the type engine Issue creation is WIP. Ahh I missed this when creating #1810! Going to close this in favour of the new issue as it provides a little more context.
gharchive/issue
2022-05-26T18:03:34
2025-04-01T04:32:34.037943
{ "authors": [ "mitchmindtree", "mohammadfawaz" ], "repo": "FuelLabs/sway", "url": "https://github.com/FuelLabs/sway/issues/1691", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1559466602
Clarify how different program types are used The Sway book says the following: A Sway program itself has a type: it is either a contract, a predicate, a script, or a library. The first three of these things are all deployable to the blockchain. A library is simply a project designed for code reuse and is never directly deployed to the chain. It's a bit misleading to say that scripts and predicates are "deployable". In fact, only contracts are. We should clarify the language here. Closed by #72 on quickstart, issue was actually in the developer quickstart not the docs @ControlCplusControlV actually this issue refers to the first paragraph in https://fuellabs.github.io/sway/v0.34.0/book/sway-program-types/index.html.. But indeed there was another question on the forums about the section in that's being fixed in https://github.com/FuelLabs/fuel-docs/pull/72
gharchive/issue
2023-01-27T09:50:45
2025-04-01T04:32:34.040656
{ "authors": [ "ControlCplusControlV", "mohammadfawaz" ], "repo": "FuelLabs/sway", "url": "https://github.com/FuelLabs/sway/issues/3908", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1743541589
Storage is not initialized and will revert on read() Given the contract code: contract; storage { x: u64 = 64, } abi MyContract { #[storage(read)] fn get_value() -> u64; } impl MyContract for Contract { #[storage(read)] fn get_value() -> u64 { storage.x.read() } } storage.x.read() will revert because we call unwrap() on a None value. To recreate follow the steps from the fuel book: https://fuelbook.fuel.network/master/quickstart/smart-contract.html, and use the contract example above. I tried it again. Created a new project with forc init. Added tests with cargo generate --init fuellabs/sway templates/sway-test-rs --name {SOME_NAME}. Built with forc 0.40.0. Ran tests with cargo test ( used fuels-rs 0.42 and fuel-core 0.18.2) I still get the Revert(0) panic you have to call the method. Here is the whole test. use fuels::{prelude::*, types::ContractId}; // Load abi from json abigen!(Contract( name = "MyContract", abi = "out/debug/{SOME_NAME}-abi.json" )); async fn get_contract_instance() -> (MyContract<WalletUnlocked>, ContractId) { // Launch a local network and deploy the contract let mut wallets = launch_custom_provider_and_get_wallets( WalletsConfig::new( Some(1), /* Single wallet */ Some(1), /* Single coin (UTXO) */ Some(1_000_000_000), /* Amount per coin */ ), None, None, ) .await; let wallet = wallets.pop().unwrap(); let id = Contract::load_from( "./out/debug/predicate-true.bin", LoadConfiguration::default(), ) .unwrap() .deploy(&wallet, TxParameters::default()) .await .unwrap(); let instance = MyContract::new(id.clone(), wallet); (instance, id.into()) } #[tokio::test] async fn can_get_contract_id() { let (instance, _id) = get_contract_instance().await; instance.methods().get_value().call().await.unwrap(); } SOME_NAME - use your project name. FYI I also tried the test with forc 0.40.1 and I get the same revert.
gharchive/issue
2023-06-06T10:12:26
2025-04-01T04:32:34.045517
{ "authors": [ "hal3e" ], "repo": "FuelLabs/sway", "url": "https://github.com/FuelLabs/sway/issues/4634", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
855197880
The calculation result of Calc.calc is incorrect. Calc.calc does not calculate correctly if there is a space between values or operators. When space is entered, it is taken as each argument, so it was not handled correctly. The bug has been fixed.
gharchive/issue
2021-04-11T01:57:11
2025-04-01T04:32:34.047777
{ "authors": [ "Fukuda-B" ], "repo": "Fukuda-B/BBBot_discord", "url": "https://github.com/Fukuda-B/BBBot_discord/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1437302967
有写入的序列化,怎么没有读取的序列化呢? 感谢指出,ByteArray 和 Parcelable 确实是写漏了。已补全相关代码和示例,并提交 commit。将在 v1.1.2 版本中发版。 v1.1.3 已发版,包含此项修复。此 issue 将被关闭,如有其余问题欢迎继续交流
gharchive/issue
2022-11-06T06:28:31
2025-04-01T04:32:34.070920
{ "authors": [ "FunnySaltyFish", "guozhiqiang123" ], "repo": "FunnySaltyFish/ComposeDataSaver", "url": "https://github.com/FunnySaltyFish/ComposeDataSaver/issues/4", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
666711838
Arreglado el responsive , y seteadas las imagenes del carousel Otro cambio en el responsive Cancelamos el pull, para repararlo desde consola
gharchive/pull-request
2020-07-28T02:51:47
2025-04-01T04:32:34.072607
{ "authors": [ "Furok-Dev" ], "repo": "Furok-Dev/HouseTime", "url": "https://github.com/Furok-Dev/HouseTime/pull/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2097022551
🛑 Home Assistant is down In 7e02906, Home Assistant (https://ha.fussel.tv) was down: HTTP code: 400 Response time: 595 ms Resolved: Home Assistant is back up in 031930c after 53 minutes.
gharchive/issue
2024-01-23T21:33:16
2025-04-01T04:32:34.090295
{ "authors": [ "FusselTV" ], "repo": "FusselTV/status", "url": "https://github.com/FusselTV/status/issues/674", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
75451860
<link rel="prefetch" src="foo.htm" type="text/html" /> This issue was imported from Google Moderator Moderator votes: +14 Issue added by GedByrne on 2011-11-03 To vote this issue up or down, simply include +1 or -1 in your comment. Duplicate of #981 so moving half the votes there
gharchive/issue
2015-05-12T04:04:44
2025-04-01T04:32:34.116178
{ "authors": [ "Fyrd" ], "repo": "Fyrd/caniuse", "url": "https://github.com/Fyrd/caniuse/issues/1046", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
172018545
chromestatus.com/feature vs. chromestatus.com/features Should the Chrome status link be to the individual feature (focuses, loads much faster), or to features? Example: at http://caniuse.com/#feat=ambient-light, click "in development", vs. without the 's': http://www.chromestatus.com/feature/5298357018820608. Good idea, has now been changed to the /feature/ link. Thanks!
gharchive/issue
2016-08-18T22:46:55
2025-04-01T04:32:34.118731
{ "authors": [ "Fyrd", "dandv" ], "repo": "Fyrd/caniuse", "url": "https://github.com/Fyrd/caniuse/issues/2737", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
224001331
battery-status: Reflect removal from Firefox Fixes #3393 Thanks!
gharchive/pull-request
2017-04-25T02:24:37
2025-04-01T04:32:34.119507
{ "authors": [ "Fyrd", "cvrebert" ], "repo": "Fyrd/caniuse", "url": "https://github.com/Fyrd/caniuse/pull/3396", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1036420319
Add options Add description field for cdn resource struct webp optimization option browser cache settings option After acceptance of this PR next one will be created: https://github.com/G-Core/terraform-provider-gcorelabs/compare/master...skyeng:additions_from_skyeng?expand=1 @ujhgj thank you for PR, will be reviewed ASAP @ujhgj I close it as all options have been added. I'll add them to https://github.com/G-Core/terraform-provider-gcorelabs as well.
gharchive/pull-request
2021-10-26T15:06:50
2025-04-01T04:32:34.122152
{ "authors": [ "ujhgj", "vvelikodny" ], "repo": "G-Core/gcorelabscdn-go", "url": "https://github.com/G-Core/gcorelabscdn-go/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1123398202
New game: Stratego Adds the game Stratego to the framework (implemented by Jonny Betts). Some adjustments made (game runs), but it needs further testing on functionality (and potentially some more simplifications are possible). I've made some small changes; but on the whole this looks OK. Approved.
gharchive/pull-request
2022-02-03T18:04:17
2025-04-01T04:32:34.138302
{ "authors": [ "hopshackle", "rdgain" ], "repo": "GAIGResearch/TabletopGames", "url": "https://github.com/GAIGResearch/TabletopGames/pull/178", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1953852257
Include input uncertainty in output estimation At the moment the tool can estimate the uncertainty of the output given an input. But what happens when the input is also uncertain? In many practical cases the uncertainty in the input could be as large as the uncertainty of the output. Would it be possible to integrate input uncertainty to the tool? Dear manna88aero, Sorru for the delay. The models you refer to are called GP-LVM (Gaussian process latent variable model). In its current state, MOGPTK has not implemented GPLVMs, hopefully we can do it in the future. Best, Felipe. On 20-10-2023, at 06:02, manna88aero @.***> wrote: At the moment the tool can estimate the uncertainty of the output given an input. But what happens when the input is also uncertain? In many practical cases the uncertainty in the input could be as large as the uncertainty of the output. Would it be possible to integrate input uncertainty to the tool? — Reply to this email directly, view it on GitHub https://github.com/GAMES-UChile/mogptk/issues/67, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACT3KG47RJPU7INDACAZUHTYAI42XAVCNFSM6AAAAAA6IWI7ECVHI2DSMVQWIX3LMV43ASLTON2WKOZRHE2TGOBVGIZDKNY. You are receiving this because you are subscribed to this thread. On a side note, you can set Y_err when calling mogptk.Data(...) (see https://games-uchile.github.io/mogptk/data.html#mogptk.data.Data) to pass the standard deviation of the input data. When using the Exact model (the default), these will be passed as data_variance to https://games-uchile.github.io/mogptk/gpr/model.html#mogptk.gpr.model.Exact. Hopefully that satisfies your need.
gharchive/issue
2023-10-20T09:02:24
2025-04-01T04:32:34.143131
{ "authors": [ "felipe-tobar", "manna88aero", "tdewolff" ], "repo": "GAMES-UChile/mogptk", "url": "https://github.com/GAMES-UChile/mogptk/issues/67", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
616315383
Add a formatting script for Travis CI Please check if the PR fulfills these requirements: [x] The commit message follows our guidelines. For bug fixes and features: [ ] You tested the changes. [ ] You updated the docs or changelog. What kind of change does this PR introduce? This is a simple script which automatically checks and corrects file formatting to comply with POSIX standards. The POSIX standard for text files says that text files have lines delimited with newline characters and the file ends with a newline character. GitHub shows warnings in the diffs when files don't end with newline characters. To run on Travis, this repo needs to be added on the Travis CI website. The formatting script can also be run locally to fix formatting, which I did with the 2nd commit. Starting with Godot 3.2, Godot will automatically ensure files end in newlines with its internal editor, but this script can help with fixing up old files, and ensuring this is done with all text files. The only other change is that a BOM was removed from multiplayer-outline-part-2.md. P.S. I noticed there is also a checkerrors.sh script. If you want, I could try adding this to Travis in another PR, so it's run automatically. Does this PR introduce a breaking change? No. Thanks
gharchive/pull-request
2020-05-12T02:43:21
2025-04-01T04:32:34.164274
{ "authors": [ "NathanLovato", "aaronfranke" ], "repo": "GDQuest/godot-demos", "url": "https://github.com/GDQuest/godot-demos/pull/61", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2584121039
[Feature]: Add an About Us Page @rishicds Atleast assign me this issue to add an about Us. Assigned.
gharchive/issue
2024-10-13T17:11:44
2025-04-01T04:32:34.171166
{ "authors": [ "inkerton", "rishicds" ], "repo": "GDSC-RCCIIT/gdg-website", "url": "https://github.com/GDSC-RCCIIT/gdg-website/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1001658030
Docs - Added documentation to our project Issue Closes #17 What is changed / added added docusaurus to our project under src/documentation Screenshots fixing changed files
gharchive/pull-request
2021-09-21T01:45:04
2025-04-01T04:32:34.175520
{ "authors": [ "TheFatPanda97" ], "repo": "GDSCUTM-CommunityProjects/ActNow", "url": "https://github.com/GDSCUTM-CommunityProjects/ActNow/pull/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
450586866
Movement: Wall jump and slide Add the ability to jump against walls, and have the character slowly slide down slowly while against the wall With "ability" you mean a new Skill? If so, Skills are following a State pattern, right? Could you tell me if these snippets are on the lines for this task? WallSlide WallJump <- These add a vertical gravity to make the walls behave like a floor, making the character perform an arc similar to as if it was on the floor jumping and moving horizontally. Skill.gd is specific to the hook and likely to go out of the project. Check the Player scene, it uses a state machine now (look for the StateMachine.gd script and Player/**/State.gd) The character shouldn't be able to climb on a single wall indefinitely imo: the jump is weak in the game, the hook is your main tool. The wall jump's purpose is to help you change direction upon hitting a wall without losing momentum, e.g. to reach a hooking point that's just a little too high otherwise. Another use would be to climb between two walls that are close to one another. Regarding your snippets, I don't want things like dividing or multiplying the gravity depending on the context, as it makes the motion unreliable: change the world's gravity and the arc of your jumps won't be the same anymore in different states. I prefer the jump to always work the same way, even from a Wall. For now, you can keep it simple. The wall should mainly feel good and support the core mechanic: the hook. I would create a Wall state as a child of the Move state, define the wall motion there (e.g. sliding down), and use the Jump state to jump away from the wall. You can already use StateMachine.transition_to(state, msg={'velocity': ...} to make the player jump in the direction you want, using the velocity you want. Check the Move and Jump nodes in the Player scene.
gharchive/issue
2019-05-31T02:31:34
2025-04-01T04:32:34.180431
{ "authors": [ "NathanLovato", "henriiquecampos" ], "repo": "GDquest/godot-metroidvania-2d", "url": "https://github.com/GDquest/godot-metroidvania-2d/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
78464579
Add override for request state It would be nice to have a feature to alter the account request state. The case occurred where an account request was being handled simultaneously by two people and the request_state ended up as "EMAILED_LEADS" instead of "APPROVED". There may be other cases where a manual state change is desirable. Imported from trac ticket #60, created by ahelsing on 10-23-2013 at 17:56, last modified: 04-02-2014 at 15:47 If y'all do this via the web UI, it should log-and-audit the state change. Trac comment by chaos on 10-23-2013 at 18:02 Sorry, too much shorthand: audit == send mail to the audit address. Trac comment by chaos on 10-23-2013 at 18:03 Someone other than Pam should own this. Trac comment by jbs on 04-02-2014 at 15:47
gharchive/issue
2015-05-20T10:18:54
2025-04-01T04:32:34.183477
{ "authors": [ "ahelsing" ], "repo": "GENI-NSF/geni-ar", "url": "https://github.com/GENI-NSF/geni-ar/issues/60", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
468271409
Move install directory parallel to build/ Per discussion with @pchakraborty, move the install directory to parallel with build/. The reason is that in the build/ directory there is a bin/ directory which we felt could confuse people if (when) we said "Go to the bin directory" or the like. I'm going to close and kill this branch. We can always reopen a new one if needed
gharchive/pull-request
2019-07-15T18:27:49
2025-04-01T04:32:34.201696
{ "authors": [ "mathomp4" ], "repo": "GEOS-ESM/ESMA_cmake", "url": "https://github.com/GEOS-ESM/ESMA_cmake/pull/12", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1853324019
Feature/tclune/mapl3 latlongeomfactory Description Related Issue Motivation and Context How Has This Been Tested? Types of changes [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) [ ] Trivial change (affects only documentation or cleanup) Checklist: [ ] I have tested this change with a run of GEOSgcm (if non-trivial) [ ] I have added one of the required labels (0 diff, 0 diff trivial, 0 diff structural, non 0-diff) [ ] I have updated the CHANGELOG.md accordingly following the style of Keep a Changelog @tclune, I don't see a hook in the new GeomFactory for the equivalent of append_var_metadata that was in the old factory The one I would insist on being restored before I approve is append_variable_metadata. Why does the metadata object need to be 1st created somewhere else? The create metadata routine is the former "append" plus the assumption that it will be first in the pipeline. I can't immediately recall how I convinced myself of this, but will certainly amend the interface as the use cases come in. Trying to keep it simple for the moment - just the stuff I can immediately see how it fits in the process. The one I would insist on being restored before I approve is append_variable_metadata. Why does the metadata object need to be 1st created somewhere else? The create metadata routine is the former "append" plus the assumption that it will be first in the pipeline. I can't immediately recall how I convinced myself of this, but will certainly amend the interface as the use cases come in. Trying to keep it simple for the moment - just the stuff I can immediately see how it fits in the process. Please go look at what the old append_variable_metdata is doing, I don't think you are understanding what it is for. This is NOT for appending the overall file metadata. We add grid specific meta to the file variables, at least in the case of the cube, go look in the old factory. We add specific metadata to the variables so that panoply knows how to plot them. The append_variable_metadata was a way to know for a particular grid, if any extra metadata needs to be attached to the file variables we will be writing. Someone needs to know, if I'm a cube, every variable in the file needs these extra attributes. In reality every factory that is not a basic lat-lon should be adding "coordinate" attribute to the variables according to CF. So there really does need to be a place in general if I'm grid foo, there may need to be some extra attributes that need to be added to the variables. OK - I understand. But still, it can wait for now. It will be straightforward enough to bring in when the time comes. Yes - we have to implement it for each factory subclass, but it's the same work now as it will be then, and my short term need is to use the factory to produce distinct grids so that generic can induce a regrid operation.
gharchive/pull-request
2023-08-16T14:06:14
2025-04-01T04:32:34.209672
{ "authors": [ "bena-nasa", "tclune" ], "repo": "GEOS-ESM/MAPL", "url": "https://github.com/GEOS-ESM/MAPL/pull/2309", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
505470582
Add support for GEOSwift 5.1 When CocoaPods tries to install the library, the following message is displayed: [!] CocoaPods could not find compatible versions for pod "GEOSwift": In snapshot (Podfile.lock): GEOSwift (= 5.1.0) In Podfile: GEOSwift GEOSwiftMapKit was resolved to 1.0.0, which depends on GEOSwift (= 5.0.0) PS: The CocoaPods version is 1.8.3 Hi, can you attach a sample project that reproduces the issue? Also, I wonder whether you just need to $ pod repo update You are absolutely right. Thank you for your help.
gharchive/issue
2019-10-10T19:39:02
2025-04-01T04:32:34.212254
{ "authors": [ "dasilva-carlos", "macdrevx" ], "repo": "GEOSwift/GEOSwiftMapKit", "url": "https://github.com/GEOSwift/GEOSwiftMapKit/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
70450772
Developers-Information Developers-Information in der englischen Version ist auf Deutsch. In Version 0.9.8 auf Android 4.4.4 in der englischen Version.
gharchive/issue
2015-04-23T16:14:50
2025-04-01T04:32:34.229040
{ "authors": [ "NIPE-SYSTEMS" ], "repo": "GGDevelopers/SchulinfoAPP", "url": "https://github.com/GGDevelopers/SchulinfoAPP/issues/35", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
681546090
Are the "cutting plane" and "clipping plane" the same thing? The 3D docs reference a "cutting plane" and a "clipping plane". Are these the same thing? If so, probably one word should be chosen and made consistent. Yes they are, and yes we should fix that :) I personally have a slight preference for "cutting" than "clipping", but both can work. The term "clipping plane" comes from OpenGL, e.g. the function glClipPlane(...), the flag GL_CLIP_PLANE0, etc. In the context of GLVis though, "cutting plane" makes sense too. I'm fine with either choice, if you want to unify the terminology. @rw-anderson do you want to propose a simple PR for this? Addressed in https://github.com/GLVis/glvis/pull/123/commits/49bc42c7e12b1859bc7f14acb1207c1472671df0 as part of the v4.0-dev branch
gharchive/issue
2020-08-19T04:57:48
2025-04-01T04:32:34.265672
{ "authors": [ "rw-anderson", "tzanio", "v-dobrev" ], "repo": "GLVis/glvis", "url": "https://github.com/GLVis/glvis/issues/107", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2760751892
[BUG] Noticeable delay when opening/scrolling through menus Version & Platform Version [1.0.5.0] (this was happening in 1.0.3.4 as well) Singleplayer Describe the bug There is a noticeable delay/framerate drop when opening/scrolling through menus (like the map, settings page, PowerTools menu, etc.). This is a very minor "bug", but can be slightly jarring. The delay seems to be roughly half a second or less. This occurs both when only this mod is selected, and when in a full list of mods (~70 currently). Deselecting this mod removes the aforementioned delay. To Reproduce Steps to reproduce the behavior: Load into the game with mod active Open any menu (map, settings page, PowerTools menu) Expected behavior In theory, there should be no delay opening any menus and should be smooth going between them. Log files Log shows no errors or warnings related to this, as far as I can tell, but I'll include it anyway. log.txt Thanks for the report @BurcoGames, I'll take a look. I suspect it has to do with the UI menu that we've added, but I don't know if there's anything we can do about it. Yes, I confirm that too, it happens to me with the latest version (releases) 1.0.5.0. Without the mod installed, the menus are all smooth; however when I open the same in-game while with the mod installed, it lags quite a bit. Yes, I confirm that too, it happens to me with the latest version (releases) 1.0.5.0. Without the mod installed, the menus are all smooth; however when I open the same in-game while with the mod installed, it lags quite a bit. PS.: it's possible add inside a Refresh Contracts + Clear all contracts?? Please post suggestions or feature requests as their own issue, not appending to other people's completely unrelated issues. ...and No, that feature will not be part of Contract Boost. That feature will be handled by FS25_BetterContracts, which is intended to live side by side with this mod. No, because i don't use BetterContracts (only one at once, just for tried), so i use only ContractBoost.. @BurcoGames We've been testing this bug report (thank you for the detailed report), but at this time we're not able to replicate any noticeable delay when opening or going between menus. The only thing that this mod does is add a somewhat large number of options to the general settings menu, after which - it's up to the game engine to render those properly. I can only guess that the game engine isn't optimized properly for long menus within the settings menu area. At this time, there's nothing we can do that's actionable other than potentially disabling the addition of the menu entirely. Since all of the settings also live in a preferences file that is user editable, the only thing I can think of would be to add a configuration option (that would be true by default and not accessible in the UI) something like enableInGameSettingsMenu - and let you (or anyone else experiencing that same delay and wanting to do something about it) to disable the menu entirely. Other than that, it's not a delay that I have any control over. Thoughts? Would you like that setting, or no? @GMNGjoy I appreciate you looking into this. If nothing else, I think that config option could be useful not only to me but others as well, so yeah that would be nice to have, if it's easy to implement. I did have a thought, however - I know mods like Courseplay have an entirely separate menu for all the configurations, so would that be something that could solve this? Basically a dedicated menu with all the options and a keybind to access said menu. Just an idea, if that wouldn't work, then of course the solution you provided would work just fine. I did have a thought, however - I know mods like Courseplay have an entirely separate menu for all the configurations, so would that be something that could solve this? Basically a dedicated menu with all the options and a keybind to access said menu. Just an idea, if that wouldn't work, then of course the solution you provided would work just fine. This would be the only other "solution" which would take a significant amount of work, but may or may not solve the "issue"; it doesn't make sense to dive into that deep of a rewrite without knowing what the root cause is. I'll see about adding a config param, and we can take it from there. This would be the only other "solution" which would take a significant amount of work, but may or may not solve the "issue"; it doesn't make sense to dive into that deep of a rewrite without knowing what the root cause is. I'll see about adding a config param, and we can take it from there. That makes sense. Yeah, the config option is a good start, and we'll see what happens. Again, I really appreciate you looking into this. done. 1.0.5.3 Please let me know if the performance issue / delay is gone with this setting turned off. Just tested it, and I can confirm that the delay is fixed with the in-game settings disabled. Happy that worked! 👍 On Sat, Dec 28, 2024 at 5:12 PM BurcoGames @.***> wrote: Just tested it, and I can confirm that the delay is fixed with the in-game settings disabled. — Reply to this email directly, view it on GitHub https://github.com/GMNGjoy/FS25_ContractBoost/issues/45#issuecomment-2564575162, or unsubscribe https://github.com/notifications/unsubscribe-auth/AZSFL3QBKIJSIVXEBJ7DSTD2H5D6VAVCNFSM6AAAAABUIPN4U6VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKNRUGU3TKMJWGI . You are receiving this because you modified the open/close state.Message ID: @.***> I also tested the latest update and confirm the above reported: now with this setting disabled, every menu and switching from one section to another is much faster, smoother, returning as if there was no mod installed, plus I also gained several fps. Very Good work. 👍🏻
gharchive/issue
2024-12-27T11:12:48
2025-04-01T04:32:34.281039
{ "authors": [ "BurcoGames", "FirenzeIT", "GMNGjoy" ], "repo": "GMNGjoy/FS25_ContractBoost", "url": "https://github.com/GMNGjoy/FS25_ContractBoost/issues/45", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
565324160
DescriptorPool leaks when exceeding m_maxSetsPerPool DescriptorPool leaks descriptor sets when the number of descriptor sets exceeds the m_maxSetsPerPool value. The reason is simple : in FreeDescriptorSets we override m_currentAllocationPoolIndex by the freed descriptor set pool index, but we don't test if m_currentAllocationPoolIndex was already inferior to it which leads in loosing information of previously freed descriptor sets. There are two ways to fix this, but I think both must be used, just in case : The first is to use canonical destruction of descriptor sets by reversing order of destruction of transient resources in StreamEncoder. The second and safest is to add the above described test in DescriptorPool::FreeDescriptorSet : if (poolIndex < m_currentAllocationPoolIndex) m_currentAllocationPoolIndex = poolIndex; 🆙 @vlmillet it's been a few years, but could you explain a bit more about what you mean when it comes to reversing the order of destruction in StreamEncoder? Do you still have the code change?
gharchive/issue
2020-02-14T13:23:07
2025-04-01T04:32:34.332080
{ "authors": [ "EvilTrev", "vlmillet" ], "repo": "GPUOpen-LibrariesAndSDKs/V-EZ", "url": "https://github.com/GPUOpen-LibrariesAndSDKs/V-EZ/issues/72", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1221510403
VX_NN - vxUpsampleNearestLayer node implementation on HIP backend Implementation of vxUpsampleNearestLayer node on HIP backend which is needed for yolov3 models. $ ./anntest weights.bin OK: loaded 39 kernels from libvx_nn.so OK: OpenVX using GPU device#0 AMD Radeon VII (gfx906:sramecc+:xnack-) (with 60 CUs) on PCI bus 44:00.0 OK: graph initialization with annAddToGraph() took 10895.731 msec OK: vxProcessGraph() took 1619.843 msec (1st iteration) OK: vxProcessGraph() took 27.555 msec (average over 100 iterations) OK: HIP buffer usage: 531069860, 376/376 OK: successful This fixes #802 upsample issue. @hansely can you try OpenCL & HIP backend using dockers? @hansely can you try OpenCL & HIP backend using dockers? @kiritigowda On OCL backend: Runs successfully with batch size 4. On HIP backend: ./anntest weights.bin OK: loaded 39 kernels from libvx_nn.so "hipErrorNoBinaryForGpu: Unable to find code object for all current devices!" Aborted (core dumped) @hansely let me know when this is ready to be merged?
gharchive/pull-request
2022-04-29T20:49:44
2025-04-01T04:32:34.335155
{ "authors": [ "hansely", "kiritigowda" ], "repo": "GPUOpen-ProfessionalCompute-Libraries/MIVisionX", "url": "https://github.com/GPUOpen-ProfessionalCompute-Libraries/MIVisionX/pull/828", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
936611042
Gatsby-image is deprecated See #12
gharchive/issue
2021-07-05T02:03:18
2025-04-01T04:32:34.489704
{ "authors": [ "GabeEddyT", "GabyTinoco" ], "repo": "GabeEddyT/gatsby-image-background-slider", "url": "https://github.com/GabeEddyT/gatsby-image-background-slider/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2612569712
Criar modal home Descrição Precisamos de um modal com uma newsletter Não encontrei o arquivo da home. Alguém sabe onde está?
gharchive/issue
2024-10-24T21:13:41
2025-04-01T04:32:34.490844
{ "authors": [ "Gabriel-Alves02" ], "repo": "Gabriel-Alves02/ecommerce_organiza-o_x", "url": "https://github.com/Gabriel-Alves02/ecommerce_organiza-o_x/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
211571623
The Fold and FoldM docs don't link to the foldl package The turtle docs for these types aren't useful. The useful docs reside in the foldl package, but it was very difficult to figure that out. So the issue here is that if you re-export a type (like how Fold is re-exported in the Turtle module) then haddock prefers to link to the local package's re-export instead of the original package One thing I could do is add documentation pointing to the foldl haddocks so that users can figure out how to use Fold/FoldM more effectively, but I was wondering where the most discoverable place to do that would be Yet another argument against re-exports. :) In all seriousness, I'd put it in the haddock comment for Shell. Actually, looking at this more closely, the "Folds" section of the tutorial does link to the Control.Foldl module, so that seems like a reasonably discoverable place It's discoverable if you go through the tutorial. But not very discoverable for me browsing the haddocks. It's up to you though. Yeah, I think the tutorial is the appropriate place to reference the foldl library
gharchive/issue
2017-03-03T01:41:23
2025-04-01T04:32:34.494007
{ "authors": [ "Gabriel439", "mightybyte" ], "repo": "Gabriel439/Haskell-Turtle-Library", "url": "https://github.com/Gabriel439/Haskell-Turtle-Library/issues/220", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
785321974
work with output from long-running processes I am trying to wrap a process that watches for changes on a value, and so intentionally does not terminate; as it stands, it appears that I have no way to use the abstractions provided by turtle to do this. Here is an example that demonstrates the problem: *Main Lib Paths_etcdctl> import Turtle *Main Lib Paths_etcdctl Turtle> import qualified Control.Foldl as Fold *Main Lib Paths_etcdctl Turtle Fold> :t fold (inproc "yes" [] mempty) Fold.head fold (inproc "yes" [] mempty) Fold.head :: MonadIO io => io (Maybe Line) *Main Lib Paths_etcdctl Turtle Fold> fold (inproc "yes" [] mempty) Fold.head which blocks indefinitely. Now I know you are saying in the issue at https://github.com/Gabriel439/Haskell-Foldl-Library/issues/85 that this wont make sense for the foldl library, but I also cannot seem to make it work with the lower-level turtle functions, e.g. *Main Lib Paths_etcdctl Turtle Control.Foldl Fold> foldShell (inproc "yes" [] mempty) (FoldShell (\_ a-> pure $ Just a) Nothing pure) I saw this answer on stack overflow and it seemed to perhaps show a way: https://stackoverflow.com/questions/28421469/how-to-drop-lines-when-streaming-from-a-file-using-haskell-and-the-turtle-librar But testing it: *Main Lib Paths_etcdctl Turtle Control.Foldl Fold> view $ limit 10 (inproc "yes" [] mempty) Line "y" Line "y" Line "y" Line "y" Line "y" Line "y" Line "y" Line "y" Line "y" Line "y" ... which still blocks indefinitely, but no longer provides any output. I have seen a few issues that seem like they may perhaps be related (especially #22), but I believe this is separate. I have had to drop down to the process library in the past to do some things, but this just really surprised me, since turtle is built with streaming in mind. @joelmccracken: It is not possible to write a lazy version of limit. The Shell type is implemented in such a way that this cannot be done. The main reason why is due to the need to support safe exception handling, which doesn't play nice with terminating streams early. @joelmccracken: It is not possible to write a lazy version of limit. The Shell type is implemented in such a way that this cannot be done. The main reason why is due to the need to support safe exception handling, which doesn't play nice with terminating streams early. I see. I found this quite surprising given that Turtle attempts to support streaming in general with Shell; Can I add this to documentation somewhere? I can totally understand that there are things that Turtle doesn't attempt to handle, I just found this to be very surprising and it took me a while to track down what was going on. I see. I found this quite surprising given that Turtle attempts to support streaming in general with Shell; Can I add this to documentation somewhere? I can totally understand that there are things that Turtle doesn't attempt to handle, I just found this to be very surprising and it took me a while to track down what was going on.
gharchive/issue
2021-01-13T18:05:40
2025-04-01T04:32:34.500011
{ "authors": [ "Gabriel439", "joelmccracken" ], "repo": "Gabriel439/Haskell-Turtle-Library", "url": "https://github.com/Gabriel439/Haskell-Turtle-Library/issues/398", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1920968153
Add CODEOWNERS Description Adds the CODEOWNERS file. lgtm, this will make the auto reviewer suggestions better. But we should feel free to assign someone else if appropriate
gharchive/pull-request
2023-10-01T20:52:29
2025-04-01T04:32:34.535038
{ "authors": [ "RiscadoA", "luishfonseca" ], "repo": "GameDevTecnico/cubos", "url": "https://github.com/GameDevTecnico/cubos/pull/651", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2306529408
[Enhancement]: Addition of a Dedicated Play Button in Squared Game. Do you want to have the enhancement of existing game ? 😀 Describe yourself.. Game enhancement The current implementation of the Squared Game includes a "Tap to Play" text that changes color, prompting users to start the game. However, there is no specific button for playing. Adding a dedicated play button would enhance usability and improve the overall user experience by providing a clearer and more intuitive interface. Description: In the Squared Game, the "Tap to Play" text serves as the interactive element to start the game. While this text changes color to indicate interactivity, the absence of a specific play button can be confusing for users, especially those who might expect a more conventional interface element like a button. A dedicated play button would make the action more obvious, reduce ambiguity, and provide a more intuitive interaction point for users. Steps to Reproduce: Open the Squared Game interface. Observe the "Tap to Play" text that changes color when hovered over or tapped. Note the absence of a specific play button, which may lead to confusion about how to start the game. Expected Behavior: There should be a dedicated play button that clearly indicates the action to start the game. This button should have a recognizable design, such as a "Play" label or an icon, making it immediately apparent to users how to begin playing. Current Behavior: The game currently uses a "Tap to Play" text that changes color to indicate interactivity, but lacks a specific button for playing, which can be less intuitive and potentially confusing for some users. Describe the solution you'd like Solution steps Identify the Current Implementation. Design the Play Button Update the HTML Structure Style the Play Button with CSS Include styles for different states of the button, such as normal, hover, and active states, to provide visual feedback to users. Bind the Button with JavaScript Update or write JavaScript to link the play button to the game's start functionality. Ensure that clicking the play button triggers the existing game-start logic. Test the Implementation. Adjust Existing "Tap to Play" Text program in which contributing is Gssoc'24. this can be the steps .... please assign me this issue as a contributor i would love to decorate the game for good user experience. Select program in which you are contributing Other Code of Conduct [X] I follow CONTRIBUTING GUIDELINE of this project. Please raise any other issue, i'll assign you If you have any doubt you can directly send message on our discord (https://discord.gg/rZb46cCMmK), we have created separate channals for all projects... Thank you :)
gharchive/issue
2024-05-20T18:25:31
2025-04-01T04:32:34.544212
{ "authors": [ "Durgesh4993", "tanya-54" ], "repo": "GameSphere-MultiPlayer/Squard-line", "url": "https://github.com/GameSphere-MultiPlayer/Squard-line/issues/68", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1075059491
Add a cache for hash queries Currently, to query hashes on the blockchain, every block from genesis onwards is queried. This means the same block is queried multiple times if a query is run again. There is potential to save a lot of interaction time with the blockchain if a cache system is developed If left too long, a long enough blockchain could significantly slow down the query process Currently aiming to implement a store house that listens for newly committed blocks using https://iroha.readthedocs.io/en/develop/develop/api/queries.html#fetch-commits The storage for this information will be a python dictionary mapping domains to lists of hashes to start, but in future it may be better to change this to a key-value store, or even just files (in case the python object grows too large with large enough blockchains)
gharchive/issue
2021-12-09T02:08:28
2025-04-01T04:32:34.555519
{ "authors": [ "Gamma749" ], "repo": "Gamma749/IrohaFileHashing", "url": "https://github.com/Gamma749/IrohaFileHashing/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
598269966
Make sage calls compatible with virtual envs I'm using RsaCtfTool in a conda env with sage installed, but the script doesn't detect it. That's because of how the check is done: subprocess.check_output(['sage', '-v']) and all other subprocess' calls are not executed in the caller's activated conda env (same thing would happen with any other virtual env tool). Since sage can be called directly from Python, I propose not to use it with subprocess calls anymore but to directly use imports and function calls, that way the code would be cleaner and anyone would be able to use sage without needing to install it globally. If you agree with the idea, I could make a pull request. And thank you for providing us with this great tool :) Hi ! I've looked how to call sage directly from python, but it seem difficult with py3... Can you make an example on an attack (like ecm) so i'll try to add it everywhere it's needed? Closing for the moment.
gharchive/issue
2020-04-11T13:49:28
2025-04-01T04:32:34.557722
{ "authors": [ "Chadys", "Ganapati" ], "repo": "Ganapati/RsaCtfTool", "url": "https://github.com/Ganapati/RsaCtfTool/issues/95", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1360171029
OOM airflow-init_1 | WARNING!!!: Not enough memory available for Docker. airflow-init_1 | At least 4GB of memory required. You have 2.0G airflow-init_1 | airflow-init_1 | airflow-init_1 | WARNING!!!: You have not enough resources to run Airflow (see above)! Docker 기본용량 2G -> 4G 증설로 이슈 해결
gharchive/issue
2022-09-02T13:05:42
2025-04-01T04:32:34.560653
{ "authors": [ "Garden92" ], "repo": "Garden92/air-flow", "url": "https://github.com/Garden92/air-flow/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1418355157
[CONF-22] Add Prolive as silver sponsor Waiting for logo Logo for dark background ready
gharchive/issue
2022-10-21T13:36:19
2025-04-01T04:32:34.602678
{ "authors": [ "elfgoh" ], "repo": "GeekcampSG/geekcampsg.github.io", "url": "https://github.com/GeekcampSG/geekcampsg.github.io/issues/216", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
372894084
gfx-renderer macOS GLSL version 130 problems I'm trying to run imgui-gfx-renderer on this 10 year old macbook pro (I know right?). It does support all the way up to OpenGL 3.3, but for some reason on a 3.3 context, #version 130 of GLSL is not supported. #version 140, 150 and 330 all work, just not 130. Problem is that imgui-gfx-renderer only provides a 130 shader version for all of OpenGL 3+. So.. I can't run the gfx renderer (I can on a compatibility profile but the rest of this program relies on a core profile so..) I recompiled the crate with a version 150 shader and the program loaded up just fine. So would it be possible to provide a shaders version using a higher GLSL version? Pretty sure it'd fix the problem. Sure, this sounds like a doable idea. While supporting 10 year old stuff isn't top priority, in this case it's very simple and doesn't do any harm so let's do it :smile: I think I'll add version 150 shaders, because OpenGL 3.2 / GLSL 1.50 added the core/compatibility profile split, which is a major change and sounds like a good version to have explicit support for. Works like a charm!
gharchive/issue
2018-10-23T09:17:28
2025-04-01T04:32:34.615928
{ "authors": [ "Gekkio", "PJB3005" ], "repo": "Gekkio/imgui-rs", "url": "https://github.com/Gekkio/imgui-rs/issues/170", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
181189760
Include databinding folder in example the example references for example import com.genius.groupie.example.databinding.ItemSquareCardBinding; but there is no databinding folder… Bahh didn't realize it was something auto generated
gharchive/issue
2016-10-05T15:35:47
2025-04-01T04:32:34.659332
{ "authors": [ "maruf89" ], "repo": "Genius/groupie", "url": "https://github.com/Genius/groupie/issues/18", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1964036927
🛑 WWWW xyz is down In 7986490, WWWW xyz (https://www.crushing.xyz) was down: HTTP code: 502 Response time: 1474 ms Resolved: WWWW xyz is back up in b2bbebb after 8 minutes.
gharchive/issue
2023-10-26T17:34:35
2025-04-01T04:32:34.663134
{ "authors": [ "GentlemanHu" ], "repo": "GentlemanHu/own-status", "url": "https://github.com/GentlemanHu/own-status/issues/1466", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2129443169
🛑 AI flowise is down In c59dcb9, AI flowise (https://aiflow.crushing.xyz) was down: HTTP code: 0 Response time: 0 ms Resolved: AI flowise is back up in 27f011b after 7 minutes.
gharchive/issue
2024-02-12T05:25:14
2025-04-01T04:32:34.665451
{ "authors": [ "GentlemanHu" ], "repo": "GentlemanHu/own-status", "url": "https://github.com/GentlemanHu/own-status/issues/2730", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
404861316
Record screen produces video that appears black in Quicktime Here's the command I ran on version 1.6: scrcpy -m 1280 -r Desktop/some-file.mp4 -t Here are the logs for my recording attempt: /usr/local/Cellar/scrcpy/1.6/share/scr...shed. 1.2 MB/s (19464 bytes in 0.015s) 2019-01-30 17:30:33.950 scrcpy[21276:1317781] INFO: Enable show_touches 2019-01-30 17:30:34.673 scrcpy[21276:1317781] INFO: Initial texture: 720x1280 2019-01-30 17:30:41.662 scrcpy[21276:1317781] INFO: Disable show_touches Here's the output of scrcpy --version: scrcpy 1.6 dependencies: - SDL 2.0.9 - libavcodec 58.35.100 - libavformat 58.20.100 - libavutil 56.22.100 OS: MacOS Mojave 10.14 (18A391) While the command was running (until I clicked the close button on the window), the screen was displayed correctly on my computer, but the output file, results in black footage when viewed from Quicktime. Fixed on dev.
gharchive/issue
2019-01-30T16:44:09
2025-04-01T04:32:34.677006
{ "authors": [ "LouisCAD", "rom1v" ], "repo": "Genymobile/scrcpy", "url": "https://github.com/Genymobile/scrcpy/issues/416", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
304251944
Input Method Manager(IME) intercept pasted text IME intercept pasted text, which is undesirable. IME: Gboard with English and Chinese Languages. sample pasted text: scrcpy:用电脑控制 Android 手机,做演讲时很有用。比 vysor 的延迟和画面质量都好多了,开源,支持 Windows, Mac, Linux。目前Mac, Linux 端需要自行编译。 result: IME(in either English or Chinese mode) intercept scrcpy and discard text behind, no text pasted into TextView duplicate of #37
gharchive/issue
2018-03-12T06:15:10
2025-04-01T04:32:34.679289
{ "authors": [ "wuairc" ], "repo": "Genymobile/scrcpy", "url": "https://github.com/Genymobile/scrcpy/issues/48", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1044424551
GeoAI_2021_unit_01_EX_Warm_up_R-spatial 2 - check CRS and other info #-----------------------# raster::crs(rasterStack) raster::crs(buildings) ->I think here we need to put "orchards" instead of "buildings", right? crs(rasterStack) == crs(buildings) ->here as well raster::crs(orchards) -> also, here I get "NA" as result.. so this "crs(rasterStack) == crs(orchards)" also doesn't work Unit 01 - EX | Warm Up R-spatial Step 5 - Save the results for later usage 5 - stack and save as RDS #-----------------------# marburg_stack <- stack(rasterStack, rgbI) saveRDS(marburg_stack, (file.path(envrmt$data_processed, "dop_indices.rds")) The closing bracket is missing. And there is the following error: Error in if (file == "") stop("'file' must be non-empty string") : argument is of length zero Should we include the exercises from the chapters from Lovelace to our PDF file? If you like you can do so. I think it is a good training for producing Rmarkdown based documentation, even if it is not urgently necessary. I tried to get the tif File, but I got following message: Response [http://85.214.102.111/geo_data/data/01_raw_data/aerial/marburg_dop.tif] Date: 2021-11-25 20:27 Status: 401 Content-Type: text/html; charset=iso-8859-1 Size: 461 B 401 Unauthorized Unauthorized This server could not verify that you are authorized to access the document requested. Either you supplied the wrong credentials (e.g., bad password), or your browser doesn't understand how to supply How can I fix this problem? @Muenchj4 like always I need the commands u have used... most obously you did not provide the correct credentials. Try to download the data manually and it is a good idea to use an incognito window for haveing a wiped cache. @gisma I now have used the correct credentials and I got a response, but obviously it did not write into the folder. So I got the message: "Error in h(simpleError(msg, call)) : Fehler bei der Auswertung des Argumentes 'x' bei der Methodenauswahl für Funktion 'stack': object 'marburg_dop.tif' not found" @gisma I have used these commands: require(envimaR) MANDANTORY: defining the root folder DO NOT change this line rootDIR = "C:/Users/jomue/edu/geoAI" #-- Further customization of the setup by the user this section #-- can be freely customized only the definition of additional packages #-- and directory paths MUST be done using the two variables #-- appendpackagesToLoad and appendProjectDirList #-- feel free to remove this lines if you do not need them define additional packages uncomment if necessary appendpackagesToLoad = c("httr") define additional subfolders uncomment if necessary appendProjectDirList = c("data/dymmy-folder/") MANDANTORY: calling the setup script also DO NOT change this line source(file.path(envimaR::alternativeEnvi(root_folder = rootDIR),"src/geoAI_setup.R"),echo = TRUE) 1 - start script #----------------------------- geoai_user= "geoai" pw="ck@jx|xc?m2w" httr::GET("http://85.214.102.111/geo_data/data/01_raw_data/aerial/marburg_dop.tif", authenticate(geoai_user,pw), write_disk = (destfile=file.path(envrmt$path_data_data,"marburg_dop.tif", sep="/"))) Then I tryed to get a rater stack with rasterStack = raster::stack(file.path(envrmt$data/marburg_dop.tif)), but he could not find the object. It is not in the folder, obviously. Can you imagine, why? @gisma this problem is solved now. But there is a new problem: Trying to save a marburg_stack failed, since the rasterStack and the RGBI have got different extent. You have to crop or resample it. Good night I´l try it. Good night. Zitat von Chris Reudenbach @.***>: You have to crop or resample it. Good night -- You are receiving this because you were mentioned. Reply to this email directly or view it on GitHub: https://github.com/GeoMOER/geoAI/issues/1#issuecomment-979591629 @gisma @Baldl @dirkzeuss I tried to login at copernicus. But it failed. I tried to reset my password, but it failed again. Does anybody know what the problem could be? I am really sure to have written the password correct. @gisma Besides I tryed to a sentinel retrieval object and got following Error: [2021-11-27 21:12:35] #### Starting sen2r execution. #### Error in all(is.na(pm$extent)) || length(nn(pm$extent)) == 0 : invalid 'x' type in 'x || y' @Muenchj4 regarding Copernicus - no idea maybe check here regarding sentinel retrieval I think the error indicates a wron or missing extent probably due to referenced file you use.
gharchive/issue
2021-11-04T07:13:25
2025-04-01T04:32:34.709164
{ "authors": [ "Baldl", "Muenchj4", "gisma", "katharinakiem" ], "repo": "GeoMOER/geoAI", "url": "https://github.com/GeoMOER/geoAI/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
216017900
Doesn't renew with extender It doesn't renew using ext3nder tweak while normal yalu does I never heard of that tweak. It's not a tweak it's an app made by saurik The tweak it's just a way to use it without devlepoer account That is Extender not ext3nder. And Yalu Dark works fine with Cydia Exntender. Yes I know ext3nder is a tweak that makes extender work without developer id And yes Yalu dark works with ext3nder for me but it doesn't open when I'm in non jailbroken state You have Immortal Installed? No I confirm having ext3nder getting installed correctly the Yalu Dark It installs but I can open it after rebooting Can't * Trust the profile? Closed as the user hasn't replied in over 12 days.
gharchive/issue
2017-03-22T10:37:57
2025-04-01T04:32:34.741645
{ "authors": [ "GeoSn0w", "portalgamesmais", "ridalarry" ], "repo": "GeoSn0w/Yalu-Jailbreak-iOS-10.2", "url": "https://github.com/GeoSn0w/Yalu-Jailbreak-iOS-10.2/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1834939768
CI use all action fix #27 I found out that he will use the default action with command gepetuto -vv should we replace the snippet code by something else (not using pinocchio) or install pinocchio in the CI ? I think we should replace the snippet by something else. Simple math, event without numpy, would do https://nbconvert.readthedocs.io/en/latest/install.html#installing-nbconvert ;) should we re-add the 3 pipx install ruff,black and isort or maybe create a requirement.txt file ? No, they are already in pyproject.toml. But I had forgotten they were optional. I'll fix this, sorry.
gharchive/pull-request
2023-08-03T12:12:03
2025-04-01T04:32:34.748460
{ "authors": [ "TheoMF", "nim65s" ], "repo": "Gepetto/gepetuto", "url": "https://github.com/Gepetto/gepetuto/pull/28", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
1071225848
1.18 port (Needs a new branch) Ported to 1.18 (forge) Fix mod description containing control characters Fix block models not working Fix block entity saving Reopen on: https://github.com/German-Immersive-Railroading-Community/GIRC-Redstone/tree/1.18-master Weird that github disabled the feature to change the target branch of a pull request
gharchive/pull-request
2021-12-04T14:57:27
2025-04-01T04:32:34.765426
{ "authors": [ "HyCraftHD", "MrTroble" ], "repo": "German-Immersive-Railroading-Community/GIRC-Redstone", "url": "https://github.com/German-Immersive-Railroading-Community/GIRC-Redstone/pull/18", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1733921184
Add a source choices field to Task. Values: CLIMATE_ACTION_PLAN, SUGGESTED Rationale: Tasks may be officially decided (e.g. part of a climate action plan (KAP)) or mere suggestions of the LocalZero local team and therefore completely ignored by the government or something in between. As a first step we want to allow local teams to add to their (possibly insufficient) KAP. The added tasks should be discernible from the official ones. [x] Add class TaskSource as an IntegerChoices with valuesCLIMATE_ACTION_PLAN = 0, "KAP" SUGGESTED = 1, "Vorschlag" [x] Set default CLIMATE_ACTION_PLAN, also in migration for existing Tasks. [x] Come up with a way to show on task card and task page. (e.g. show title of SUGGESTED in cursive.) [ ] Come up with a way for users to understand the difference. (A legend would add clutter to the pages. Maybe a hover pop-up?) [ ] Allow to filter to show only one of the two kinds. (Optional, to be discussed, split into new issue) [ ] Add additional test cases Should suggested tasks be viewed exactly like normal tasks when it comes to aggregating the amount of tasks and their statuses in the city view? @La-Cezanne @mdrie three tasks in the description remain unchecked, is this realy closed?
gharchive/issue
2023-05-31T10:45:41
2025-04-01T04:32:34.769013
{ "authors": [ "La-Cezanne", "fblampe", "mdrie" ], "repo": "GermanZero-de/klimaschutzmonitor", "url": "https://github.com/GermanZero-de/klimaschutzmonitor/issues/196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1836955865
258 fix layout closes #258 @kleins05 please check/merge Great! Looks all good now. Seems like you understand the margins issue, at least your fixed it. And the overlap of the charts is gone indeed, cannot reproduce either. Thanks for fixing this so quickly and completely!
gharchive/pull-request
2023-08-04T15:27:27
2025-04-01T04:32:34.770514
{ "authors": [ "Holger-GZ", "kleins05" ], "repo": "GermanZero-de/klimaschutzmonitor", "url": "https://github.com/GermanZero-de/klimaschutzmonitor/pull/260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126405122
Adding the package results in an error. No compatible binary build found for this package Hey! Really excited that stream now has a meteor package as well. I can't add the package though. Whenever I try to I get this message: While checking for getstream:bin-deps@0.1.1: error: No compatible binary build found for this package. Contact the package author and ask them to publish it for your platform. Hope someone can help out :) Hi @novamul what platform (linux?) are you trying to install stream-meteor on? HI. I'm on Windows. @novamul ok i investigated and it seems the binary dependencies aren't build correctly on the Windows platform. I will need some more time to find out if there is a solution to this problem. Ok, too bad. Thanks for your quick responses and effort. I hope it can be fixed. Hi @matthisk. This is currently a blocking issue for us. Is there anything we can do to help? I searched around but cannot find the source code of getstream:bin-deps. @lewtds try running meteor publish-for-arch getstream:bin-deps@0.1.1 on a Windows machine, I have tried this but it results in an error message. Unfortunately I haven't gotten around to debugging this error message. The gestream:bin-deps project is currently not on github I will try to create a repo for it so we can fix this issue. Looks like node-fibers failed to build because most Windows machines don't come with VC++ and Python (and somehow the prebuilt binary is not used). Do we really need that package? I thought Meteor is already running on fibers? We managed to remove the bin-deps dependency and make it work by just Npm.require-ing fibers/future and getstream directly from this package (Meteor comes bundled with fibers already). Will try to upstream the changes ASAP. That wouldn't change the issue though would it? You still need to build the 'fibers' npm package for all target machines (i.e. linux/osx/windows). You can not use the fibers bundled with Meteor as far as I know? If you want to use fibers/futures you have to retrieve them from npm. I separated the npm dependencies into a different meteor package to avoid having to rebuild for every target architecture when I update stream-meteor. Hi @matthisk I just bumped into this problem when I was porting an app to Meteor Galaxy hosting ( running Ubuntu 14.04). Don't suppose you've had the chance to look more into it? I'm getting a slightly different error when I build: => Errors while initializing project: While checking for getstream:stream-meteor@0.4.2: error: No compatible binary build found for this package. Contact the package author and ask them to publish it for your platform. Thanks! Hi @kantle and @matthisk - I've encountered this one myself. Take a look at this: http://docs.meteor.com/#/full/meteorpublishforarch Meteor packages must be published for each platform you want to run them on: "Currently, the supported architectures for Meteor are 32-bit Linux, 64-bit Linux and Mac OS. Galaxy's servers run 64-bit Linux." Hi @ErikAugust - that's interesting. I've been running the app successfully on Digital Ocean, which is also running Ubuntu 14.04 x64. Shouldn't the package also fail on DO if it didn't support 64 bit? Hi guys maybe I did something wrong on the latest release (you have to do "publish for arch" for every release), I will make sure it is published for all architectures. Btw Windows built is still broken (@erik are you on Windows or OS X?), I can not get the package to build on windows and the Meteor support is also of no assistence Thanks for taking a look @matthisk - just for more info I'm on OSX. I was able to deploy to Digital Ocean using MUP without a problem. I tried removing the GetStream package and was able to deploy to Galaxy as well - so there is something about the deploy process to Galaxy that is not interacting well with this version of the package. @matthisk - Thanks, I'm on OS X like @kantle. I have forked this library and have it installed locally - and will be taking a look at a solution this weekend. Thanks again! @matthisk @ErikAugust I downgraded this package to version 0.3.7 and was able to build and deploy just fine to Galaxy, so I think it has something to do with how version 0.4.0 was published. @kantle @ErikAugust I just build the package for linxu (x86 and x64) so everything should work as expected. Please let me know if you are still experiencing any issues. @matthisk Awesome. Thanks! Awesome - just upgraded back to 0.4.0 and deployed no problem. Thanks @matthisk! Hey Guys! Read through the above thread. Did a Windows fix ever get implemented? Hitting the "No compatible binary build found for this package." @matthisk do we have an update on this? No currently we do not have a solution for this. Support from Meteor publishforarc on Windows doesn't seem to be in great shape. I am unable to use their architecture to publish the package. I tried running the build on a VM windows machine without success. I will give their build farm a try again right now, maybe they have applied some fixes to their architecture. @matthisk Appreciate the update and your time! getting the same problem. any more update? Any updates? we are still not able to run it on windows Any updates on this? is there any similar package for meteor?
gharchive/issue
2016-01-13T12:08:25
2025-04-01T04:32:34.851225
{ "authors": [ "ErikAugust", "Nisthar", "enoziak", "kantle", "lewtds", "matthisk", "monstrfolk", "novamul", "sayed-ali", "tbarbugli" ], "repo": "GetStream/stream-meteor", "url": "https://github.com/GetStream/stream-meteor/issues/1", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1669935053
Create LICENSE Ghepes Create LICENSE Exemple - ok!
gharchive/pull-request
2023-04-16T12:17:48
2025-04-01T04:32:34.871299
{ "authors": [ "Ghepes", "NextWromo" ], "repo": "Ghepes/Ghepes", "url": "https://github.com/Ghepes/Ghepes/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1494956558
Add enable toggle Currently, the vtt-observe is always enabled upon installation. That means when the user loads into a Roll20 campaign, vtt-observe will always remove their UI elements. The current workaround is to disable the extension as a whole from chrome://extensions. This is not preferable. We should add a toggle. Implemented in f999d38fda45e93f69336593d63d6ea5c505484a.
gharchive/issue
2022-12-13T18:50:26
2025-04-01T04:32:34.872579
{ "authors": [ "Ghifari160" ], "repo": "Ghifari160/vtt-observe", "url": "https://github.com/Ghifari160/vtt-observe/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1920434378
Add Log4Net Goal: Replace current Log with Log4Net library to log information More info see https://logging.apache.org/log4net/release/manual/configuration.html @josesimoes what do you think about? The "formated" logging is there for debugging purpose, basically to cross check with native output. Why do you think we need a different logging mechanism?
gharchive/issue
2023-09-30T23:44:16
2025-04-01T04:32:34.874416
{ "authors": [ "Ghislain1", "josesimoes" ], "repo": "Ghislain1/nf-tools", "url": "https://github.com/Ghislain1/nf-tools/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
806116625
Rotate doesn't behave properly with undo For some strange reason, rotating a big structure some angle that is not multiple of 90 degrees and the undoing it leaves blocks behind. I didn't do any code analysis to check why that's the case. Result of rotating an octahedron (into a space that was formerly just air) After undoing the operation I did a tiny bit of code analysis and I think I got the issue pinned down (from past experiece, when I did my undo in other apps). rotate is likely assigning to one block the rotated position of multiple blocks, meaning it sets that block multiple times. That results in the undo history having a block saved more than once. The first time it's saved, the old block it references is the actual old block, whatever was there before the operation. The second time it saves that position, what's saved as old block is what the previous setting operation did, hence not the actual old block (glass, in this case, instead of air). I can think of two ways to fix this: when undoing, run the list of blocks to set (affected_blocks) in reverse order. That ensures that the first block to be placed is the last to be removed, fixing the issue. when saving to history, don't allow to save one position more than once. This would require modifying set_block to do the check each time it sets a block. This solution would make undo more efficient, but all the other set operations less efficient, so I'm leaning for solution 1.
gharchive/issue
2021-02-11T06:28:35
2025-04-01T04:32:34.877726
{ "authors": [ "Firigion" ], "repo": "Ghoulboy78/World-edit-scarpet", "url": "https://github.com/Ghoulboy78/World-edit-scarpet/issues/57", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
1585198882
🛑 Licence Manager - Login API is down In e77718f, Licence Manager - Login API (https://licence.gidsimulation.com/v1/Auth/Login?fingerprint=gidsimulation&machine_name=status) was down: HTTP code: 404 Response time: 1894 ms Resolved: Licence Manager - Login API is back up in 2d9a7ce.
gharchive/issue
2023-02-15T04:28:37
2025-04-01T04:32:34.880424
{ "authors": [ "jginternational" ], "repo": "GiDHome/status-licence-manager", "url": "https://github.com/GiDHome/status-licence-manager/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
339206987
Issue with ORDER BY (SELECT 1) This cannot be interpreted by the VistaDB engine, for workaround, see: https://github.com/aspnet/EntityFrameworkCore/issues/12519 We may be able to easily extend the VistaDB engine to support SELECT 1 as syntax in this position; it not being supported is largely a lack of imagination that this was valid and useful. Reading through the EF ticket I'm surprised that ORDER BY is treated as "order by "; that's a SQL trick I didn't know. I'll look for the official MS TSQL Documentation on this point so we can refer that to the VistaDB team. OK, it looks like effectively (SELECT 1) is meant to largely obviate the order by expression but still allow offset/fetch (cheeky!) and get around their parser order. We'll see if we can get this put into the VistaDB engine as effectively "don't order by".
gharchive/issue
2018-07-08T08:15:43
2025-04-01T04:32:34.888619
{ "authors": [ "ErikEJ", "kendallmiller" ], "repo": "GibraltarSoftware/VistaDB.EFCore", "url": "https://github.com/GibraltarSoftware/VistaDB.EFCore/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2015604258
🛑 Gigadrive Accounts is down In 99cf3cb, Gigadrive Accounts (https://old.gigadrivegroup.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Gigadrive Accounts is back up in f4c014d after 30 minutes.
gharchive/issue
2023-11-29T01:04:48
2025-04-01T04:32:34.901582
{ "authors": [ "GigadriveBot" ], "repo": "Gigadrive/status.gigadrive.network", "url": "https://github.com/Gigadrive/status.gigadrive.network/issues/973", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
231069842
Error : Bad value (line 497, file "Code") I had done all the settings as you described at Link. But at the end, when i click on the menu Run->start it displays an error that "Bad value (line 497, file "Code")". Google Scriprt which I used is posted herein below, /* global Logger CalendarApp ScriptApp ContactsApp Utilities Calendar UrlFetchApp MailApp */ /* * Thanks to this script you are going to receive an email before the birthday of each of your contacts. * The script is easily customizable via some variables listed below. */ // START MANDATORY CUSTOMIZATION // You need to personalize these values, otherwise the script won't work. /* * GOOGLE EMAIL ADDRESS * * First of all specify the gmail address of your Google Account. * This is needed to retrieve informations about your contacts. */ var myGoogleEmail = 'keynote.arjun@gmail.com'; /* * NOTIFICATION EMAIL ADDRESS * * Now specify to which email address the notifications should be sent. * This can be the same email address of the previous line or any other email address. */ var myEmail = 'rajesh@femaonline.com'; /* * ID OF THE BIRTHDAY CALENDAR * * Open up https://calendar.google.com, in the menu on the left click on the arrow next to the birthday calendar * and choose 'Calendar settings', finally look for a the "Calendar ID field" (it should be something similar to * #contacts@group.v.calendar.google.com): copy and paste it between the quotes in the next line. */ var calendarId = 'keynotearjun@gmail.com'; /* * YOUR TIMEZIONE * * If you need to adjust the timezone of the email notifications use this variable. * * Accepted values: * GMT/UTC - examples: 'UTC+2' 'GMT-4' * regional timezones: 'Europe/Berlin' (See here for a complete list: http://joda-time.sourceforge.net/timezones.html) */ var myTimeZone = 'UTC+05:30'; /* * HOUR OF THE NOTIFICATION * * Specify at which hour of the day would you like to receive the email notifications. * This must be a number between 0 and 23. */ var notificationHour = 9; /* * HOW MANY DAYS BEFORE BIRTHDAY * * Here you have to decide when you want to receive the email notification. * Insert between the square brackets a comma-separated list of numbers, where each number * represents how many day before a birthday you want to be notified. * If you want to be notified only once then enter a single number between the brackets. * * Examples: * [0] means "Notify me the day of the birthday"; * [0, 7] means "Notify me the day of the birthday and 7 days before"; * [0, 1, 7] means "Notify me the day of the birthday, the day before and 7 days before"; * * Note: in any case you will receive one email per day: all the notifications will be grouped * together in that email. */ var anticipateDays = [0]; /* * LANGUAGE * * For internationalization (translation) enter the two-digit lang-code here (to add your language just fill in the * 'i18n' hash below and change lang here to match that). */ var lang = 'en'; // For places where an indent is used for display reasons (in plaintext email), this number of spaces is used. var indentSize = 2; // END MANDATORY CUSTOMIZATION // START DEBUGGING OPTIONS // When debugging is not wanted you can set this true to disable debugging calls, for a slight speedup. var noLog = false; // When debugging (noLog == false) and you want the logs emailed too, set this to true. var sendLog = false; /* * The test() function can be run on a specified date as if it is "today". Specify that date here in the format * YEAR/MONTH/DAY HOUR:MINUTE:SECOND * Choose a date you know should trigger a birthday notification. */ var fakeTestDate = '2017/02/14 06:00:00'; // END DEBUGGING OPTIONS /* * There is no need to edit anything below this line. * The script will work if you inserted valid values up until here, however feel free to take a peek at my code ;) */ var version = '2.1'; // Merge an array at the end of an existing array. if (typeof Array.prototype.extend === 'undefined') { Array.prototype.extend = function (array) { var i; for (i = 0; i < array.length; ++i) { this.push(array[i]); } return this; }; } if (typeof String.prototype.format === 'undefined') { String.prototype.format = function () { var args; args = arguments; return this.replace(/\{(\d+)\}/g, function (match, number) { return typeof args[number] !== 'undefined' ? args[number] : match ; }); }; } var indent = Array(indentSize + 1).join(' '); var i18n = { // For all languages, if a translation is not present the untranslated string // is returned, so just leave out translations which are the same as the English. // An entry for 'en' marks it as a valid lang config-option, but leave it empty // to just return unaltered phrases. 'en': {}, 'el': { 'UNKNOWN': 'ΑΓΝΩΣΤΟ', 'Age': 'Ηλικία', 'Birthday': 'Γενέθλια', 'Birthday today': 'Γενέθλια σήμερα', 'Birthday tomorrow': 'Γενέθλια αύριο', 'Birthday in {0} days': 'Γενέθλια σε {0} ημέρες', 'Hey! Don\'t forget these birthdays': 'Μην ξεχάσετε αυτά τα γενέθλια', 'Google Calendar Contacts Birthday Notification': 'Ενημερώσεις Γενεθλίων του Ημερολογίου Google', 'version': 'εκδοχή', 'by': 'από τον', // τον=masculine,την=feminine (using the masculine, in one place, for now but may need more context in future) 'dd-MM-yyyy': 'dd-MM-yyyy', 'send email now': 'στείλτε email τώρα', 'Mobile phone': 'Κινητό', 'Work phone': 'Τηλέφωνο εργασίας', 'Home phone': 'Τηλέφωνο οικίας', 'Main phone': 'Κύριο τηλέφωνο', }, 'it': { 'UNKNOWN': 'SCONOSCIUTO', 'Age': 'Età', 'Birthday': 'Compleanno', 'Birthday today': 'Compleanno oggi', 'Birthday tomorrow': 'Compleanno domani', 'Birthday in {0} days': 'Compleanno fra {0} giorni', 'Hey! Don\'t forget these birthdays': 'Hey! Non dimenticare questi compleanni', 'version': 'versione', 'by': 'by', 'dd-MM-yyyy': 'dd-MM-yyyy', 'send email now': 'invia email ora', 'Mobile phone': 'Cellulare', 'Work phone': 'Telefono di lavoro', 'Home phone': 'Telefono di casa', 'Main phone': 'Telefono principale', }, /* To add a language: '[lang-code]': { '[first phrase]': '[translation here]', '[second phrase]': '[translation here]', ... } */ }; var birthdayCalendar = CalendarApp.getCalendarById(calendarId); var calendarTimeZone = birthdayCalendar ? birthdayCalendar.getTimeZone() : null; var inlineImages; // Replace a Field.Label object with its "beautified" text representation. function beautifyLabel (label) { switch (label) { case ContactsApp.Field.MOBILE_PHONE: return _('Mobile phone'); case ContactsApp.Field.WORK_PHONE: return _('Work phone'); case ContactsApp.Field.HOME_PHONE: return _('Home phone'); case ContactsApp.Field.MAIN_PHONE: return _('Main phone'); default: return label; } } /* * Get the translation of a string. * If the language or the chosen string is invalid return the string itself. */ function _ (string) { return i18n[lang][string] || string; } function doLog (arg) { noLog || Logger.log(arg); } /* * Look for birthdays on a certain date. * If testDate is not specified Date.now() will be used. */ function checkBirthdays (testDate) { var anticipate, subjectPrefix, subjectBuilder, bodyPrefix, bodySuffix1, bodySuffix2, bodyBuilder, htmlBodyBuilder, now, subject, body, htmlBody; doLog('Starting run of GoogleCalendarBirthdayNotifications version ' + version + '.'); // The script needs this value in milliseconds, but the user entered it in days. anticipate = anticipateDays.map(function (n) { return 1000 * 60 * 60 * 24 * n; }); // Verify that the birthday calendar exists. if (!birthdayCalendar) { doLog('Error: Birthday calendar not found!'); doLog('Please follow the instructions at this page to activate it: https://support.google.com/calendar/answer/6084659?hl=en'); return; } // Start building the email notification text. subjectPrefix = _('Birthday') + ': '; subjectBuilder = []; bodyPrefix = _('Hey! Don\'t forget these birthdays') + ':'; bodySuffix1 = _('Google Calendar Contacts Birthday Notification') + ' (' + _('version') + ' ' + version + ')'; bodySuffix2 = _('by ') + 'Giorgio Bonvicini'; // The email is built both with plain text and HTML text. bodyBuilder = []; htmlBodyBuilder = []; // Use the testDate if specified, otherwise use todays' date. now = testDate || new Date(); doLog('Date used: ' + now); inlineImages = {}; /* * Look for birthdays on each of the days specified by the user. * timeInterval represents how many milliseconds in the future to check. */ anticipate.forEach( function (timeInterval) { var optionalArgs, birthdays, formattedDate, whenIsIt; // Set the search filter to include only events happening 'timeInterval' milliseconds after now. optionalArgs = { // Filter only events happening between 'now + timeInterval'... timeMin: Utilities.formatDate(new Date(now.getTime() + timeInterval), calendarTimeZone, 'yyyy-MM-dd\'T\'HH:mm:ss\'Z\''), // ... and 'now + timeInterval + 1 sec'. timeMax: Utilities.formatDate(new Date(now.getTime() + timeInterval + 1000), calendarTimeZone, 'yyyy-MM-dd\'T\'HH:mm:ss\'Z\''), // Treat recurring (like birthdays) events as single events. singleEvents: true }; doLog('Checking birthdays from ' + optionalArgs.timeMin + ' to ' + optionalArgs.timeMax); // Get all the matching events. birthdays = Calendar.Events.list(calendarId, optionalArgs).items; doLog('Found ' + birthdays.length + ' birthdays in this time range.'); // If no event is found for this particular timeInterval skip it. if (birthdays.length < 1) { return; } formattedDate = Utilities.formatDate(new Date(now.getTime() + timeInterval), calendarTimeZone, _('dd-MM-yyyy')); // Build the headers of birthday grouping by date. bodyBuilder.push(' * '); htmlBodyBuilder.push('<dt style="margin-left:0.8em;font-style:italic">'); switch (timeInterval / (24 * 60 * 60 * 1000)) { case 0: whenIsIt = _('Birthday today') + ' (' + formattedDate + ')'; break; case 1: whenIsIt = _('Birthday tomorrow') + ' (' + formattedDate + ')'; break; default: whenIsIt = _('Birthday in {0} days').format(timeInterval / (24 * 60 * 60 * 1000)) + ' (' + formattedDate + ')'; } bodyBuilder.push(whenIsIt, ':\n'); htmlBodyBuilder.push(whenIsIt, '</dt><dd style="margin-left:0.4em;padding-left:0"><ul style="list-style:none;margin-left:0;padding-left:0;">'); // Add each of the birthdays for this timeInterval. birthdays.forEach( function (event, i) { var contact; doLog('Contact #' + i); contact = new Contact(event); subjectBuilder.push(contact.fullName); bodyBuilder.extend(contact.getPlainTextLine()); htmlBodyBuilder.extend(contact.getHtmlLine()); } ); bodyBuilder.push('\n'); htmlBodyBuilder.push('</ul></dd>'); } ); // If there is an email to send... if (bodyBuilder.length > 0) { subject = subjectPrefix + subjectBuilder.join(' - '); body = [bodyPrefix, '\n\n'] .concat(bodyBuilder) .concat(['\n\n', indent, bodySuffix1, '\n', indent, bodySuffix2, '\n']) .join(''); htmlBody = ['<h3>', bodyPrefix, '</h3><dl>'] .concat(htmlBodyBuilder) .concat(['</dl><hr/><p style="text-align:center;font-size:smaller"><a href="https://github.com/GioBonvi/GoogleCalendarBirthdayNotifications">', bodySuffix1, '</a><br/>', bodySuffix2, '</p>']) .join(''); // ...send the email notification. doLog('Sending email...'); MailApp.sendEmail({ to: myEmail, subject: subject, body: body, htmlBody: htmlBody, inlineImages: inlineImages }); doLog('Email sent.'); } // Send the log if the debug options say so. if (!noLog && sendLog) { MailApp.sendEmail({ to: myEmail, subject: 'Logs for birthday-notification run', body: Logger.getLog() }); } } /* * Extract contact data from a birthday event and integrate it with additional data * recovered directly from Google Contact through the contactId field if present. */ var Contact = function (event) { var eventData, googleContact, currentYear, birthdayYear, phoneFields; // Extract basic data from the event description. eventData = event.gadget.preferences; this.id = (typeof eventData['goo.contactsContactId'] === 'undefined') ? '' : eventData['goo.contactsContactId']; this.fullName = (typeof eventData['goo.contactsFullName'] === 'undefined') ? '' : eventData['goo.contactsFullName']; this.email = (typeof eventData['goo.contactsEmail'] === 'undefined') ? '' : eventData['goo.contactsEmail']; this.photo = (typeof eventData['goo.contactsPhotoUrl'] === 'undefined') ? '' : eventData['goo.contactsPhotoUrl']; this.age = ''; this.phoneFields = []; if (this.email !== '') { doLog('Has email.'); } if (this.fullName !== '') { doLog('Has full name'); } if (this.photo !== '') { doLog('Has photo.'); } // If the contact has a contactId field try to get the Google Contact corresponding to that contactId. if (this.id !== '') { googleContact = ContactsApp.getContactById('http://www.google.com/m8/feeds/contacts/' + encodeURIComponent(myGoogleEmail) + '/base/' + this.id); } // If a valid Google Contact exists extract some additional data. if (googleContact) { // Extract contact's age if the contact has the birthday year. if (googleContact.getDates(ContactsApp.Field.BIRTHDAY)[0]) { doLog('Has birthday year.'); currentYear = Utilities.formatDate(new Date(event.start.date.replace(/-/g, '/')), calendarTimeZone, 'yyyy'); birthdayYear = googleContact.getDates(ContactsApp.Field.BIRTHDAY)[0].getYear(); this.age = birthdayYear !== '' ? (currentYear - birthdayYear).toFixed(0) : ''; } // Extract contact's phone numbers. phoneFields = googleContact.getPhones(); if (phoneFields.length > 0) { this.phoneFields = phoneFields; doLog('Has phones.'); } } /* * Use the extracted data to build a plain line of text displaying all the * collected data about the contact. */ this.getPlainTextLine = function () { var line; line = []; // Full name. line.push('\n', indent, this.fullName); // Age. if (this.age !== '') { line.push(' - ', _('Age'), ': ', this.age); } if (this.email !== '' || typeof this.phoneFields !== 'undefined') { line.push(' ('); // Email address. if (this.email !== '') { line.push(this.email); } // Phone numbers. this.phoneFields.forEach(function (phoneField, i) { var label; if (i !== 0 || this.email !== '') { line.push(' - '); } label = phoneField.getLabel(); if (label !== '') { line.push('[', beautifyLabel(label), '] '); } line.push(phoneField.getPhoneNumber()); }); line.push(')'); } line.push('\n'); return line; }; /* * Use the extracted data to build a line of HTML text displaying all the * collected data about the contact. */ this.getHtmlLine = function () { var line, imgCount; line = []; line.push('<li>'); // Profile photo. if (this.photo !== '') { imgCount = Object.keys(inlineImages).length; inlineImages['contact-img-' + imgCount] = UrlFetchApp.fetch(this.photo).getBlob().setName('contact-img-' + imgCount); line.push('<img src="cid:contact-img-' + imgCount + '" style="height:1.4em;margin-right:0.4em" />'); } // Full name. line.push(this.fullName); // Age. if (this.age !== '') { line.push(' - ', _('Age'), ': ', this.age); } if (this.email !== '' || typeof this.phoneFields !== 'undefined') { line.push(' ('); // Email address. if (this.email !== '') { line.push(this.email); } // Phone fields. this.phoneFields.forEach(function (phoneField, i) { var label; if (i !== 0 || this.email !== '') { line.push(' - '); } label = phoneField.getLabel(); if (label !== '') { line.push('[', beautifyLabel(label), '] '); } line.push('<a href="tel:', phoneField.getPhoneNumber(), '">', phoneField.getPhoneNumber(), '</a>'); }); line.push(')'); } // Mailto link. if (this.email !== '') { line.push(' <a href="mailto:', this.email, '">', _('send email now'), '</a>'); } return line; }; }; // Start the notification service. function start () { stop(); ScriptApp.newTrigger('normal') .timeBased() .atHour(notificationHour) .everyDays(1) .inTimezone(myTimeZone) .create(); } // Stop the notification service. function stop () { var triggers; // Delete all the triggers. triggers = ScriptApp.getProjectTriggers(); for (var i = 0; i < triggers.length; i++) { ScriptApp.deleteTrigger(triggers[i]); } } // Check if notification service is running. function status () { var toLog = 'Notifications are'; if (ScriptApp.getProjectTriggers().length < 1) { toLog += ' not'; } toLog += ' running.'; Logger.log(toLog); if (!noLog && sendLog) { MailApp.sendEmail({ to: myEmail, subject: 'Status for birthday-notification', body: Logger.getLog() }); } } // Normal function call (This function is called by the timed trigger). function normal () { checkBirthdays(); } /* * Use this function to test the script. Edit the date in the debugging * configuration above and click "Run"->"test" in the menu at the top * of the Google script interface. */ function test () { var testDate; testDate = new Date(fakeTestDate); doLog('Testing.'); doLog('Test date: ' + testDate); checkBirthdays(testDate); } Please help me and thanks in advance. regards, Arjun Chavda Hello @arj0903, thank you for reaching out. At a first glance the only thing I can see that could cause some problems is line 35: are you sure you enterd the right calendar ID? Please make sure you followed the procedure correctly: /* * ID OF THE BIRTHDAY CALENDAR * * Open up https://calendar.google.com, in the menu on the left click on the arrow next to the birthday calendar * and choose 'Calendar settings', finally look for a the "Calendar ID field" (it should be something similar to * #contacts@group.v.calendar.google.com): copy and paste it between the quotes in the next line. */ What you enterd for that line looks more like an email address than a calendar ID... I am not fully convinced, however, as this should not produce the kind of error you are describing. Please verify this firts hypothesis and if this does not solve your problem I will be happy to further assit you. @arj0903 Actually it looks like the problem lies in line 46 (myTimeZone declaration). Google documentation about timezones in Google Scripts is extremely poor and I did some test to understand them better, although it looks like I failed to notice that UTC is NOT supported. You should change UTC+05:30 to something else (GMT or name-based timezones should work) and this should solve the issue. I'll soon add a piece of code to verify that the timezone is actually valid and throw a meaningful error if not. Hope this helps! I've added an additional check which should cause a meaningful error message to be displayed in cases such as this In the latest version. I0ve removed the UTC example as well. Hello, @arj0903 Are you still experiencing this problem? Best regards Hello @arj0903, on 24/08/2017 this issue will be closed in compliance with the rule about unresponsive help requests. If you are still experiencing this problem just reply to this issue explaining what problems you are experiencing. Best regards This issue has been closed in compliance with the rule about unresponsive help requests. @arj0903: if you are still experiencing this problem post a message in this issue asking for it to be re-opened. Best regards
gharchive/issue
2017-05-24T14:57:50
2025-04-01T04:32:34.935696
{ "authors": [ "GioBonvi", "arj0903" ], "repo": "GioBonvi/GoogleCalendarBirthdayNotifications", "url": "https://github.com/GioBonvi/GoogleCalendarBirthdayNotifications/issues/12", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1189465048
Improve test coverage Description Our test coverage is currently at 93%. This sounds good, but some files actually cheat and make too much use of # pragma: no cover. Here is a table of the top 3 (ab)users of this directive: Module statements missing excluded coverage ggshield/ci.py 10 0 154 100% ggshield/dev_scan.py 39 0 75 100% ggshield/hook_cmd.py 30 0 42 100% To do [ ] Inspect these 3 files [ ] If possible, write tests to cover some of the currently excluded code and remove the "no cover" directive This has gotten better nowadays.
gharchive/issue
2022-04-01T08:46:39
2025-04-01T04:32:34.974368
{ "authors": [ "agateau-gg" ], "repo": "GitGuardian/ggshield", "url": "https://github.com/GitGuardian/ggshield/issues/191", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
194552674
Remove lodash dependency lodash is a big dependency in the editor, and we can use ES6 features and Immutable.js most of the time (or small npm modules). Fixed by #29
gharchive/issue
2016-12-09T09:45:00
2025-04-01T04:32:34.991146
{ "authors": [ "SamyPesse" ], "repo": "GitbookIO/repofs", "url": "https://github.com/GitbookIO/repofs/issues/25", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2599964417
🛑 twoc is down In 16a024a, twoc (twoc.co.uk) was down: HTTP code: 0 Response time: 0 ms Resolved: twoc is back up in 1e0ab6a after 13 minutes.
gharchive/issue
2024-10-20T04:28:28
2025-04-01T04:32:34.993541
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/102396", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2606564764
🛑 cnn is down In 4578ff0, cnn (cnn.com) was down: HTTP code: 0 Response time: 0 ms Resolved: cnn is back up in cb7b8c3 after 16 minutes.
gharchive/issue
2024-10-22T21:33:54
2025-04-01T04:32:34.995836
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/102516", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2724973745
🛑 saunta is down In 650e441, saunta (saunta.com) was down: HTTP code: 0 Response time: 0 ms Resolved: saunta is back up in a2605b0 after 21 minutes.
gharchive/issue
2024-12-08T05:57:55
2025-04-01T04:32:34.998349
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/104894", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2739340494
🛑 xxlhd is down In 0a01fb8, xxlhd (xxlhd.com) was down: HTTP code: 0 Response time: 0 ms Resolved: xxlhd is back up in 4541839 after 22 minutes.
gharchive/issue
2024-12-13T22:58:12
2025-04-01T04:32:35.000649
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/105337", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1052852683
🛑 lings is down In 31cc67b, lings (lings.com) was down: HTTP code: 403 Response time: 415 ms Resolved: lings is back up in 23d9279.
gharchive/issue
2021-11-14T05:37:22
2025-04-01T04:32:35.002921
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/11222", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1060920505
🛑 orkut is down In 60a50e6, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in d6e61ae.
gharchive/issue
2021-11-23T07:39:58
2025-04-01T04:32:35.005210
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/11746", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1062022863
🛑 zshareblog is down In 5da06e1, zshareblog (zshareblog.com) was down: HTTP code: 0 Response time: 0 ms Resolved: zshareblog is back up in 00f4722.
gharchive/issue
2021-11-24T05:37:14
2025-04-01T04:32:35.007533
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/11802", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1086507664
🛑 orkut is down In 859dc99, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in 5f2966f.
gharchive/issue
2021-12-22T06:47:43
2025-04-01T04:32:35.010044
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/13380", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1091816747
🛑 orkut is down In 95a4c5c, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in cdd2743. Resolved: orkut is back up in cdd2743.
gharchive/issue
2022-01-01T13:21:57
2025-04-01T04:32:35.013131
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/13860", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1138117683
🛑 orkut is down In d8780c7, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in d7fff27.
gharchive/issue
2022-02-15T03:09:24
2025-04-01T04:32:35.015905
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/17166", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1148856704
🛑 orkut is down In ea6bb69, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in 9a23cee.
gharchive/issue
2022-02-24T04:46:09
2025-04-01T04:32:35.018175
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/17908", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1178944036
🛑 orkut is down In 13f1e8b, orkut (orkut.com.br) was down: HTTP code: 403 Response time: 231 ms Resolved: orkut is back up in d4a73d9.
gharchive/issue
2022-03-24T04:08:07
2025-04-01T04:32:35.020468
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/19966", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1251513017
🛑 ojolink is down In 15c9d2c, ojolink (ojolink.fr) was down: HTTP code: 0 Response time: 0 ms Resolved: Ojolink is back up in a1c2705.
gharchive/issue
2022-05-28T07:34:15
2025-04-01T04:32:35.022984
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/23593", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1257041807
🛑 ojolink is down In ed014ba, ojolink (ojolink.fr) was down: HTTP code: 0 Response time: 0 ms Resolved: Ojolink is back up in 4a7ba80.
gharchive/issue
2022-06-01T21:06:49
2025-04-01T04:32:35.025300
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/23977", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1263672512
🛑 orkut is down In 578a99b, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in e0acbe0.
gharchive/issue
2022-06-07T17:58:18
2025-04-01T04:32:35.027566
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/24542", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1271919272
🛑 hcr is down In 497a46f, hcr (hcr.co.uk) was down: HTTP code: 401 Response time: 717 ms Resolved: hcr is back up in e5ec3d9.
gharchive/issue
2022-06-15T09:09:46
2025-04-01T04:32:35.029865
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/25767", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
941226237
🛑 feest-idee is down In f20d214, feest-idee (feest-idee.nl) was down: HTTP code: 0 Response time: 0 ms Resolved: feest-idee is back up in 56a218c.
gharchive/issue
2021-07-10T10:56:26
2025-04-01T04:32:35.032226
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/2628", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1279538450
🛑 hcr is down In 5903782, hcr (hcr.co.uk) was down: HTTP code: 401 Response time: 1399 ms Resolved: hcr is back up in 50aaa3a.
gharchive/issue
2022-06-22T05:09:42
2025-04-01T04:32:35.034706
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/26922", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1376990218
🛑 penispedia is down In 8e56144, penispedia (penispedia.de) was down: HTTP code: 0 Response time: 0 ms Resolved: penispedia is back up in fac3345.
gharchive/issue
2022-09-18T08:57:27
2025-04-01T04:32:35.037087
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/34027", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1391070391
🛑 163 is down In 9063a29, 163 (163.com) was down: HTTP code: 0 Response time: 0 ms Resolved: 163 is back up in cfd44ba.
gharchive/issue
2022-09-29T15:30:31
2025-04-01T04:32:35.039337
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/34542", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1391981589
🛑 feest-start is down In f688bb7, feest-start (feest-start.nl) was down: HTTP code: 0 Response time: 0 ms Resolved: feest-start is back up in 375801f.
gharchive/issue
2022-09-30T07:33:15
2025-04-01T04:32:35.041671
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/34566", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1398132168
🛑 ojolink is down In 77b30c1, ojolink (ojolink.net) was down: HTTP code: 403 Response time: 503 ms Resolved: Ojolink is back up in 92eae7c.
gharchive/issue
2022-10-05T17:31:23
2025-04-01T04:32:35.043957
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/34911", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1410453265
🛑 ojolink is down In 6bb696a, ojolink (ojolink.net) was down: HTTP code: 403 Response time: 519 ms Resolved: Ojolink is back up in 24b34e8.
gharchive/issue
2022-10-16T11:04:43
2025-04-01T04:32:35.046453
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/35658", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1428816693
🛑 ojolink is down In 677016b, ojolink (ojolink.net) was down: HTTP code: 403 Response time: 508 ms Resolved: Ojolink is back up in c4c251c.
gharchive/issue
2022-10-30T14:03:14
2025-04-01T04:32:35.048729
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/37234", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1429341235
🛑 orkut is down In b2bd9cf, orkut (orkut.com.br) was down: HTTP code: 429 Response time: 764 ms Resolved: orkut is back up in d2f42a6.
gharchive/issue
2022-10-31T06:43:34
2025-04-01T04:32:35.050969
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/37330", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1451726154
🛑 wvtaxsales is down In e7d0e0c, wvtaxsales (wvtaxsales.com) was down: HTTP code: 0 Response time: 0 ms Resolved: wvtaxsales is back up in 850c5c5.
gharchive/issue
2022-11-16T14:39:24
2025-04-01T04:32:35.053492
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/39720", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
960550309
🛑 orkut is down In 692bdd2, orkut (orkut.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: orkut is back up in 76ff1ed.
gharchive/issue
2021-08-04T14:32:57
2025-04-01T04:32:35.055790
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/4421", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1501233067
🛑 inkcityusa is down In 6695b6e, inkcityusa (inkcityusa.com) was down: HTTP code: 0 Response time: 0 ms Resolved: inkcityusa is back up in 797504a.
gharchive/issue
2022-12-17T07:23:03
2025-04-01T04:32:35.058282
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/44951", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1515084175
🛑 orkut is down In 71efe3b, orkut (orkut.com.br) was down: HTTP code: 429 Response time: 975 ms Resolved: orkut is back up in f22be2c.
gharchive/issue
2022-12-31T10:05:19
2025-04-01T04:32:35.060643
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/47410", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1528790528
🛑 orkut is down In 6468e32, orkut (orkut.com.br) was down: HTTP code: 429 Response time: 965 ms Resolved: orkut is back up in 995edac.
gharchive/issue
2023-01-11T10:27:36
2025-04-01T04:32:35.062917
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/49198", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
970041778
🛑 harlotwear is down In 69ab3c8, harlotwear (harlotwear.com) was down: HTTP code: 504 Response time: 12911 ms Resolved: harlotwear is back up in 7b50d5d.
gharchive/issue
2021-08-13T04:04:37
2025-04-01T04:32:35.065219
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/4944", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1554468292
🛑 ojolink is down In d058275, ojolink (ojolink.net) was down: HTTP code: 403 Response time: 1450 ms Resolved: Ojolink is back up in aaa2b2d.
gharchive/issue
2023-01-24T07:36:33
2025-04-01T04:32:35.067480
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/50815", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1561051546
🛑 citybuild is down In 130d8cc, citybuild (citybuild.bg) was down: HTTP code: 0 Response time: 0 ms Resolved: citybuild is back up in 3b88460.
gharchive/issue
2023-01-28T23:38:08
2025-04-01T04:32:35.069936
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/51347", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1563634884
🛑 Mangahype is down In d1dac3f, Mangahype (https://www.mangahype.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Mangahype is back up in 278fb2f.
gharchive/issue
2023-01-31T04:28:59
2025-04-01T04:32:35.072253
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/51622", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1563698851
🛑 orkut is down In 6f560d9, orkut (orkut.com.br) was down: HTTP code: 429 Response time: 1129 ms Resolved: orkut is back up in 931653b.
gharchive/issue
2023-01-31T05:50:12
2025-04-01T04:32:35.074511
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/51629", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1594105334
🛑 orkut is down In f26bd73, orkut (orkut.com.br) was down: HTTP code: 429 Response time: 832 ms Resolved: orkut is back up in b7aff49.
gharchive/issue
2023-02-21T21:11:33
2025-04-01T04:32:35.076779
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/53496", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1786946387
🛑 ojolink is down In d9f3ed2, ojolink (ojolink.fr) was down: HTTP code: 0 Response time: 0 ms Resolved: Ojolink is back up in 1539929.
gharchive/issue
2023-07-04T00:05:17
2025-04-01T04:32:35.079034
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/63796", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1802502778
🛑 ojolink is down In 56787bd, ojolink (ojolink.fr) was down: HTTP code: 0 Response time: 0 ms Resolved: Ojolink is back up in ed813f7.
gharchive/issue
2023-07-13T08:38:09
2025-04-01T04:32:35.081509
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/64514", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1816900206
🛑 ojolink is down In 20dc563, ojolink (ojolink.fr) was down: HTTP code: 0 Response time: 0 ms Resolved: Ojolink is back up in c7efcdf.
gharchive/issue
2023-07-22T21:26:36
2025-04-01T04:32:35.083868
{ "authors": [ "GiuseppeFilingeri" ], "repo": "GiuseppeFilingeri/upgraded-symmetrical-waddle", "url": "https://github.com/GiuseppeFilingeri/upgraded-symmetrical-waddle/issues/65325", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }