added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:39:57.998010
2021-11-17T11:21:42
1056011351
{ "authors": [ "0x7CFE", "atenjin", "athei", "pepyakin" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9502", "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/issues/10297" }
gharchive/issue
Refactor and rename sandboxing implementation The sandboxing code inside the client which is responsible for running the contracts of pallet-contracts needs a refactor. Currently, the different execution engines for contracts are crammed into sc-executor-common. Neither should different execution engines be in the same crate nor should runtime execution and sandbox execution be intertwined in this way. For that reason we want to refactor that code in order to have proper abstractions that allows us to move the sandboxing code to its own set of crates. Rename After a discussion with @pepyakin (see comments below) we agreed that we should also move away from the term "sandboxing" for this whole thing. One proposal what I also back is to use the terminology "Wasm Virtualization". TODO [x] #10563 [ ] Create a trait to abstract over sandboxing execution engines ("backends"). [ ] Use dynamic dispatch with that trait to make engine selection possible at runtime (command line argument). [ ] Move sandboxing code into its own crate client/wasm-virt with one additional crate per execution engine (sc-wasm-virt-wasmi, sc-wasm-virt-wasmer). This should also move all wasm virtualization related traits that needs to be implement by the runtime executor to the new location. Another open question is whether we need to expand the API to allow for chunked execution so that code merkelization can be implemented. However, current thinking is that we can and should get away without execution engine support. I think it is reasonable to go forward with a non chunked version risking that we might need to maintain two APIs. Otherwise this is blocked by #9431. I think, by carefully designing the API we can future-proof it to support potential chunked execution similar to what we have done for wasmi during our experiment. Maybe. It adds complexity, though. For example, we would need some kind of feature discovery then because not all execution engines will support this right off the bat. As a matter of fact none will support it as we only put it in for future proofing. Just creating a new version of the API when the need arises is better in my opinion. Future proofing APIs is worth nothing when users will depend on the implementation. Excerpt from our discussion about changing the terminology from sandoxing to a term based on virtualization, e.g. wasm virtualization API, hypervisor API, or whatever. Specifically, I think it is just too general and does not represent what it actually does. What is usually understood by sandbox? Well, it's when some code is put into an environment where it cannot reach to anything except things that was explicitly designated to be used by that sandboxed code. If we agree on that definition, then the "runtime" is exactly that. The thing that executes "runtime" (which would be called runtime by normal people) can be also called a sandbox. However, virtualization is really more appropriate term here I feel. This is because it fits really an analogy. Virtualization is where you have some medium that provides virtualization facilities, some hypervisor/supervisor that virtualizes or multiplexes the underlying hardware, emulating other and providing services, and the guest which consumes those services, who uses the hardware through the hypervisor, and ultimately is controlled by the hypervisor Updated the top post with new information after a discussion I had with @0x7CFE, today. Refined the top posting again after the picture got a bit clearer on how the final refactoring should look like. I narrowed down the scope if this issue and created a a proper task list. What exactly do you *need it for? run a piece of wasm in runtime You can still do that: Use sp_sandbox::embedded_executor. It doesn't rely on this interface but includes a wasmi into the runtime. No one is working on this anymore. Since we decided to go with wasmi for now and we don't really have the resources to work on that we will close it.
2025-04-01T06:39:58.001465
2022-09-09T09:26:49
1367545201
{ "authors": [ "Ank4n", "MrishoLukamba" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9503", "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/issues/12227" }
gharchive/issue
MaxUnlockingChunks should be configurable via Runtime StakingLedger in the pallet-staking is using a hardcoded value of MaxUnlockingChunks. Pallet-staking also has a MaxUnlockingChunks in the configuration which is not used anywhere. I discussed this with @kianenigma, we should remove the hard coded value and use the value we get from Config. Taking it Taking it @MrishoLukamba Feel free to pick it up.
2025-04-01T06:39:58.006398
2022-04-21T08:43:57
1210711490
{ "authors": [ "GopherJ", "anantasty", "bkchr" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9504", "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/pull/11256" }
gharchive/pull-request
Fix pallet_assets no_std compilation Could you port this to polkadot-v0.9.19 as well? thanks Hmm, it is included in the bin/node/runtime. So, it should already be compiled as part of the CI in wasm. Not sure why it didn't failed there :thinking: What compiler version are you using? @bkchr we are using: nightly-2021-11-07, somehow I cannot compile into wasm yet due to missing vec import, wierd I found the problem. I bet with you that you have the std feature of sp-std somewhere in your project enabled. Please check this. Not sure if it's caused by compiler though, it's also reporting some others: let me check thanks @bkchr you are awesome! fixed now I am running into the same issue in v0.9.26 I have tried wit the sp-std/std flag enabled and disabled. Suggestions? @anantasty it can also be any other crate that propagates std feature to sp-std. You can use cargo tree to find out which crate enables the std feature. Thanks @bkchr
2025-04-01T06:39:58.008439
2022-05-14T15:26:15
1236029803
{ "authors": [ "athei" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9505", "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/pull/11414" }
gharchive/pull-request
contracts: Get rid of #[pallet::without_storage_info] All the pallet storage now implements MaxEncodedLen and hence we can remove #[pallet::without_storage_info]. Most of the changes were replacing Vec with BoundedVec and adding a derive for MaxEncodedLen. cumulus companion: https://github.com/paritytech/cumulus/pull/1261 bot merge bot merge
2025-04-01T06:39:58.010119
2023-01-13T16:34:24
1532595964
{ "authors": [ "ggwpez" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9506", "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/pull/13140" }
gharchive/pull-request
[DNM] Test CI Various CI testing. Works es expected @rcny :+1: bot help bot help wrong repo bot bench $ pallet dev pallet_balances
2025-04-01T06:39:58.014029
2023-02-06T18:49:41
1573101368
{ "authors": [ "kianenigma", "liamaharon", "paritytech-cicd-pr" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9507", "repo": "paritytech/substrate", "url": "https://github.com/paritytech/substrate/pull/13319" }
gharchive/pull-request
track total value locked in pools onchain could resolve #12838 as well. [ ] fix tests [ ] migration The CI pipeline was cancelled due to failure one of the required jobs. Job name: cargo-check-each-crate Logs: https://gitlab.parity.io/parity/mirrors/substrate/-/jobs/2357659 The CI pipeline was cancelled due to failure one of the required jobs. Job name: test-linux-stable-int Logs: https://gitlab.parity.io/parity/mirrors/substrate/-/jobs/2357655 Not at the top of my TODO list, but plan to getting around to it in the next couple of weeks. Closing in favor of https://github.com/paritytech/substrate/pull/14775
2025-04-01T06:39:58.045304
2022-09-24T08:51:38
1384593768
{ "authors": [ "lootsorrow", "parlance-zz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9508", "repo": "parlance-zz/g-diffuser-lib", "url": "https://github.com/parlance-zz/g-diffuser-lib/issues/41" }
gharchive/issue
Add new g-diffuser command "enhance" Rescale the input image to a higher resolution and use inpainting with a constant mask of some opacity, effectively using SD for super-resolution. The same function could be aliased as a style transfer function, since it would do the same thing depending on opacity value and the prompt supplied. optional parameter to increment or decrement certain parameters like scale or strength during said operation so that the longer it goes on the more insane the images get as the str gets higher or the scale gets lower
2025-04-01T06:39:58.054218
2016-11-17T19:19:15
190138145
{ "authors": [ "MikhailKlemin", "han2015", "kervinson", "luckcry" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9509", "repo": "parnurzeal/gorequest", "url": "https://github.com/parnurzeal/gorequest/issues/117" }
gharchive/issue
I guess AddCookie/AddCookies doesnt work? Hi! I just tried to use it vs sites which use cookies to confirm language/currency, and it's just do not work, I am not sure why... I can only confirm that same sites with standard approach (cookiejar and http.Client(jar: cookieJar)) work just fine. Hi @MikhailKlemin I just have the same issue, but it works well if i put the AddCookies back the post/get. i just have the same issue, i have try many method, but i have not find the way, so i need you help@han2015, this is mine code , func GetGoRequest() { request := gorequest.New() //loginCookie := GetLoginCookie() // //fmt.Println("go request logincookie",loginCookie) cookie := &http.Cookie{Name: "PHPSESSID", Value: "t2cdimadecpoel6i8rcdomk1m6"} fmt.Println("go request cookie--->", cookie) _, _, err := request. Get("https://test-xxxxxx.cn/api/index.php?r=detail/brokerage-flows&token=cdkqqf1407307954&recommend_id=39ee6603-e648-49b1-12f5-e803efbf383a&devopenid=88374"). Send("r":"detail/recommend-flows","token":"cdkqqf1407307954","recommend_id":"39ee6603-e648-49b1-12f5-e803efbf383a"). AddCookie(cookie). End() if err != nil{ fmt.Printf("%v \n", err) } fmt.Println(request.Header) // this is always map[] } please, give me same help, thank you @kervinson correct your Send sentence refer doc(https://github.com/parnurzeal/gorequest/blame/develop/README.md#L120). go request cookie---> PHPSESSID=t2cdimadecpoel6i8rcdomk1m6 ++++++++++++++1 ++++++++++++++TypeJSON ++++++++++++++5 [Get https://test-xxxxxx.cn/api/index.php?devopenid=88374&r=detail%2Fbrokerage-flows&recommend_id=39ee6603-e648-49b1-12f5-e803efbf383a&token=cdkqqf1407307954: dial tcp: lookup test-xxxxxx.cn: no such host] [PHPSESSID=t2cdimadecpoel6i8rcdomk1m6] I met the same problem, my code is bvelow env : window 10, go 1.15, gorequest 1.6.0 package main import ( "github.com/parnurzeal/gorequest" "net/http" "strings" "log" ) func buildCookie(cookie string) []*http.Cookie { cookieSlice := make([]*http.Cookie, 0) params := strings.Split(cookie, ";") for _, line := range params { res := strings.Split(line, "=") cookieSlice = append(cookieSlice, &http.Cookie{Name: res[0], Value: res[1]}) } return cookieSlice } func main() { cookies := []string {"cookie1", "cookie2", "cookie3", "cookie4"..} r := gorequest.New() r = r.Set("Accept","text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8"). Set("Accept-Encoding","gzip, deflate, br"). Set("Accept-Language","zh-CN,zh;q=0.8,zh-TW;q=0.7,zh-HK;q=0.5,en-US;q=0.3,en;q=0.2"). Set("Host","www.tianyancha.com"). // Set("Referer","https://www.tianyancha.com/search/ocD?base=taiyuan"). Set("User-Agent","Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Firefox/68.0") index := 0 req, _ := http.NewRequest("GET", "https://www.tianyancha.com/company/3192219802", nil) req.Header.Set("Accept","text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8") req.Header.Set("Accept-Encoding","gzip, deflate, br") req.Header.Set("Accept-Language","zh-CN,zh;q=0.8,zh-TW;q=0.7,zh-HK;q=0.5,en-US;q=0.3,en;q=0.2") req.Header.Set("Host","www.tianyancha.com") // Set("Referer","https://www.tianyancha.com/search/ocD?base=taiyuan"). req.Header.Set("User-Agent","Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Firefox/68.0") for i := 0; i<3; i ++ { // cookies[index%len(cookies)]) r.AddCookies(buildCookie(cookies[index%len(cookies)])) // log.Println(r.Header) // r = r.Set("Cookie", cookies[index%len(cookies)]) response, _, _ := r.Get("https://www.tianyancha.com/company/3192219802").End() log.Println("url, ", response.Request.URL.String()) req.Header.Set("Cookie", cookies[index%len(cookies)]) client := &http.Client{} resp, _ := client.Do(req) log.Println(resp.Request.URL.String()) index ++ } } output 2020/11/11 16:20:42 url, https://www.tianyancha.com/login?from=https%3A%2F%2Fwww.tianyancha.com%2Fcompany%2F3192219802 2020/11/11 16:20:43 https://www.tianyancha.com/company/3192219802 2020/11/11 16:20:43 url, https://www.tianyancha.com/login?from=https%3A%2F%2Fwww.tianyancha.com%2Fcompany%2F3192219802 2020/11/11 16:20:43 https://www.tianyancha.com/company/3192219802 2020/11/11 16:20:44 url, https://www.tianyancha.com/login?from=https%3A%2F%2Fwww.tianyancha.com%2Fcompany%2F3192219802 2020/11/11 16:20:44 https://www.tianyancha.com/company/3192219802 anyway, i cannot get it through via gorequest~~~ Now I've solved this problem. Its really a misuse of cookie. When I read source code of gorequest, I found it would clear the agent stat every time before a new request if DoNotClearSuperAgent is default to false. So the solution is add this code between you set cookie and start a new request. agent := gorequest.New() agent.SetDoNotClearSuperAgent(true)
2025-04-01T06:39:58.077766
2017-12-20T08:51:40
283489592
{ "authors": [ "coveralls", "hit023" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9510", "repo": "parsonsmatt/kale", "url": "https://github.com/parsonsmatt/kale/pull/22" }
gharchive/pull-request
Add 'Syntax' data type. Something along these lines? References #9 . Coverage decreased (-0.8%) to 84.566% when pulling a5e1e0c7d6a8230e750122414a46a05bbd072f6c on hit023:cleaner-imports into 0468c68f1df8cb7d07d746085e020bc219daae33 on parsonsmatt:master. Coverage decreased (-0.8%) to 84.566% when pulling a5e1e0c7d6a8230e750122414a46a05bbd072f6c on hit023:cleaner-imports into 0468c68f1df8cb7d07d746085e020bc219daae33 on parsonsmatt:master. Coverage decreased (-0.8%) to 84.566% when pulling a5e1e0c7d6a8230e750122414a46a05bbd072f6c on hit023:cleaner-imports into 0468c68f1df8cb7d07d746085e020bc219daae33 on parsonsmatt:master. Coverage decreased (-0.8%) to 84.566% when pulling 6126353fbeb6a126cd3f3b412508f83d4bd1bcbf on hit023:cleaner-imports into 0468c68f1df8cb7d07d746085e020bc219daae33 on parsonsmatt:master.
2025-04-01T06:39:58.097498
2016-07-26T17:50:40
167669720
{ "authors": [ "chimon2000", "kevinSuttle", "pascalduez" ], "license": "unlicense", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9511", "repo": "pascalduez/postcss-apply", "url": "https://github.com/pascalduez/postcss-apply/issues/9" }
gharchive/issue
How to use with webpack? I am attempting to use this feature in webpack, but I cannot figure out how to inject custom properties. This is fairly easy to do for var let variables = { 'primary-toolbar': '#415464', 'secondary-toolbar': "#488fb4" } require("postcss-cssnext")({ features: { customProperties: { variables: variables, 'conversation-progress': ` { margin: 0 auto; position: fixed; top: 50%; left: 50%; }; ` }, applyRule: { } } }) However when attempting to do something similar for apply I get the no custom properties set declared error. Hi, this has nothing to do with webpack, it's just that there's no mechanism implemented to create custom property sets from js/config. That's what I assumed, thanks for the clarification. Since I'm using angular2, my use case isn't as simple as just plugging everything into one file. As a workaround, I'm using postcss-inject to inject :root into my components' css. @chimon2000: See https://github.com/postcss/postcss-custom-properties/issues/32#issuecomment-237363712 Refs #15
2025-04-01T06:39:58.135795
2024-07-12T08:13:55
2405014707
{ "authors": [ "ValterGames-Coder" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9512", "repo": "pastasfuture/com.hauntedpsx.render-pipelines.psx", "url": "https://github.com/pastasfuture/com.hauntedpsx.render-pipelines.psx/issues/44" }
gharchive/issue
Dont render camera in Build I see https://github.com/pastasfuture/com.hauntedpsx.render-pipelines.psx/issues/28, but i have a lastet realise in project I change Graphics API for Linux to Vulkan and errors lose, but i also have a gray screen My OS: Ubuntu 20.04 LTS Ohh, well, i changed unity version to 2021 LTS and this works! For me it's not problem what version i use, but someone it's can interfere
2025-04-01T06:39:58.164635
2024-07-24T17:52:29
2428159292
{ "authors": [ "mchitre", "patel999jay" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9513", "repo": "patel999jay/ProtocolDataUnits", "url": "https://github.com/patel999jay/ProtocolDataUnits/pull/2" }
gharchive/pull-request
Update to ProtocolDataUnits with Nested PDU Support and Improved API This pull request introduces significant enhancements to the ProtocolDataUnits Python module (as discussed in pull request #1 ). The following features have been added or improved: Nested PDU Support: PDUs can now contain other PDUs, allowing for more complex and hierarchical data structures. Enhanced API: Simplified and user-friendly API for creating PDU formats. New helper function create_pdu_format to streamline PDU creation. Updated Encoding/Decoding Functions: Improved handling of various data types. Support for nested PDUs in both encoding and decoding processes. Comprehensive Documentation: Detailed docstrings for the encode and decode functions to explain their functionality and usage. Updated README: Revised to reflect the new features and provide clear instructions for installation, usage, and contribution. Changes Made: Added nested_pdu method to the PDU class. Updated encode and decode methods to handle nested PDUs. Refactored the API for defining PDUs with create_pdu_format. Enhanced documentation and examples in the README. Testing: TODO: Extensive testing has to perform to ensure the new features work as expected. All existing tests have been updated to accommodate the new functionality. New tests have been added for nested PDU support. Impact: These changes improve the flexibility and usability of the ProtocolDataUnits module, making it easier to define and work with complex PDUs. Please review the changes and provide feedback. If everything looks good, kindly approve the pull request so we can merge it into the main branch. A bit busy these couple of weeks with field tests. Will review by early Aug. No rush. Good luck with the trials. @mchitre Thank you so much for reviewing, much appreciated. Sorry i have missed the notification for this for some reason. I do agree with the comments, will made the suggested changes and send for another review.
2025-04-01T06:39:58.253640
2024-07-23T14:42:41
2425398404
{ "authors": [ "lockwo", "zhengqigao" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9514", "repo": "patrick-kidger/diffrax", "url": "https://github.com/patrick-kidger/diffrax/issues/466" }
gharchive/issue
Accelerate ODE solver [What did I miss?] Hi, I am playing around with diffrax's ODE solving functionality. In a nutshell, I define a simple feedforward MLP with random initialization and benchmark the runtime of using it as the temporal derivatives of an ODE. I wrote the following code to record the run-time of ODE solving and got run-time around 3.7 sec, which seems much slower compared to other ODE solver frameworks. I am new to jax and diffrax. What did I miss in my code implemenation? import equinox as eqx import jax import diffrax import jax.numpy as jnp import time class MLPeqx(eqx.Module): layers: list activation: callable = eqx.static_field() def __init__(self, hidden_dims): super().__init__() tmp_key = jax.random.split(jax.random.PRNGKey(0), len(hidden_dims) - 1) self.layers = [eqx.nn.Linear(hidden_dims[i], hidden_dims[i + 1], key=tmp_key[i]) for i in range(len(hidden_dims) - 1)] self.activation = jax.nn.relu def __call__(self, x): for i in range(len(self.layers) - 1): x = self.activation(self.layers[i](x)) x = self.layers[-1](x) return x class ODEjax(eqx.Module): func: MLPeqx def __init__(self, hidden_dims): super().__init__() self.func = MLPeqx(hidden_dims) def __call__(self, t, y, args=None): return self.func(y) def solve_ode(input_x, t, func, cfg): sol = diffrax.diffeqsolve( diffrax.ODETerm(func), cfg['method'], t0=t[0], t1=t[-1], y0=input_x, dt0=None, saveat=diffrax.SaveAt(ts=t), stepsize_controller=diffrax.PIDController(atol=cfg['atol'], rtol=cfg['rtol']), ) return sol.ys def run_diffrax(hidden_dims, input_x, t, num_t, cfg): t = jnp.linspace(t[0], t[1], num_t) func = ODEjax(hidden_dims) y = jax.vmap(solve_ode, in_axes=(0, None, None, None))(input_x, t, func, cfg) return y if __name__ == '__main__': batch_size = 128 hidden_dims = [100, 100, 100] input_x = jax.random.normal(jax.random.PRNGKey(0), (128, 100)) start_time = time.time() run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}) end_time = time.time() print(f"run time = {end_time - start_time:.3f} (sec)") jax compile times will be longer for first iteration (and are generally excluded in benchmarks) with async dispatch you need a block until ready With the following code I got: import equinox as eqx import jax import diffrax import jax.numpy as jnp import time class MLPeqx(eqx.Module): layers: list activation: callable = eqx.static_field() def __init__(self, hidden_dims): super().__init__() tmp_key = jax.random.split(jax.random.PRNGKey(0), len(hidden_dims) - 1) self.layers = [eqx.nn.Linear(hidden_dims[i], hidden_dims[i + 1], key=tmp_key[i]) for i in range(len(hidden_dims) - 1)] self.activation = jax.nn.relu def __call__(self, x): for i in range(len(self.layers) - 1): x = self.activation(self.layers[i](x)) x = self.layers[-1](x) return x class ODEjax(eqx.Module): func: MLPeqx def __init__(self, hidden_dims): super().__init__() self.func = MLPeqx(hidden_dims) def __call__(self, t, y, args=None): return self.func(y) def solve_ode(input_x, t, func, cfg): sol = diffrax.diffeqsolve( diffrax.ODETerm(func), cfg['method'], t0=t[0], t1=t[-1], y0=input_x, dt0=None, saveat=diffrax.SaveAt(ts=t), stepsize_controller=diffrax.PIDController(atol=cfg['atol'], rtol=cfg['rtol']), ) return sol.ys @eqx.filter_jit def run_diffrax(hidden_dims, input_x, t, num_t, cfg): t = jnp.linspace(t[0], t[1], num_t) func = ODEjax(hidden_dims) y = jax.vmap(solve_ode, in_axes=(0, None, None, None))(input_x, t, func, cfg) return y batch_size = 128 hidden_dims = [100, 100, 100] input_x = jax.random.normal(jax.random.PRNGKey(0), (128, 100)) _ = run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}).block_until_ready() %%timeit _ = run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}).block_until_ready() 19.4 ms ± 4.31 ms per loop (mean ± std. dev. of 7 runs, 10 loops each) Thanks so much! I have tried on my end and observed similar run-time metrics. I have another follow-up question. Say I first want to run with atol=rtol=1e-5, and later in my code I want it to run with atol=rtol=1e-4. I observe again the method run_diffrax runs slower again when changing from 1e-5 to 1e-4 because of compilation( I guess). Namely, # first time of atol=rtol=1e-5, takes ~2secs run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}).block_until_ready() # second time of atol=rtol=1e-5, takes ~0.008secs run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}).block_until_ready() # first time of atol=rtol=1e-4, takes ~2secs run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}).block_until_ready() # second time of atol=rtol=1e-4, takes ~0.008secs run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': 1e-5, 'rtol': 1e-5}).block_until_ready( Is this behavior expected? I wonder if there is a way to compile only once for arbitrary atol=rtol values, and can always run around millisecond level regardless of atol and rtol. Yes, this behavior is expected. The python floats are getting marked as static by the filtering that happens before jit. You can make them not static by making them jax types (e.g. arrays). start_time = time.time() run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': jnp.array(1e-5), 'rtol': jnp.array(1e-5)}).block_until_ready() end_time = time.time() print(f"run time = {end_time - start_time:.3f} (sec)") start_time = time.time() run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': jnp.array(1e-5), 'rtol': jnp.array(1e-5)}).block_until_ready() end_time = time.time() print(f"run time = {end_time - start_time:.3f} (sec)") start_time = time.time() run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': jnp.array(1e-4), 'rtol': jnp.array(1e-5)}).block_until_ready() end_time = time.time() print(f"run time = {end_time - start_time:.3f} (sec)") start_time = time.time() run_diffrax(hidden_dims, input_x, [0.0, 1.0], 100, { 'method': diffrax.Dopri5(), 'atol': jnp.array(1e-4), 'rtol': jnp.array(1e-5)}).block_until_ready() end_time = time.time() print(f"run time = {end_time - start_time:.3f} (sec)") run time = 4.057 (sec) run time = 0.016 (sec) run time = 0.013 (sec) run time = 0.013 (sec)
2025-04-01T06:39:58.259677
2023-12-15T13:55:33
2043772276
{ "authors": [ "darkkatarsis", "zweckj" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9515", "repo": "patrickhilker/tedee_hass_integration", "url": "https://github.com/patrickhilker/tedee_hass_integration/issues/51" }
gharchive/issue
State does not change the icon I noticed that the entity that is from integration does not update the icon when changing the status. It seems to me that this is due to the fact that the device does not have the right device_class or the supported_features are different. I am using beta version 2023.11.0b5 Probably just because the icon is set statically. Will change it in the next update, but it should be fine in tile or mushroom cards anyways. shouldn't happen on the non-beta versions, the beta branch will not be taken further, as that part will move to HA core.
2025-04-01T06:39:58.272497
2023-09-24T09:25:25
1910167420
{ "authors": [ "DarKDinDoN" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9516", "repo": "patrickporto/journal-whiteboard", "url": "https://github.com/patrickporto/journal-whiteboard/issues/9" }
gharchive/issue
GMScreen compatibily is broken on Chronium based browser Describe the bug Hello, thanks for the last issue resolved! When editing a whiteboard and saving it, GM Screen preview is not updated until I hit the GM Screen refresh button. To Reproduce Steps to reproduce the behavior: Go to GM Screen (with a Whiteboard page attached) Click on Edit the journal note and edit the whiteboard then save GM Screen doesn't show changes Hit the refresh button of GM Screen, changes are shown within GM Screen Expected behavior GM Screen should show changes once they are saved (might be an event relative to note updated not fired?) Environment (please complete the following information): Foundry VTT: 11.309 Browser tested on chrome & brave Module Version latest Thx! Working perfectly 😃
2025-04-01T06:39:58.276943
2023-11-18T00:42:05
2000122622
{ "authors": [ "patrix87", "wizzbangwa" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9517", "repo": "patrix87/PowerShellGSM", "url": "https://github.com/patrix87/PowerShellGSM/issues/33" }
gharchive/issue
Icarus Server Not Listening on Port 17777 Followed your instructions. Server is finally running after suffering through figuring out how to get rid of the Visual C++ error message. I am on a fresh Windows 11 install. The Icarus server is not visible in the "Dedicated Servers" list, and is not available by direct connect either. netstat -an | Select-String 27015 shows the server machine is listening on the port. netstat -an | Select-String 17777 doesn't return anything, showing the server is not listening on this port. icarus.psm1 is set to both ports above. I have opened ports 27015 and 17777 TCP and UDP on both routers (I have two separate networks). Windows firewall has all ports opened for Icarus for UDP and TCP for private, public, and domain networks. The server's console window is open and showing the server is running, but no confirmation the server is listening on what ports. The server does not show in the "Dedicated Servers" list. The server is not available through direct connect in Icarus. Don't know what to do from here. Can you post your configuration and logs ? I managed to get it running a few minutes ago. Shortly after my original post I switched to my own startup cmd which got everything running. This time around I tried the automated script by @BananaAcid. Now the server is running just fine. Now I need to configure daily or hourly automated backups, if that's possible. So far I have to manually restart the launcher script, not quite helpful if someone wipes out the prospect world and I haven't restarted recently. Happened once already. Automated Backups and restart are built-in powershell GSM, they are also enabled by default in the configuration file. They are launched by the task-scheduler in windows.
2025-04-01T06:39:58.285557
2019-03-24T10:53:35
424599413
{ "authors": [ "boaz0", "patternfly-build" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9518", "repo": "patternfly/patternfly-next", "url": "https://github.com/patternfly/patternfly-next/issues/1618" }
gharchive/issue
form control should have ellipsis when text overflows According to the image below when the text overflows, ellipsis should be shown: I believe it's easier to implement that using css's text-overflow than writing it in JavaScript. :tada: This issue has been resolved in version 2.0.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
2025-04-01T06:39:58.291843
2020-03-10T21:59:03
578885408
{ "authors": [ "mcarrano", "mcoker", "mturley", "patternfly-build", "redallen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9519", "repo": "patternfly/patternfly-next", "url": "https://github.com/patternfly/patternfly-next/pull/2807" }
gharchive/pull-request
fix(file-upload): removed message container, added form to error example fixes https://github.com/patternfly/patternfly-next/issues/2797 Preview: https://patternfly-next-pr-2807.surge.sh @mcarrano it just removes the file message/helper text at the bottom (instructions/error text) from the file upload component and relies on the file upload component being in a form for that helper text, so the helper text comes from the form component instead. That also means: Any updates we make to the form component's helper text (descriptive, error, success, etc) will be available when used with file upload since the file upload is no longer creating its own helper text. The text doesn't get a blue overlay when you drag/hover the file upload component, since that text is no longer part of the file upload component. old: new: Sounds good to me. Thanks for the explanation @mcoker ! Ah, I'm glad you guys made that change, I forgot to mention I needed to make .pf-c-file-upload a div instead of a form in the React component so I could get my FormGroup example to work (React will refuse to let you nest a in another ). :tada: This PR is included in version 2.68.4 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
2025-04-01T06:39:58.295088
2018-08-28T21:57:38
354919420
{ "authors": [ "mattnolting", "michael-coker", "patternfly-build" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9520", "repo": "patternfly/patternfly-next", "url": "https://github.com/patternfly/patternfly-next/pull/658" }
gharchive/pull-request
Extract components from page layout fixes https://github.com/patternfly/patternfly-next/issues/637 @andresgalante I left the JS that toggles the navigation in the demo. It's easy to remove if we want to. For the sake of consistency and for future demos, I don't think we should write javascript in our workspace, and just present the different states instead. To see the interaction/animations, maybe link to a demo of the react components in a sample application since that's officially supported. What do you think? Deploy preview for pf-next ready! Built with commit 994767c170f99c03907627550aec2d9db70ec898 https://deploy-preview-658--pf-next.netlify.com @michael-coker @andresgalante I agree that best practice is to present different states. With more complex components, it's just a lot more work to do so when including very basic JS makes it much easier to understand. It doesn't seem to be a big issue while we're still in Alpha, but will be in Beta+. Insights will be using all interactive components, so that would be a good opportunity to link to a live demo when they become available.
2025-04-01T06:39:58.377492
2021-09-22T17:07:01
1004527348
{ "authors": [ "StepS-", "pattlebass" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9521", "repo": "pattlebass/Music-DJ", "url": "https://github.com/pattlebass/Music-DJ/issues/4" }
gharchive/issue
[Feature]: Import original MusicDJ MIDI files Describe the problem you're having I've got a bunch of MusicDJ MIDI files made on SE K750i back in 2006, however nothing can open or convert them. Edit: VLC with fluidsynth can in fact open them (thanks to some people for the tips). MusicDJ data appears to be stored in the SEM1 chunk as noted in losnoco/foo_midi#1 and confuses most MIDI decoders. MusicDJ_files.zip Describe the solution you'd like I was wondering if it would eventually be possible to import the files as projects in this new MusicDJ recreation. However, since I imagine this use case is super rare I'm just leaving it as a "nice to have" in the hopes someone would be compelled enough to work on this. Thank you! @StepS- Good news! v1.5-beta just added this.
2025-04-01T06:39:58.396837
2024-07-15T21:44:06
2409696311
{ "authors": [ "WilliamAGH", "berwinsingh", "bitnom", "flamerged", "jackyliang", "pascalandy", "paul-gauthier" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9522", "repo": "paul-gauthier/aider", "url": "https://github.com/paul-gauthier/aider/issues/867" }
gharchive/issue
Add support for 8k tokens with Claude 3.5 Sonnet Issue Claude 3.5 Sonnet now supports a maximum output/response token length of 8,192 tokens (from 4,096 currently). To support 8k tokens, we need to add the header "anthropic-beta": "max-tokens-3-5-sonnet-2024-07-15" to our API calls: Issue Claude 3.5 Sonnet now supports a maximum output/response token length of 8,192 tokens (from 4,096 currently). To support 8k tokens, we need to add the header "anthropic-beta": "max-tokens-3-5-sonnet-2024-07-15" to our API calls: How did you get past this: official docs about the feat: https://docs.anthropic.com/en/docs/about-claude/models#model-comparison did anybody submit a PR yet? Thanks for trying aider and filing this issue. Aider is already able to receive unlimited output from Sonnet. See the article below. This capability was introduced back in v0.41.0. https://aider.chat/2024/07/01/sonnet-not-lazy.html That said, it would be beneficial to support the new 8k output limit, mainly to somewhat reduce the cost and latency of longer LLM responses. I added support for 8k output tokens. The change is available in the main branch. You can get it by installing the latest version from github: python -m pip install --upgrade git+https://github.com/paul-gauthier/aider.git If you have a chance to try it, let me know if it works better for you. I'm going to close this issue for now, but feel free to add a comment here and I will re-open or file a new issue any time. I am trying to do this on the AWS Bedrock hosted version of 3.5 Sonnet. However, to my surprise I found that it isn't really updated all that much meaning if I pass this parameter I get an error of extra headers not allowed. Then doing some searching I found that Sonnet on bedrock hasn't been updated since it was launched i.e., June 2024 which is really surprising. Any idea if Amazon or Anthropic will update the models? Is this on by default, or do I need to point Cursor to use a specific model/API? Aider is using 8k output tokens for sonnet: aider --sonnet --verbose ... Model metadata: { "key": "claude-3-5-sonnet-20241022", "max_tokens": 8192, ...
2025-04-01T06:39:58.409332
2015-09-05T16:23:02
105035609
{ "authors": [ "SeerUK", "TEWatson", "Tolsi", "Woodz", "agile-jordi", "barkhorn", "eyalroth", "mdirkse" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9523", "repo": "paulbutcher/ScalaMock", "url": "https://github.com/paulbutcher/ScalaMock/issues/117" }
gharchive/issue
"Null expectation context - missing withExpectations?" w/ ScalaTest FunSpec Hi, I'm currently trying to create a basic test suite using ScalaTest and wanted to try my hand at mocking too (I'm quite new to Scala so am just wanting to get used to a few things and pin down which libraries I can use to perform common tasks like testing, and mocking etc). I'm encountering a strange exception when I run my tests, one test succeeds, then all of the following tests fail. I've tried to change how this works a few times to restructured when things are created, and if they're only created once, but I still just get exceptions. Here's my tests: Here's the output: I've just encountered this as well with FunSuite. It seems to happen when you mock stuff in before/beforeEach. A temporary workaround is to define whatever in before in a method and calling it at the beginning of each test (in other words, behave like there's no "before" feature). +1 +1 +1 minimal failing example: class CatTest extends FunSuite with BeforeAndAfter with MockFactory { var cat: Cat = _ trait Cat { def meow(): String } before { cat = stub[Cat] (cat.meow _).when().returns("meow") } test("meows") { cat.meow() === "meow" } test("meows again") { cat.meow() === "meow" } } working example: class CatTest extends FunSuite with BeforeAndAfter with MockFactory { var cat: Cat = _ trait Cat { def meow(): String } before { withExpectations { cat = stub[Cat] (cat.meow _).when().returns("meow") } } test("meows") { cat.meow() === "meow" } test("meows again") { cat.meow() === "meow" } } The example above only works because there is no assertion on the ===. This code with added asserts looks like this: class CatTest extends FunSuite with BeforeAndAfter with MockFactory { var cat: Cat = _ trait Cat { def meow(): String } before { withExpectations { cat = stub[Cat] (cat.meow _).when().returns("meow") } } test("meows") { assert(cat.meow() === "meow") } test("meows again") { assert(cat.meow() === "meow") } } Which fails the assertions with cat.meow() being equal to null. The problem seems to occur whenever the mock is done within before{}. The only way I have been able to get around this is using OneInstancePerTest instead of BeforeAndAfter and implementing similarly. @barkhorn can you confirm why this bug was closed? It seems to be confirmed but not fixed? Hi, I can't see where the bug is here, so this should remain closed. To share mocks within a suite, i'd recommend reading up on this page: http://scalamock.org/user-guide/sharing-scalatest/ Using the OneInstancePerTest is the correct way of sharing mocks (or a Fixture, whatever style you like better). You could kind of use withExpectations too, but not in a before call. A higher order function wrapping around it might work for you. See here for the implementation, maybe that helps. I would suggest using a Fixture though. https://github.com/paulbutcher/ScalaMock/blob/5fd4e563143d752b9638d32e05defcd8ddb8bb82/shared/src/main/scala/org/scalamock/MockFactoryBase.scala#L47
2025-04-01T06:39:58.415610
2019-10-30T13:01:26
514651042
{ "authors": [ "borisllona", "bus710", "pauldemarco", "rjcasson" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9524", "repo": "pauldemarco/flutter_blue", "url": "https://github.com/pauldemarco/flutter_blue/issues/418" }
gharchive/issue
SOLVED - Error importing androidx.core in FlutterBluePlugin.java :\Users\boris\AppData\Roaming\Pub\Cache\hosted\pub.dartlang.org\flutter_blue-0.6.3+1\android\src\main\java\com\pauldemarco\flutter_blue\FlutterBluePlugin.java:43: error: package androidx.core.app does not exist import androidx.core.app.ActivityCompat; ^ C:\Users\boris\AppData\Roaming\Pub\Cache\hosted\pub.dartlang.org\flutter_blue-0.6.3+1\android\src\main\java\com\pauldemarco\flutter_blue\FlutterBluePlugin.java:44: error: package androidx.core.content does not exist import androidx.core.content.ContextCompat; I have this error and I've already imported android x and refractored. What should I do? I've tested with diferent versions of flutter ble too. @borisllona, check out my comment here: https://github.com/pauldemarco/flutter_blue/issues/402#issuecomment-546425715 Or my pull request here: #416 Also, staying at 0.6.2 is not a bad idea (in pubspeck.yaml). I had to delete the .flutter0plugins and pubspec.lock before flutter pub get. It works for me but I am looking forward the next version if it comes with the solution of @rjcasson @borisllona, check out my comment here: #402 (comment) Or my pull request here: #416 It worked ! I was starting going crazy and updating my own app gradle but I haven't realized that it has to be changed the library one. Thanks so much! @borisllona Chances are the problem is with outdated project files. Stop into the discord if you'd like some help getting things up and running. Please follow the instructions here and let me know the results: https://github.com/pauldemarco/flutter_blue/issues/415#issuecomment-548186492
2025-04-01T06:39:58.798125
2023-06-12T16:09:01
1753089839
{ "authors": [ "OpenBagTwo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9525", "repo": "paulevsGitch/EdenRing", "url": "https://github.com/paulevsGitch/EdenRing/pull/66" }
gharchive/pull-request
Port to 1.20 What's Working The mod builds, and Minecraft loads with it and BCLib 3.0.2 I can create a world without crashing I can construct an Eden Portal I can travel to the Eden Ring I saw a discwing (didn't grab a screenshot :disappointed:) Eden ring items show up in the creative inventory What's Not There's no Eden Ring tab The Eden Ring guidebook is barely legible What I haven't tested Crafting Mining / block breaking (see below) Multiplayer Anything having to do with gravity manipulation What to Pay Extra Attention to The most non-obvious changes (read: the ones I probably got wrong) were: rewriting the guide book rendering (and I clearly messed that up pretty badly, haha!) migrating to the tag-based system of determining which tools break which blocks, explicitly the portion of the program that determines what items are mineable by what. I followed BetterEnd's example, but I was stumbling around in the dark and have no idea if I did it right, or whether I was even matching intent when it came to hammer-compatibility. Thank you so much for creating this gorgeous mod--there's nothing like spending a day and a half poring over someone's codebase to make you appreciate the impressiveness of their body of work. If there's anything I can do to assist in the future development of this mod (including sponsoring its development via KoFi, Patreon, etc.), please let me know. Also thank you to @Treetrain1 (whose review, testing and feedback I would greatly appreciate!) for all the work on getting the mod to 1.19.3 (especially the datagen gradle task!) Ha! Okay, update: placing a gravity compressor crashed the game: java.lang.IllegalArgumentException: Cannot get property class_2746{name=extended, clazz=class java.lang.Boolean, values=[true, false]} as it does not exist in Block{edenring:gravity_compressor} I'm guessing there's a problem with the piston mixin... Anyway, this is why this PR is likely to need a bit more attention than, say, #61 I'm closing this PR, as any port to 1.19.4+ really should be using the new tags system. I may try again, starting from @Treetrain1's 1.19.3 port, but I don't think I'll be reusing much of my own work from this PR.
2025-04-01T06:39:58.867141
2015-10-06T21:37:59
110106661
{ "authors": [ "binarykitchen", "es128", "paulmillr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9526", "repo": "paulmillr/chokidar", "url": "https://github.com/paulmillr/chokidar/issues/363" }
gharchive/issue
Why is fsevents mentioned twice in package.json? This as a dependency and as an optional dependency both? Furthermore I am unable to use npm shrinkwrap on my Ubuntu machine anymore because fsevents is targeted for OS machines only. It is breaking our build in our organization :( Looking at your issues, I think fsevents has caused lots of trouble. Maybe time to drop it? If not, how else can I solve the shrinkwrap issue then? fsevents is an optional dependency: https://github.com/paulmillr/chokidar/blob/master/package.json#L41 If you can't do shrinkwrap, that seems like a npm issue, not brunches. NPM should always respect optionalDependencies. Please report it here: https://github.com/npm/npm/issues https://github.com/npm/npm/issues/9865
2025-04-01T06:39:58.870486
2019-10-09T19:25:39
504845756
{ "authors": [ "XhmikosR", "exx8", "paulmillr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9527", "repo": "paulmillr/chokidar", "url": "https://github.com/paulmillr/chokidar/pull/897" }
gharchive/pull-request
allow watch for recursive not existing folder Solves #872 This pull request allows user to watch specific not existing folders. Was achieved via introducing a new function _mostSpecifiedExistingAncestor We should consider making the function public, as users might be interested in this functionality. We need tests for that feature. Also, ensure you conform to chokidar code style e.g. { locations. I added tests. I've made the changes that you asked. Ping? If i’m reading this correctly, we’ll have a huge performance degradation because of the new IO call for every file. Need to investigate this thoroughly and so some benchmarks. If i’m reading this correctly, we’ll have a huge performance degradation because of the new IO call for every file. Need to investigate this thoroughly and so some benchmarks. maybe we should use a flag to enable or disable this functionality? This needs a rebase. Also the curly brackets shouldn't be placed on new lines, there's an extra space in a test message and extra newlines.
2025-04-01T06:39:58.878506
2015-12-12T20:56:02
121879600
{ "authors": [ "Xotic750", "cscott", "ljharb" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9528", "repo": "paulmillr/es6-shim", "url": "https://github.com/paulmillr/es6-shim/issues/382" }
gharchive/issue
Question: When is Set/Map not? I am writing some type checking routines isSet and isMap among others. I have them using Object#toString when toStringTag is not available, .and when it is available then I'm using the size getter to perform a check. This is working nicely across realm and on all environments that I have tested on. I then threw es6-shim into the mix, which has now caused me some dilemma. On ES5 environments the shim is used, it walks, talks, eats, sleeps, and even farts like a Duck :) but it's DNA is not of a Duck, Object#toString is [object Object]. Ok, so the alternative method of using size says that it is a Duck, and I thought "that's ok then". But using this method without es6-shim then fails due to bugs in old FireFox where size is a function rather than a getter. So I'm wondering what thoughts are regarding type checkers and shims which implement a Duck in all but DNA. Should these checkers support es shims or should they count a shim as not a real Duck and say false. I'm not asking for code, just a perspective from shim makers as to whether isSet should return true for the Set shim in an ES5 environment or even a Set patch in an ES6 environment where the constructor has been replaced? You can go one of two ways - either isSet only returns true for a set that has native behavior (which is impractical to determine in JS), or, assume that if Set exists, cache the Set.prototype.size getter and use that (an old Firefox Set is not actually a Set if "size" is a non-getter). Thanks for that perspective, my head is still in dilemma about what to do about it, how much to check (is a buggy Set not a Set?). Seems likely that there is a need to drop Object#toString checking for Map and Set and be more duck-typed. Maybe sleeping on it will clear my view. Yes, Object#toString is not reliable whatsoever in an environment with Symbol.toStringTag, so it's best to avoid it except as an optimization in older engines. Yeah, I was only using it where toStringTag didn't exist, but it didn't exist in the buggy Firefox. So this begs the question of how much of an implementation to test before you say "this is a Set/Map", sure I could require that es6-shim is loaded, but unlike es5-shim (which I fully recommend/require) these shims still feel like they are in their infancy, and some even require true ES5 environments. It feels a bit of a "between a rock and a hard place" situation. es6-shim should still be used on every engine since there are tons of bugs in all of them with their implementations. Based on what you've said, I am going back over all my work to add it in. It's going to throw up some issues I'm sure. :) The only really reliable mechanism is to compile a list of bugs on each platform, and test for them individually. That happens to be exactly what es6-shim is doing -- it then loads a shim to fix those bugs when it detects them. But there are certainly ES6 features which are unshimmable. Our goal is that correctly written ES6 code which uses only the shimmable subset of ES6 will run correctly on every platform. If you use unshimmable features we can't make any guarantees. (ToStringTag is an unshimmable feature.) I'm currently moving away from toStringTag checks in all of my code, and adding es6-shim to all projects. Often it means implementing more costly code, but nothing that I'm working on needs to be so performant. :) Closing this, I'm very happy that the shim implements Map and Set as per spec, and that my isMap and isSet are happy. I look forward to a new release that includes the MapIterator and SetIterator changes that are sitting in the master. Thanks.
2025-04-01T06:39:58.880909
2022-03-21T14:28:50
1175470808
{ "authors": [ "jasonbrudvik" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9529", "repo": "paulscherrerinstitute/tomorec", "url": "https://github.com/paulscherrerinstitute/tomorec/pull/3" }
gharchive/pull-request
Added docker image creation I added dockerfiles and instructions necessary to build a docker image containing: JuptyerLab Tomorec Kernel In the build of the docker image, I use as input the current user's uid. I did this in order to simplify the permissions issues that occur when writing to the host file system from within a docker container. Not sure if this is the best solution, as it make the image specific to that user, but we can change this later and use a different method if needed. Please test it out and see if it works for you! I tested it myself and it works, but it really pushes my laptop to the limit in terms of resource usage. It might be necessary to modify the analysis script further to avoid this. #! Jason I have updated the code, will make a new pull request
2025-04-01T06:39:58.938265
2022-05-16T09:04:09
1236845044
{ "authors": [ "emilleszczak2", "m-pastuszek" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9531", "repo": "pay-now/paynow-php-sdk", "url": "https://github.com/pay-now/paynow-php-sdk/issues/65" }
gharchive/issue
Laravel 9 - empty array on getPaymentMethods Hello, I'm integrating Paynow with Laravel 9 project / PHP 7.4. public function availablePaymentMethods() { $client = new Client(config('paynow.api_key'), config('paynow.signature_key'), Environment::SANDBOX); try { $payment = new Payment($client); $paymentMethods = $payment->getPaymentMethods(); $availablePaymentMethods = $paymentMethods->getAll(); return response()->json($availablePaymentMethods); } catch (PaynowException $exception) { return response()->json('Error occurred', 400); } } API returns 200, but it's content is only empty elements: [ {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {}, {} ] Http Client: httpClient: Paynow\HttpClient\HttpClient {#499 ▼ #client: Symfony\Component\HttpClient\Psr18Client {#503 ▼ -client: Symfony\Component\HttpClient\CurlHttpClient {#506 ▶} -responseFactory: Nyholm\Psr7\Factory\Psr17Factory {#504} -streamFactory: Nyholm\Psr7\Factory\Psr17Factory {#505} } Have you tried to call the REST API from the postman or curl for your credentials? @emilleszczak2, calling API directly from postman returns data correctly. I've made it this way: public function availablePaymentMethods(): \Illuminate\Http\JsonResponse { $client = new Client(config('paynow.api_key'), config('paynow.signature_key'), Environment::SANDBOX); try { $payment = new Payment($client); $paymentMethods = $payment->getPaymentMethods(); $availablePaymentMethods = $paymentMethods->getOnlyCards(); $responseTable = array(); foreach ($availablePaymentMethods as $availablePaymentMethod) { $responseTable[] = [ 'id' => $availablePaymentMethod->getId(), 'type' => $availablePaymentMethod->getType(), 'name' => $availablePaymentMethod->getName(), 'description' => $availablePaymentMethod->getDescription(), 'image' => $availablePaymentMethod->getImage(), 'status' => $availablePaymentMethod->getStatus(), 'authorizationType' => $availablePaymentMethod->getAuthorizationType() ]; } return response()->json($responseTable); } catch (PaynowException $exception) { return response()->json('Error occurred: '. $exception->getMessage(), 400); } } We can close the issue.
2025-04-01T06:39:58.962898
2023-01-28T15:12:18
1560891839
{ "authors": [ "Kikky", "jmikrut", "joornby" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9532", "repo": "payloadcms/payload", "url": "https://github.com/payloadcms/payload/issues/1963" }
gharchive/issue
Some entries disappear from collection page view table Hello! Bug Report After I've added draft system to this collection, I've got a problem with view table of the collection: 12 entries from 14 don't appear in table: But the view becomes ok if I sort with "-" by some field: So this /pages?sort=-_status&page=1 will be ok and this pages?sort=_status&page=1 will lose 12 entries. Other Details Payload version: 1.5.9 For what it's worth, I was experiencing this issue and had to go all the way back to version 1.5.4 before it worked as expected. There were several version in between 1.5.4 and 1.5.9 that did slightly different things. Hey @Kikky and @joornby — this is now resolved and released in 1.6.1. Note that there are some migration steps that you need to follow in order to get to the new version. See here: https://github.com/payloadcms/payload/releases/tag/v1.6.1 @jmikrut , I've updated to 1.6.3. After created new collection with such versions options, because I just need draft system: And when I create new entry to this collection, edit it and click Publish - I get 1 version is found. If I edit it and click Publish again I will get - no versions found. One more time - 1 version is found. One more - no versions found ... etc. I don't need versions, I just need draft system, but when a doc has no versions in DB when versions are enabled on collection I can't view them in collection's view table.
2025-04-01T06:39:58.966965
2024-07-16T10:06:46
2410751966
{ "authors": [ "chrisvanmook", "paulpopus" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9533", "repo": "payloadcms/payload", "url": "https://github.com/payloadcms/payload/issues/7163" }
gharchive/issue
Switching locale doesn't update field UI based on certain access Link to reproduction https://github.com/chrisvanmook/payload/commit/c0be6a12eb7bc17b906f88832253e9d28c0a1fe6 Payload Version 3.0.0-beta.60 Node Version v20.11.1 Next.js Version 15.0.0-canary.58 Describe the Bug When switching locale, and you have an access method (update) that should disable the field based on a language, it doesn't update the UI. Only after a full refresh the expected result is visible (see video). https://github.com/user-attachments/assets/ed029167-fc44-4016-933e-e93af0486475 Reproduction Steps Add code in reproduction url Switch from english to NL Note that the field is not disabled Refresh Note that the field is now disabled as it should Adapters and Plugins db-postgres This will be fixed in the next beta release!
2025-04-01T06:39:59.002007
2023-02-02T13:51:59
1568101243
{ "authors": [ "Oooooori", "saperi22", "tdchow" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9534", "repo": "paypal/android-checkout-sdk", "url": "https://github.com/paypal/android-checkout-sdk/issues/195" }
gharchive/issue
How to clear previous user paypal data? @Oooooori apologies for the delayed reply. The logout functionality is slated to be available in our next release. Closing this issue as it's a duplicate of https://github.com/paypal/android-checkout-sdk/issues/99 The changes are already merged for programmatic logout. It will be in the next release.
2025-04-01T06:39:59.006913
2017-03-13T10:25:25
213725609
{ "authors": [ "bluepnume", "danigomez", "yuuk" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9535", "repo": "paypal/paypal-checkout", "url": "https://github.com/paypal/paypal-checkout/issues/257" }
gharchive/issue
can not detect paypal dialog closed event. I can not find a close event callback when I manually close paypal dialog. This is possibly related to #250 and #211 Are you using onCancel with the iframe modal? In that case, your issue is probably due to be released with #211 I've used, but no effect. I don't know if i'm wrong. My situation is like this: When I click the checkout button, our website will generate an order by showing a loading page and the loading page is also in paypal dialog. And now when I close the popup window I want to have a close callback to do some operations just like refreshing my checkout page. This is our loading page: After our loading page, there will show paypal's payment page When I close the window of paypal's payment page, there will have a default event to jump to another page, But when I close the window in our loading page, there will nothing happened. Looks like you're using the legacy v3.5 integration. If you want to use onCancel you will need to upgrade to the v4 button -- please see the docs at https://developer.paypal.com/docs/integration/direct/express-checkout/integration-jsv4/advanced-integration/ thanks so much! I'll try it now.
2025-04-01T06:39:59.012507
2017-01-30T12:39:47
203999821
{ "authors": [ "akara", "thefourtheye" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9536", "repo": "paypal/squbs", "url": "https://github.com/paypal/squbs/pull/382" }
gharchive/pull-request
replace map with foreach Thanks for your pull request. Please review the following guidelines. [ ] Title includes issue id. [x] Description of the change added. [x] Commits are squashed. [ ] Tests added. [ ] Documentation added/updated. [x] Also please review CONTRIBUTING.md. The change looks good. Just need help with some of the formalities: Please file an issue for any change Please edit your commit to include link or reference to the issue The PR title should reference the issue (in form of #123) Sorry for the inconvenience. We do this for traceability (as described in CONTRIBUTING.md). Thanks! @akara Done! PTAL. Thanks! Build is failing. Let me go take a look what's happening there. This should not be due to this change unless there is anything unexpected. I'll get back. No action on your side.
2025-04-01T06:39:59.036771
2016-09-21T18:17:12
178423497
{ "authors": [ "azn1viet" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9541", "repo": "pbakondy/cordova-plugin-sim", "url": "https://github.com/pbakondy/cordova-plugin-sim/issues/32" }
gharchive/issue
iOS 10, xCode 8: Error in Success callbackId Expected behaviour Tell us what should happen. Actual behaviour Tell us what happens instead. Provide a log message if relevant Error in Success callbackId: Sim1675580494 : TypeError: undefined is not an object (evaluating 'navigator.connection.type') I'm seeing this behaviour on Remove this hint: these checkboxes can be checked like this: [x] [x ] iOS device [ x] iOS simulator [ ] Android device [ ] Android emulator I am using [x ] cordova [ ] ionic [ ] PhoneGap [ ] PhoneGap Developer App [ ] Intel XDK [ ] Intel App Preview [ ] Telerik [ ] Other: Hardware models Example: Samsung Galaxy S6, iPhone 6s OS versions Example: Android 4.4.2, iOS 9.2 I've checked these [ ] It happens on a fresh Cordova CLI project as well. [ ] I'm waiting for deviceready to fire. [ ] My JavaScript has no errors (window.onerror catches nothing). [ ] I'm using the latest cordova library, Android SDK, Xcode, etc. So how can we reproduce this? Provide the used components versions (cordova, ionic, etc). Provide the steps to reproduce the issue. Provide files, sources if available. Thanks. Somehow my "cordova-plugin-network-information" got removed. I just added it again and there would not be an issue. Sorry!
2025-04-01T06:39:59.082593
2015-12-05T22:30:59
120592515
{ "authors": [ "pburtchaell", "tomatau" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9542", "repo": "pburtchaell/redux-promise-middleware", "url": "https://github.com/pburtchaell/redux-promise-middleware/pull/41" }
gharchive/pull-request
Tests for everything This tests all the features we have right now that I could think of. It also makes a fix for when the promise returns null, as I re-broke that in a previous commit. So if this gets merged, can you republish a patch version. This is great. :+1: I released the patch as 2.3.1. Doesn't seem to be on npm yet Whoops. Sorry abut that. It should be now. Just FYI, I ignored the tests and example directory on npm. I’m also in the process of setting up code coverage + https://coveralls.io https://coveralls.io/. I’ll probably finish that up in a couple of weeks when I have more time. On Dec 6, 2015, at 3:01 PM, Thomas<EMAIL_ADDRESS><EMAIL_ADDRESS>wrote: Doesn't seem to be on npm yet — Reply to this email directly or view it on GitHub https://github.com/pburtchaell/redux-promise-middleware/pull/41#issuecomment-162347079.
2025-04-01T06:39:59.087866
2019-09-05T21:43:29
490024260
{ "authors": [ "codecov-io", "pbylicki" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9543", "repo": "pbylicki/rfhub-new", "url": "https://github.com/pbylicki/rfhub-new/pull/41" }
gharchive/pull-request
Add simple implementation of infinite list Fixes #29 for all list views in the application Codecov Report Merging #41 into master will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #41 +/- ## ====================================== Coverage 94.2% 94.2% ====================================== Files 33 33 Lines 569 569 ====================================== Hits 536 536 Misses 33 33 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update b7cf394...b6ab447. Read the comment docs.
2025-04-01T06:39:59.184783
2021-01-13T04:55:04
784787065
{ "authors": [ "HosokawaR", "pcottle" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9544", "repo": "pcottle/learnGitBranching", "url": "https://github.com/pcottle/learnGitBranching/pull/779" }
gharchive/pull-request
add a japanese translate on lockedMaster Thanks for the so great app !! I love this app. I added a japanese translate on lockedMaster. Sorry about the messy commits history, becouse I screwed up on the line feed code. TEST I passed the test on local. Executed 183 of 183 specs SUCCESS in 1 sec. Thank. 🎉 Thanks a bunch for adding a level translation!! 🎉 Thanks a bunch for adding a level translation!!
2025-04-01T06:39:59.202642
2014-12-20T00:57:53
52545345
{ "authors": [ "Tonybyte", "bgod" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9545", "repo": "pdaddyo/soundbounce", "url": "https://github.com/pdaddyo/soundbounce/issues/88" }
gharchive/issue
cant join rooms Noooo! I'm home and connect through my sprint phone hotspot. I can't get into rooms. Turned off firewall. Your websocket link gives all "yes", looks fine. The log: 2014-12-19 19:58:30,613 DEBUG[SoundBounce.WindowsClient.Program]: Initial startup: Main() 2014-12-19 19:58:30,694 INFO [SoundBounce.WindowsClient.SpotifyEnabledBrowser]: Cef version initialized OK. 2014-12-19 19:58:30,701 DEBUG[SoundBounce.SpotifyAPI.Spotify]: Message thread running... 2014-12-19 19:58:36,814 DEBUG[SoundBounce.SpotifyAPI.Spotify]: api_version=12 2014-12-19 19:58:36,814 DEBUG[SoundBounce.SpotifyAPI.Spotify]: api_version=12 2014-12-19 19:58:36,814 DEBUG[SoundBounce.SpotifyAPI.Spotify]: application_key_size=321 2014-12-19 19:58:36,814 DEBUG[SoundBounce.SpotifyAPI.Spotify]: cache_location=C:\Users\Tony\AppData\Local\Temp\SoundBounce_temp 2014-12-19 19:58:36,814 DEBUG[SoundBounce.SpotifyAPI.Spotify]: settings_location=C:\Users\Tony\AppData\Local\Temp\SoundBounce_temp 2014-12-19 19:58:37,223 DEBUG[SoundBounce.SpotifyAPI.Spotify]: sp_session_preferred_bitrate() to 320k succeeded! 2014-12-19 19:58:37,284 DEBUG[SoundBounce.SpotifyAPI.Spotify]: libspotify > 00:58:37.284 I [ap:1752] Connecting to AP ap.gslb.spotify.com:4070 2014-12-19 19:58:37,285 DEBUG[SoundBounce.SpotifyAPI.Spotify]: libspotify > 00:58:37.285 E [c:/Users/spotify-buildagent/BuildAgent/work/1e0ce8a77adfb2dc/client/core/network/proxy_resolver_win32.cpp:215] WinHttpGetProxyForUrl failed 2014-12-19 19:58:37,374 DEBUG[SoundBounce.SpotifyAPI.Spotify]: libspotify > 00:58:37.374 I [ap:1226] Connected to AP: <IP_ADDRESS>:4070 2014-12-19 19:58:41,900 DEBUG[SoundBounce.SpotifyAPI.Spotify]: libspotify > 00:58:41.900 I [offline-mgr:2084] Storage has been cleaned 2014-12-19 19:58:41,900 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:41,905 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:42,753 DEBUG[SoundBounce.WindowsClient.SpotifyEnabledBrowser]: From http://cdnjs.cloudflare.com/ajax/libs/react/0.12.1/JSXTransformer.js Line 318: You are using the in-browser JSX transformer. Be sure to precompile your JSX for production - http://facebook.github.io/react/docs/tooling-integration.html#jsx 2014-12-19 19:58:42,951 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:45,587 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:45,592 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:46,943 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:47,040 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:48,542 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:49,084 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:50,110 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:50,223 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:50,887 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:51,570 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:52,235 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:52,917 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:53,638 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:54,359 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:54,938 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:55,688 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:56,397 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:57,030 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:57,386 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:57,846 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:58,147 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:58,458 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:58,924 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:58:59,129 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:59:00,917 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:59:01,024 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:59:02,186 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:59:02,353 DEBUG[SoundBounce.SpotifyAPI.Spotify]: spotify > metadata_updated 2014-12-19 19:59:03,156 DEBUG[SoundBounce.SpotifyAPI.Spotify]: libspotify > 00:59:03.156 I [offline-mgr:2032] 0 files are locked. 0 images are locked 2014-12-19 19:59:03,156 DEBUG[SoundBounce.SpotifyAPI.Spotify]: libspotify > 00:59:03.156 I [offline-mgr:2058] 0 files unlocked. 0 images unlocked 2014-12-19 19:59:03,887 DEBUG[SoundBounce.SpotifyAPI.Spotify]: Successfully closed libspotify session. In the meantime, if someone finds themself locked out because they left a session running somewhere else -- as I seem to do every other day :angry: -- they can use the "Log out everywhere" button in their Spotify account overview https://www.spotify.com/us/accounts/overview/.
2025-04-01T06:39:59.215572
2019-07-15T06:02:50
2457778398
{ "authors": [ "jbigot" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9546", "repo": "pdidev/pdi", "url": "https://github.com/pdidev/pdi/issues/143" }
gharchive/issue
Use python setuptools & pip to distribute pdicfg_validate script In GitLab by @jbigot on Jul 15, 2019, 08:02 Replaced by #153 closed This is done for zpp, just need to copy now reopened
2025-04-01T06:39:59.221573
2021-01-18T09:11:08
788080842
{ "authors": [ "Potherca", "frankgerhardt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9547", "repo": "pdsinterop/php-solid-server", "url": "https://github.com/pdsinterop/php-solid-server/issues/46" }
gharchive/issue
Which features are a available and which are not? Maybe I'm blind or my browser renders it badly. From the main page: In such a list I'd expect some checked check boxes and some uncheck check boxes. Which raises the quesiton what available? Like (in Markdown): [ ] unchecked [x] checked Hi @frankgerhardt, Thank you for pointing this out! There's nothing wrong with your browser. We're just way behind with updating our documentation. :disappointed: Current @poef, @ylebre, and I are hard at work trying to complete the implementation of milestones 4. (Social web apps). Most of the other features are fully functional, we just never got around to actually checking those boxes. (Although the code still needs cleaning up and there might be some bugs caused by hard-coded values). You can get a slightly more detailed idea of where we are by following our process in the Solid test-suite repo. As our current focus is on the code, my estimate is that our documentation will remain out-of-date for another month or two. Hi @frankgerhardt, Thank you for pointing this out! There's nothing wrong with your browser. We're just way behind with updating our documentation. :disappointed: Current @poef, @ylebre, and I are hard at work trying to complete the implementation of milestones 4. (Social web apps). Most of the other features are fully functional, we just never got around to actually checking those boxes. (Although the code still needs cleaning up and there might be some bugs caused by hard-coded values). You can get a slightly more detailed idea of where we are by following our process in the Solid test-suite repo. As our current focus is on the code, my estimate is that our documentation will remain out-of-date for another month or two. Fixed in 03c8e961772c324b4718e0e778a2a38054b3ef6d
2025-04-01T06:39:59.240655
2022-09-21T17:28:34
1381265501
{ "authors": [ "garrettcollier", "peanut-brother" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9548", "repo": "peanut-brother/categorical-to-do", "url": "https://github.com/peanut-brother/categorical-to-do/issues/1" }
gharchive/issue
Add new Category AlertDialog deactivates wrong button when no text is entered The cancel button is greyed out instead of the accept button when adding a new category. @peanut-brother I can take this on @peanut-brother Just finished a pull request addressing this issue. Let me know what you think
2025-04-01T06:39:59.243286
2021-02-09T05:43:50
804228692
{ "authors": [ "pearanalytics" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9549", "repo": "pearanalytics/uptime", "url": "https://github.com/pearanalytics/uptime/issues/194" }
gharchive/issue
⚠️ Hartman Foundation has degraded performance In 3e8752f, Hartman Foundation (https://gordonhartman.com/) experienced degraded performance: HTTP code: 200 Response time: 2528 ms Resolved: Hartman Foundation performance has improved in 245fe9c.
2025-04-01T06:39:59.278278
2022-05-22T19:46:38
1244348895
{ "authors": [ "moltam89" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9550", "repo": "pedrouid/web3modal-ethers-example", "url": "https://github.com/pedrouid/web3modal-ethers-example/pull/2" }
gharchive/pull-request
Examples for accountsChanged and chainChanged events Hi @pedrouid , These examples might help newcomers like me to understand WalletConnect a little bit more. I used the https://github.com/WalletConnect/walletconnect-test-wallet repo to test my changes. Cheers, Tamas I pushed an extra commit to get the Checksummed address, isValid returned false for me in case of MetaMask.
2025-04-01T06:39:59.281162
2015-12-18T07:35:31
122897910
{ "authors": [ "benurb", "meaku" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9551", "repo": "peerigon/erroz", "url": "https://github.com/peerigon/erroz/issues/8" }
gharchive/issue
Add git tags and/or a changelog Hello there :smile: I tried to update erroz in one of our applications from 0.1.4 to 1.0.0 and it's really hard to find out which breaking change caused the major version bump. To me it's unclear which git commits were used to publish 0.1.4 and 1.0.0 to npm. Maybe you can make this a bit more clear :+1: Ben Sorry @benurb. That's something i wanted to start with the next release. I think the breaking change was that i removed the toJSON function. Additionally i added erroz.AbstractError for comparison and customizing. The documentation on master is for the upcoming release, which i will release before christmas. Just ping me if you need help upgrading :) Hey meaku :smile: Ok, I'll just give it a try. I'll come back to you if I need help. Thanks! Ben
2025-04-01T06:39:59.322199
2015-10-08T20:49:09
110538220
{ "authors": [ "orangejulius", "rmglennon" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9554", "repo": "pelias/pelias-doc", "url": "https://github.com/pelias/pelias-doc/pull/45" }
gharchive/pull-request
Updates to Get started topic Made the bullet list into paragraphs Add links to other topics Add examples of what you can do with search Take a look at the next-to-last paragraph, in particular. I wasn't as clear on the meaning of some of the bullet points as I was reworking them. In addition: update the URLs in the reverse topic with correct formatting (changing YOUR_KEY to search=XXXXX and removing extra / in api/_key in URLs. remove or fix TODO notes. Also, why do the URLs have underscores in them? For example: [/v1/reverse?api_key=search-XXXXXXX&point.lat=48.858268&point.lon=2.294471] These changes are all good and the paragraphs are much better quality than the bullet points. I have a question though. For a very high level overview doc like this one (I think this will often be one of the very first things people who are new to Pelias read), is it useful to have some bullet points that can quickly be skimmed over to get an idea of what you can do with this thing you're reading about? Good points. Thanks, @orangejulius, for giving it a read. Let me come up with about a couple of concise bullet points of what Search does and how to get started. If we're calling it a "get started", it should tell me that info up front. That can they link to other topics. Those bullet points don't have to go in this document, now that I think about it. I'd love to hear your ideas for the best place. Maybe we keep the new text here as is, and add the bullet points to the index page? Going to merge this. I have some work to do on the index page to make the docs site consistent. Will see if the bullet points belong here or in the index file.
2025-04-01T06:39:59.382392
2024-08-21T06:06:50
2477140802
{ "authors": [ "kurealnum", "morr0ne", "pennybelle" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9555", "repo": "pennybelle/pbfetch", "url": "https://github.com/pennybelle/pbfetch/pull/22" }
gharchive/pull-request
Formatted stats.py, added some documentation. Git is saying that I removed a line from config_default.txt. Not sure why, but it looks fine. Anyways, please let me know if you want to change anything with the documentation. It's pretty basic at the moment. Thank you for your contribution! The documentation updates look great and are much appreciated. However, I have some concerns about the formatting changes. Currently, the project does not have a standard formatter set up, which might explain the discrepancies between the formatting in your PR and the formatting on other systems. It’s important to ensure consistency across different environments, so I suggest we establish a standard formatter as part of our CI process in a future update. Regarding the changes to the configuration files, it would be best to remove them from this PR. Please rebase your branch to clean up those changes. In the future, be mindful to only stage (git add) the specific files or changes you intend to include in a PR. Once these adjustments are made, I believe we can proceed with merging the documentation updates. We can address setting up a formatter and integrating it into our CI in a separate PR Sorry about the unexpected close, I renamed the master branch to main which is why this closed automatically. Ive also made dev the default branch so please make new PRs on the dev branch instead of main (the replacement for master). Since this closed automatically I don't see a way to reopen it, so if you'd like to just make a new PR on dev with the documentation changes we can continue from there. That sounds fine! I'll open one soon.
2025-04-01T06:39:59.417558
2024-11-12T22:54:15
2653553116
{ "authors": [ "conorsch", "zbuc" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9556", "repo": "penumbra-zone/UIPs", "url": "https://github.com/penumbra-zone/UIPs/pull/11" }
gharchive/pull-request
Add UIP-6: Noble Forwarding Opening as draft pending discussion on forum: https://forum.penumbra.zone/t/uip-noble-address-forwarding/125 #10 is already claiming UIP-6; mind bumping to UIP-7?
2025-04-01T06:39:59.421028
2024-05-26T04:59:40
2317483490
{ "authors": [ "TalDerei" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9557", "repo": "penumbra-zone/penumbra", "url": "https://github.com/penumbra-zone/penumbra/pull/4485" }
gharchive/pull-request
specification: file organization in protocol spec Describe your changes Several parts of the protocol are not specified, while being referred to in other parts of the documentation. We should remove these empty patches in our protocol spec, and use this PR as a guiding reference of what sections were removed as we periodically fill them in over time. @conorsch can you help better format the docs? This references components X and Y in the ECC audit log. Auxiliary to this is component Z, which should probably be (but currently is not) captured here since they're also spec-related changes. This also captures supplemental spec changes based on specific spec-related comments made in the audit. Issue ticket number and link Checklist before requesting a review [x] If this code contains consensus-breaking changes, I have added the "consensus-breaking" label. Otherwise, I declare my belief that there are not consensus-breaking changes, for the following reason: @redshiftzero this doesn't have an associated issue attached to it this is a staging area for spec-related changes, and we'll focus on addressing the rest of the spec changes later this week once phase 2 is unblocked.
2025-04-01T06:39:59.424274
2024-07-17T21:40:59
2414647496
{ "authors": [ "conorsch" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9558", "repo": "penumbra-zone/penumbra", "url": "https://github.com/penumbra-zone/penumbra/pull/4720" }
gharchive/pull-request
chore: update cometbft version v0.37.5 -> v0.37.9 Describe your changes Updates the recommended CometBFT version to the latest in the 0.37.x series, and also relaxes the language in the guide, to make clear that anything in 0.37.x is acceptable. Issue ticket number and link N/A Checklist before requesting a review [x] If this code contains consensus-breaking changes, I have added the "consensus-breaking" label. Otherwise, I declare my belief that there are not consensus-breaking changes, for the following reason: docs changes and tweaks to deploy scripts, no application code changes. I also tacked on a version bump to the recommended penumbra versions, to match 0.79.1, which went out today https://github.com/penumbra-zone/penumbra/releases/tag/v0.79.1
2025-04-01T06:39:59.427802
2020-01-30T22:30:52
557795824
{ "authors": [ "124bit", "dmontagu", "ewjoachim" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9559", "repo": "peopledoc/pylogctx", "url": "https://github.com/peopledoc/pylogctx/issues/46" }
gharchive/issue
Adding support for contextvars Currently this package uses thread locals to maintain the logging context. However, that doesn't work for async code, especially ASGI servers (which may concurrently handle multiple requests with a single thread). The way things like this are typically handled in the async world is through the use of contextvars. (They were added in 3.7, but there is a backport package for 3.6.) Would there be any possibility of accepting a PR adding optional support for the use of contextvars to manage the logging context rather than thread locals? I'm imagining something along the lines of adding a top level use_contextvars() call that would attempt to import contextvars (or the 3.6 backport), and if successful would replace the pylogctx.core.context threading.Local instance with a contextvars.ContextVar or similar. Great suggestion. I need this functionality in my project I'd definitely accept a PR that switches to contextvars :) I'm also totally fine dropping support for old Pythons. Python 3.6 is still supported but only for a few months (23 Dec 2021). I think it's not unreasonable to imagine that 3.6 users will be stuck with the current version if we make a new one. This should make the work easier to you :) I just realized the original issue was old :D Sorry for the late answer !
2025-04-01T06:39:59.477716
2024-02-02T18:10:55
2115499970
{ "authors": [ "donaldcampbelljr" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9560", "repo": "pepkit/pipestat", "url": "https://github.com/pepkit/pipestat/issues/148" }
gharchive/issue
pipestat summarize: columns view should only show stats NOT objects The columns should just show stats: Currently, it is also showing objects (which doesn't make sense): I spent some time on this but reverted my changes. It is a bit tricky because the summary table is tied to the column plotting (due to jinja template). I would need to take out all non-stats objects from the summary table. But these results are used here in the summary table AND for the individual record pages. Basically, this might require more refactoring that I originally thought. Can probably tackle this issue during the same refactor as well: https://github.com/pepkit/pipestat/issues/150 This should now be solved with the above PR #182
2025-04-01T06:39:59.506673
2023-11-16T08:04:38
1996289212
{ "authors": [ "JNKPercona", "inelpandzic" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9561", "repo": "percona/percona-postgresql-operator", "url": "https://github.com/percona/percona-postgresql-operator/pull/578" }
gharchive/pull-request
K8SPG-430: Add labels to backup objects CHANGE DESCRIPTION Problem: Labels not aligned for backup objects (repo sts, jobs, and repo PVC). Solution: Add missing labels CHECKLIST Jira [x] Is the Jira ticket created and referenced properly? [x] Does the Jira ticket have the proper statuses for documentation (Needs Doc) and QA (Needs QA)? [x] Does the Jira ticket link to the proper milestone (Fix Version field)? Tests [x] Is an E2E test/test case added for the new feature/change? [x] Are unit tests added where appropriate? Config/Logging/Testability [x] Are all needed new/changed options added to default YAML files? [x] Are the manifests (crd/bundle) regenerated if needed? [x] Did we add proper logging messages for operator actions? [x] Did we ensure compatibility with the previous version or cluster upgrade process? [x] Does the change support oldest and newest supported PG version? [x] Does the change support oldest and newest supported Kubernetes version? Test name Status demand-backup passed init-deploy passed monitoring passed operator-self-healing passed scaling passed scheduled-backup passed self-healing passed start-from-backup passed telemetry-transfer passed users passed custom-extensions passed We run 11 out of 11 commit: https://github.com/percona/percona-postgresql-operator/pull/578/commits/98519d6a16bbec9bb8cf75efe08c70e6c7b4f187 image: perconalab/percona-postgresql-operator:PR-578-98519d6a1
2025-04-01T06:39:59.550513
2023-04-25T16:21:38
1683484669
{ "authors": [ "JNKPercona", "tplavcic" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9562", "repo": "percona/percona-server-mysql-operator", "url": "https://github.com/percona/percona-server-mysql-operator/pull/359" }
gharchive/pull-request
CLOUD-783 - Cleanup test cluster before deletion CHANGE DESCRIPTION Problem: Short explanation of the problem. Cause: Short explanation of the root cause of the issue if applicable. Solution: Short explanation of the solution we are providing with this PR. CHECKLIST Jira [ ] Is the Jira ticket created and referenced properly? [ ] Does the Jira ticket have the proper statuses for documentation (Needs Doc) and QA (Needs QA)? [ ] Does the Jira ticket link to the proper milestone (Fix Version field)? Tests [ ] Is an E2E test/test case added for the new feature/change? [ ] Are unit tests added where appropriate? Config/Logging/Testability [ ] Are all needed new/changed options added to default YAML files? [ ] Are the manifests (crd/bundle) regenerated if needed? [ ] Did we add proper logging messages for operator actions? [ ] Did we ensure compatibility with the previous version or cluster upgrade process? [ ] Does the change support oldest and newest supported PS version? [ ] Does the change support oldest and newest supported Kubernetes version? Test name Status async-ignore-annotations passed auto-config passed config passed config-router failure demand-backup passed gr-bootstrap failure gr-demand-backup passed gr-ignore-annotations passed gr-init-deploy passed gr-one-pod passed gr-scaling passed gr-tls-cert-manager passed haproxy passed init-deploy passed limits passed monitoring passed one-pod passed scaling passed semi-sync passed service-per-pod passed sidecars passed tls-cert-manager passed users passed version-service passed We run 24 out of 24 commit: https://github.com/percona/percona-server-mysql-operator/pull/359/commits/6ef395e25c546e3ff92044065331570abc69e237 image: perconalab/percona-server-mysql-operator:PR-359-6ef395e Test name Status async-ignore-annotations passed auto-config passed config passed config-router passed demand-backup passed gr-bootstrap failure gr-demand-backup passed gr-ignore-annotations passed gr-init-deploy passed gr-one-pod passed gr-scaling passed gr-tls-cert-manager passed haproxy passed init-deploy passed limits passed monitoring passed one-pod passed scaling passed semi-sync passed service-per-pod passed sidecars passed tls-cert-manager passed users passed version-service passed We run 24 out of 24 commit: https://github.com/percona/percona-server-mysql-operator/pull/359/commits/6ef395e25c546e3ff92044065331570abc69e237 image: perconalab/percona-server-mysql-operator:PR-359-6ef395e
2025-04-01T06:39:59.584483
2017-12-21T07:48:46
283798845
{ "authors": [ "fluca1978", "zoffixznet" ], "license": "Artistic-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9563", "repo": "perl6/Pod-To-HTML", "url": "https://github.com/perl6/Pod-To-HTML/pull/29" }
gharchive/pull-request
Use of a debug environment variable. Instead of defining a couple of methods with the same name 'Debug' to enable or disable the debugging output, borrow the same P6DOC_DEBUG environment variable. Example of usage: % export P6DOC_DEBUG=true % perl6 --doc=HTML ~/your_pod_file.pod Tested on: % perl6 --version This is Rakudo version 2017.11 built on MoarVM version 2017.11 implementing Perl 6.c. This would precompile the debug at whatever value P6DOC_DEBUG was at during module installation time and it won't change later on, even if P6DOC_DEBUG env var is changed. Change constant to my https://docs.perl6.org/language/traps#Constants_are_Compile_Time Sorry I pushed a wrong commit in a rush, applying my to constant. I believe it is possible to cherry pick only 3689fa3. I can provide a new pr if that is cleaner. 👍 Thanks!
2025-04-01T06:39:59.604882
2022-03-07T11:16:02
1161278679
{ "authors": [ "celian-garcia" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9564", "repo": "perses/common", "url": "https://github.com/perses/common/pull/40" }
gharchive/pull-request
ETCD - Filter out locks from query results According to the ETCD concurrency code directly https://github.com/etcd-io/etcd/blob/8ac44ffa5fcccc7928876be4682c07f50b5e3b7e/client/v3/concurrency/mutex.go#L114 The value of the lock seems to be "" empty string. Thus I would like to filter empty values, hoping that it will prevent us from having this kind of issue in query all : error decoding the value associated with the key '<queried-key>/<one-real-element>/3477f3fdab5002b': unexpected end of JSON input It is probably not a good idea to check for the value to determinate if we encounter a lock or not. Despite the fact that I'm pretty sure that a lock value is empty string, one can simply set a real object with the empty string and would like to query it. Probably a better approach would be to use leasing instead. Like if the object contains a lease, then it shouldn't be part of the result, but is there a case where we'd like to retrieve a real object (not a lock) with a lease inside ? I think not. Another solution would be to not fail on Unmarshal issues, and continue. This is actually the better for me as currently we have a critical failure. You can currently make everything fail with that command etcdctl put <query key>/0 "". Because 0 will appears lexically at first result, and the first result being not parsable, it will fail and not make the rest
2025-04-01T06:39:59.612289
2022-10-28T13:44:34
1427287778
{ "authors": [ "Nexucis", "saminzadeh" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9565", "repo": "perses/perses", "url": "https://github.com/perses/perses/pull/703" }
gharchive/pull-request
put the variable in a sticky header when you scroll down This PR fixed #576. I slightly modified the variable component to made it. I hope you won't mind if it's modified like that. Perhaps if you don't want to have this behaviour by default we could add a boolean to activate it. So if someone want to use this component outside of perses it won't be bothered by this behaviour. https://user-images.githubusercontent.com/4548045/198618810-053ad10c-26c7-42cf-bb30-8150433a9368.mov /cc @saminzadeh Signed-off-by: Augustin Husson<EMAIL_ADDRESS> As a side note, when you change a variable value it re-draw everything and close every row opened. Perhaps there is something to fix around that. WDYT @saminzadeh ? I think this looks great! My only question is how does this look when the Dashboard is embedded in another app where it's not full screen. I assume it should still work ok. We could always add a prop like sticky={true} if its an issue. As a side note, when you change a variable value it re-draw everything and close every row opened. Perhaps there is something to fix around that. WDYT @saminzadeh ? Yeah this is a known issue. I think we need to write some optimizations around when to render. I think this looks great! My only question is how does this look when the Dashboard is embedded in another app where it's not full screen. I assume it should still work ok. We could always add a prop like sticky={true} if its an issue. ah that's a good question. I have no idea, but yeah definitely adding a properties like you suggest should fix the issue if there is one. For that matters, I have been told there is an even simpler way to have a sticky header. We just need to use the css directive position: 'sticky' (based on https://www.w3schools.com/howto/howto_css_sticky_element.asp) but somehow it didn't work. I don't know why, maybe there is a conflict with React MUI somehow. Here what I tried : <Box display={'flex'} justifyContent="space-between" position={'sticky'} top={0}> <Stack direction={'row'} spacing={2}> {showVariables && variableDefinitions.map((v) => ( <Box key={v.spec.name} display={v.spec.display?.hidden ? 'none' : undefined}> <TemplateVariable key={v.spec.name} name={v.spec.name} /> </Box> ))} </Stack> </Box> even if I put position={'sticky'} top={0} in the previous box, it doesn't work too. @sjcobb @saminzadeh I added the optional props variableIsSticky. By default, the variable list is not sticky. And it is explicitly set at true in the app. Should be good to merge right ? I noticed in others PRs, we started to complete the Changelog with unrelease section. Should I do the same ? I added a changelog entry, we should be good to merge @sjcobb @saminzadeh
2025-04-01T06:39:59.614885
2018-03-25T19:25:02
308377461
{ "authors": [ "soumak77" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9566", "repo": "personality-insights/sunburst-chart", "url": "https://github.com/personality-insights/sunburst-chart/issues/39" }
gharchive/issue
text is not centered inside rings (d3 v4) The text displayed inside the rings is not being centered properly for D3 version 4. This is working for version 3. fixed in https://github.com/personality-insights/sunburst-chart/pull/38
2025-04-01T06:39:59.619841
2016-08-28T23:50:35
173675120
{ "authors": [ "acwwat", "s-rajaraman", "tipsy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9567", "repo": "perwendel/spark", "url": "https://github.com/perwendel/spark/issues/648" }
gharchive/issue
Spark does not start w/ only static file location set I have just started learning about Spark with the intention of using it as a lightweight application server that mostly serve static files and occasionally use templates for pages with data. I created a simply main method in my application class as follows: public static void main(String[] args) { staticFiles.location("/public"); } When I start my application, it seems that Spark will not start and immediately exits with RC 0. However, if I add a route (e.g. using get() or even adding the route overview module, then Spark would ignite and start as expected. I can also get the server to start by calling Spark.ignite() but according to the code comment I am only to call this method directly when websocket is used. I am pretty sure that I'll add a route to my application eventually so this is not really a problem here. But I'd still like to know 1) If this is a bug, or 2) if it is not a bug, what is the correct way to get Spark to ignite/start when serving only static files. Thanks. I just added a get method inside so the server is listening. import static spark.Spark.*; public class FileServer { public static void main(String [] args){ staticFiles.location("/public"); get("/", (res, req) -> "hello world"); } } Just spent a bit more time looking at the code, and the behavior seems to be by design. As with webSocket(), staticFileLocation() or externalStaticFileLocation() must be called before init() is run for the first time. Meanwhile init() is called by any of the routing methods via addRoute(), so Spark is ignited properly whenever a route is added. Otherwise init() must be manually called in the correct sequence. The requirement of method call sequence has been documented for the WebSocket case in http://sparkjava.com/documentation.html#websockets, but would be nice if the same can be mentioned in http://sparkjava.com/documentation.html#static-files. That means: staticFileLocation() or externalStaticFileLocation() must be called before init() is called or routes are added If using only static file and WebSocket, init() must be called manually. Please consider this as a doc update. Thanks. Added to docs (but not currently live)
2025-04-01T06:39:59.672735
2016-03-29T10:18:13
144221184
{ "authors": [ "digeomel", "jmhutch", "kevinlawler", "n1vux", "petdance", "stephenostermiller", "wheany" ], "license": "artistic-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9568", "repo": "petdance/ack2", "url": "https://github.com/petdance/ack2/issues/596" }
gharchive/issue
Trim or ignore long lines If ack finds a match in e.g. minified js file, or other files with just one or few very long lines, it will flood the output with the contents of the file and likely push all useful matches off-screen. I would like an option to either ignore such matches altogether (maybe list the file name), or possibly to only show some number of characters worth of context around the match. Ack should already be ignoring minified JS files. What JS files is your ack finding? Minified js was maybe a bad example because they usually have .min.js extension, but there are minified js files that don't have the extension. I think the Vaadin/GWT UI toolkit for one. Also some tools produce minified html and xml files, which also cause problems with ack. What happens when you grep these files? I don't understand the question. If I grep (or 'ack' or 'ack-grep', doesn't matter) these files and they have a match, I get a several screenfuls of text with a matches highlighted somewhere in the mess. I don't understand the question. Not a trick question. Just wanted to know what grep did. For the most part, I try to keep ack and grep behaving the same. Well, in that case, both grep and ack fill the terminal with useless amounts of text. Depending on where I run the command, that can mean thousands of rows of scrollback, if I'm unlucky enough to have multiple minified files that match. One possibility could be being able to define characters that work as line breaks depending of file format. E.g. if you find a match in a .js file with long lines, treat semicolons like linefeeds (for the purposes of -A -B and -C switches) E.g. if you find a match in a .js file with long lines, treat semicolons like linefeeds (for the purposes of -A -B and -C switches) That's a level of source awareness that we don't want to get into. It wouldn't have to be language aware code, it could be an option, just like --type-set or --ignore-dir This is probably only a problem with languages that can be minified in the first place, and those have to have some other statement separator, so it could be like --statement-separator or --record-separator or something similar. A co-worker just IM'd me with same complaint about *.js. Since not all minifiers follow .min.js quasi-convention, there's much noise. grep doesn't do any DWIM magic, we do, so expectations are higher of ack. an option to trim (long|all) lines to KWIC (keyword in context) with specified window before/after start of match (default screen width when piped). For a proper KWIC index, the start of match will be centered on screen even if left context less than half screen an option to ignore lines > NNN characters Does grep not have an option to trim lines? I'm not seeing one. which grep ? IDK. If gnu grep has one it would be good to be compatible, but we can be better. Adding this kind of option gets things pretty ugly what with highlighting the matches etc. It sounds to me that the solution should probably be more towards people excluding files to not search files they know they want to ignore. Truncating result lines so they don't explode your screen is just saying "Let's work hard to make things more palatable that we don't even want to see anyway." Truncating result lines so they don't explode your screen is just saying "Let's work hard to make things more palatable that we don't even want to see anyway." I agree they should say --perl or --type=clojure if that's what they mean, which ignores all JS whether we can tell it's .min.js or not. And if both the minified and full JS are in the tree, they should arrange .ackrc to ignore the minified directories. We detect .min.js if that is in use. Adding ignore .js in .ackrc in dir containing minified only may help sometimes. If we consulted .gitignore it might help DWIM otherwise. Classifying files with average line length() > 1024 as binary might help. But allowing users to say ignore lines > 1024 or 256 or whatever is good too. If only the minified is available -- e.g. not shipped, or compiled from Clojure -- and they want to see where the JS calls the back end, maybe specifically having asked for --type=js or looking for all mentions of domain-specific word, seeing statements instead of lines would help them with minified files. ( Maybe that's setting $/ aka $INPUT_RECORD_SEPARATOR $RS ? I don't think we support that ... nor can we ? Might require a preprocess filter co-routine to expand and give statement numbers as faux line numbers? ) In some cases it's actually desirable to see results within minified js (assuming it's all that's available) to put back-end code into context of a front-end call, for instance. Having the option to truncate excessively long lines at a specified limit or otherwise provide limited contextual results would be flexible and useful in a number of common use cases, rather than just excluding the files outright. I was (am?) a fan of the old KWOC/KWIC formats. (I say 'was' because who really needs a lineprinter corpus index (concordance) in the 21stCentury! But context index is still plausibly useful for text searching online.) That --output lets me generate KWOC and nearly-KWIC thrills me. I don't think we need --kwic-.... options. Maybe I can write-up a KWOC/KWIC idiom or wrapper for documentation ... What are KWOC/KWIC? On Tue, Mar 14, 2017 at 12:46 AM, Andy Lester<EMAIL_ADDRESS>wrote: What are KWOC/KWIC? ​https://en.wikipedia.org/wiki/Key_Word_in_Context ​ -- Bill Ricker <EMAIL_ADDRESS>https://www.linkedin.com/in/n1vux Closed and moved to wiki. https://github.com/petdance/ack2/wiki/Feature-requests Could you reconsider this? It's been an issue in ack for 10+ years. I can't imagine anyone considers scrolling through pages of the following to be desired functionality, and it's a common occurrence in most codebases these days. The simple way to do is to add an .ackrc compatible option that consists of a boolean flag and/or a width max limit. You don't have to get fancy with the trimming: put the matches in the center of the buffer when the line exceeds the width. This gives context on both sides and it's OK if the default buffer width results in a few lines of visual output (instead of thousands). It's been an issue in ack for 10+ years. It's been an issue with grep since the beginning of time. I don't understand what you mean by "put the matches in the center of the buffer when the line exceeds the width." line: the long matched line in a file match: the text that is highlighted (substring of line) buffer: the truncated line storage The issue with truncating lines is that it isn't clear how to display a partial line as opposed to a full line. By making the buffer at least as large as the match, then you can find the middle of the buffer by dividing the buffer length by two (and the middle of the match by dividing its length by two), and then you can put the middle of the match in the middle of the buffer. This gives equal context on either side of the match. This is a simple and good enough way to do it, though it is not the only way. Let's not get bogged down in the details of how it would be implemented internally, and keep it to the user interfece. It sounds like you're suggesting that in the case of overrunning --maxwidth that ack print out some portion of the line that has the match on it, right? Something like this? 47: ... stuff that is from the middle of the line **MATCHED TEXT** more but not to the end... How do we handle multiple matches per line? What if acked on a comma and there are 1000 matches on the line in your minified javascript? How do we handle lines that are longer than --maxwidth that show up in the context lines when using -A, -B and -C? I have ideas for output that I don't want to put out here yet, but I don't see a way to handle the two scenarios above and still display matches. It sounds like you're suggesting that in the case of overrunning --maxwidth that ack print out some portion of the line that has the match on it, right? Something like this? Yes I think the way to think about an option like --maxwidth is as a UI nicety instead of as something that plays nicely with rigorous output parsing scripts. The way I (and I presume from the Google hits, a lot of other people) use ack is as a nicer grep: I want to know what files trigger, primarily, and then secondarily it's nice to see what line numbers and what context, but ultimately I'm going to open that file in my editor and jump to the matched keyword. So for How do we handle multiple matches per line? What if acked on a comma and there are 1000 matches on the line in your minified javascript? my answer would be: we could highlight only those matches that fit in the buffer starting with the first match. Now the objection to this is that you drop some valid matches, but for the stated use case, this is OK---it's only not OK if you're doing some kind of piped scripting or something. One way to signal this to the user is to make the name obviously not script-friendly, e.g. --pretty-maxwidth or somesuch. Another way to make it play nicely with scripts is, possibly, to detect when you're outputting to a terminal and only do it then, say --terminal-output-maxwidth, and this functionality is, I think, already built-in for the coloring. For -A, -B, -C I think the answer is to truncate the other lines as well, and that it would be fine to left-align and right-truncate, though I'm not as experienced with these options. For -A, -B, -C I think the answer is to truncate the other lines as well, and that it would be fine to left-align and right-truncate, though I'm not as experienced with these options. Realistically, the default buffer widths that people are going to use will be orders of magnitude smaller than the largest untruncated line, meaning that if you're displaying untruncated lines there, it's not like the output is going to be readable anyway. Is it really meaningful to show the match on the line? Vs. just saying "There are 14 matches in this 47,320 character line", for example? For me, yes. I use the context to determine whether it was a "desirable" match or not. The group of people I've known over the years that use ack is fairly large and all developers, and the typical use case is "someone mentioned such and such string, or it came up somehow in my work, and now I need to know where all instances of 'isInitialBlankNavigation' occur in the codebase." So what does the sample output look like? How do we denote it's a partial line? If this is our normal output: t/illegal-regex.t 33- 34: return subtest "test_ack_with( $testcase: @args )" => sub { 35- my ( $stdout, $stderr ) = run_ack_with_stderr( @args ); maybe the partials look like t/illegal-regex.t 33- 34* ... whatever this that other subtest this other thing that goes very .... 35- my ( $stdout, $stderr ) = run_ack_with_stderr( @args ); Kevin, I'm planning to include in Ack3's cookbook section hints on displaying selective context and may even get Andy to include features to do better at it too. Workarounds for context in Ack 2 for KWIC/KWOC Keyword indexes (with short input lines): ack2 --output can sorta do KWOC/KWIC with evil before/after vars - --output '$&^I$'"'"'^I|| $`' # *KWOC* - --output '$`^I$&^I$'"'" # *pseudo KWIC* - but they’re nasty from Shell since mix quote and dollar - and tabs don’t truly line up if width variation exceeds a tab width For your purpose, monsterline uglified JS/html/etc, I'll make a long line version of ack-standalone and ack for perl 'use' statements. perl -pE 's/\n$//' ack-standalone | ack2 --output '$1 $2 $3' '(.{0,20})(\buse \w+(?:::\w+)*[^;]{0,40};?)(.{0,20})' | less ( Those are ^V^I tabs ) Note that it steps over 'use warnings;' when it immediately follows 'use strict' which may be ok because it finds each cluster. But can get each this way perl -pE 's/\n$//' ack-standalone | ack2 --output '$1 $2 $3' '(.{0,20})(\buse \w+(?:::\w+)*[^;]{0,40};?)((?=\buse)|.{0,20})' | less Bill On Mon, Apr 24, 2017 at 9:24 PM, Kevin Lawler<EMAIL_ADDRESS>wrote: For me, yes. I use the context to determine whether it was a "desirable" match or not. The group of people I've known over the years that use ack is fairly large and all developers, and the typical use case is "someone mentioned such and such string, or it came up somehow in my work, and now I need to know where all instances of 'isInitialBlankNavigation' occur in the codebase." — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/petdance/ack2/issues/596#issuecomment-296865214, or mute the thread https://github.com/notifications/unsubscribe-auth/AANS-MTqvMHv5vXy-UohyvQMAVb3810Pks5rzUsxgaJpZM4H6lqz . -- Bill Ricker <EMAIL_ADDRESS>https://www.linkedin.com/in/n1vux Use three dots on any side that's elided. Potentially color the dots. (Putting these "outside" is fine or you can put them inside and do the more complicated string math.) If you really want to get fancy you can put the number of dropped chars in brackets outside any elided side. On Apr 24, 2017, at 6:46 PM, Andy Lester<EMAIL_ADDRESS>wrote: So what does the sample output look like? How do we denote it's a partial line? — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or mute the thread. I'm also having the same issue. No more work is being done on ack2, but a related request came in the other day on ack3, and I think it might helpful here if it were implemented. I welcome input on that ticket: https://github.com/beyondgrep/ack3/issues/234 I too would like some sort of feature to deal with suppressing or truncating matches from long lines. My workaround is to filter the output of ack using grep to remove results that are longer than 300 characters. ack my-seach-string | grep -vE '.{300,}' but because using ack with a pipe turns off color by default, I usually turn that back on with a flag: ack --color my-seach-string | grep -vE '.{300,}' It would be nice to be able to put something in my .ackrc to ignore or truncate long lines by default that I could then override on the command line if I needed to. @stephenostermiller Please go comment on the current ticket at https://github.com/beyondgrep/ack3/issues/325
2025-04-01T06:39:59.676325
2016-05-26T18:19:56
157046737
{ "authors": [ "hoelzro", "jeffmacdonald" ], "license": "artistic-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9569", "repo": "petdance/ack2", "url": "https://github.com/petdance/ack2/issues/600" }
gharchive/issue
--pager="less -R" = command not found If you add --pager="less -R" to your .ackrc file and the run it you get something like this: (venv) mbp0 AWS-43 $ ack TODO sh: less -R: command not found Hi @jeffmacdonald, thanks for the report! This is duplicate of #447; you can solve the issue by omitting the quotes around less -R. I filed an issue #601 to document the options. Thanks for the info!
2025-04-01T06:39:59.678484
2017-04-18T22:07:47
222559967
{ "authors": [ "petdance" ], "license": "artistic-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9570", "repo": "petdance/ack2", "url": "https://github.com/petdance/ack2/issues/638" }
gharchive/issue
Inline calls to print_line_if_context In the main loop, if the line isn't a match, we do this: else { chomp; # XXX Proper newline handling? print_line_if_context( $filename, $_, $., '-' ); } If we're not doing context (the most common case), then the chomp is unnecessary and the print_line_if_context is unnecessary. Let's look at inlining that code, since we're not going to be making any more functionality changes in ack2, and keep this in mind in ack3. In my basic testing, it sped things up about 4%. Can't inline this.
2025-04-01T06:39:59.684164
2023-12-30T06:06:40
2060647787
{ "authors": [ "pete-mc", "wilsteady" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9571", "repo": "pete-mc/Summit", "url": "https://github.com/pete-mc/Summit/issues/34" }
gharchive/issue
Support Leader Access Could we get allowance for Support Leaders to use the add-on? It just displays an error message when trying to open within Terrain. I've tried doing a full clear of the cache and cookies with Terrain first to no avail. Hmm I am not sure how the support leaders work, I have tested with youth member accounts. I might need to do a screenshare with someone who has one of these accounts. I did not know there was even a support leader option. Do you know how you get them setup? I might be able to get one for my wife and can test with that. The Support Leader roles within Terrain are for District Leaders and above. We can request access to another formation for a set duration as either Read/Write or Write only. If we get Write access, we can do everything a Group Leader can within that Formations terrain. This is different to just being added to a formation as an adult leader, as when you get added you can view members and modify roles and UC, but you can't change Patrols unless you're from that group; Support Leader role lets you modify patrols. It also has a 'Branch Life' section seperate from Group Life where you can view all the groups you have access to in one place. Happy to do a show and tell some time. Topo is having to make a few changes too as we had it working, but a Terrain Update over Xmas has broken whatever was working. Now I know what you are talking about my DCL has been having the same issue with Topo. Can you do me a favor and go to Developer Tools in chrome and run some code to give me a list of the roles you have? Screenshot is below, steps are 1: click chrome menu, 2: click more tools, 3: click Developer Tools, 4: click Console Tab, 5: Paste the code in the cursor (as shown in the yellow box) and press Enter. Then copy the infomation in the Blue box and send it back to me. Find attached listing. Just a tad long. Thanks for that, it looks like it has to do with the new roles. Now I can see the names of the permissions, like support-leader-group-readwrite I should be able to add them to my code so it knows what to do. I will try to sort this for the next release. I am fairly sure I have solved this issue but I can't test it as I dont have the role. The fix will be in the next release, I will keep this one open until I can verify that it is working after the release. Also, apologies for not giving you the code in text, I just re-read this issue and the images and noticed I only gave you the code in a screenshot. Sorry! Hey Pete, just tried it and it's all working perfectly. Awesome, thanks for the update. Let me know if there is any other summary views or features that would be handy at the district, region and branch levels.
2025-04-01T06:39:59.711770
2015-08-31T23:42:46
104154691
{ "authors": [ "limnic", "peterolson" ], "license": "unlicense", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9572", "repo": "peterolson/BigInteger.js", "url": "https://github.com/peterolson/BigInteger.js/issues/44" }
gharchive/issue
new bigInt(array of bytes) does not work In Java, to manipulate a textual string (for example for RSA) you would convert the string to bytes, construct a new BigInteger with the bytes and then do modPow(...). The result would be the encrypted string. This however does not work with this library. Could this feature be added? My javascript equivalent test (with BigInteger.min.js version 3bd64ef38bba5e05912922f987f1b16e65798591): var test = "bla"; var bytes = []; for (var i = 0; i < test.length; i++) { bytes.push(test.charCodeAt(i)); } var bInt = new bigInt(bytes); console.log(bInt + ", " + bigInt.isInstance(bInt)); bInt.modPow(10, 100); And this yields the following output (Firefox console): [object Object], false TypeError: bInt.modPow is not a function Concluding that the constructor failed to create the object. The problem isn't that new bigInt(bytes) isn't working correctly, the problem is that no such constructor exists. First, bigInt is just a function, not a constructor, so using new will mess things up. (new is used internally for the BigInteger and SmallInteger classes, but the API doesn't expose those). Second, bigInts are represented internally using base 107, so I don't feel comfortable providing a method for creating a bigInt from an array of bytes, because people might assume that doing so is the cheapest/most low-level way to create a bigInt, while it actually would be a fairly expensive operation requiring base conversion. You can do what you want using a slightly different approach: generate a hex string from the bytes, and then create the bigInt using the hex string: var test = "bla"; var hexString = ""; for (var i = 0; i < test.length; i++) { hexString += ("0" + test.charCodeAt(i).toString(16)).slice(-2); } var bInt = new bigInt(hexString, 16); console.log(bInt + ", " + bigInt.isInstance(bInt)); console.log(+bInt.modPow(10, 100)); Which outputs 6450273, true 49 Let me know whether that resolves your question. Oh my bad, I assumed it was a
2025-04-01T06:39:59.756502
2023-03-02T10:06:03
1606504145
{ "authors": [ "leciuk81", "petrleocompel" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9573", "repo": "petrleocompel/hikaxpro_hacs", "url": "https://github.com/petrleocompel/hikaxpro_hacs/issues/17" }
gharchive/issue
Alarm control unit not available Hi Petr, can you help me? I can log in and set the central locally without authentication problems. I can install and configure the addon from hacs but then the control panel can’t find the alarm central. Modello dispositivo:DS-PWA96-M-WE Versione Firmware:V1.2.8 build 230110 Versione Web:V4.25.1 build 221212 sensors: DS-PDMCX-E-WE DS-PDPG12P-EG2-WE DS-PDPC12P-EG2-WE DS-PDMCK-EG2-WE DS-PDMC-EG2-WE sirens: DS-PS1-I-WE DS-PS1-E-WE tag readers: DS-PT1-WE wireless keyfob: DS-PKF1-WE I attach the debug error_log.txt I see main problem. AccessModuleType and that was fixed with 0.6.4. Next thing is Invalid detector type magnetShockDetector but we are not crashing on this. Please update to version 0.6.4 and restart homeassinstant. You are great!!!! Thank you very much. It seems to work fine now I still don't see the status of the batteries but it's a minor problem Everytime if something is not working. Issue + log. We will sort it out. More type of detectors are on the way. Also batteries are issue #16, currently working on it how to implement this.
2025-04-01T06:39:59.764557
2021-10-21T03:11:20
1032011570
{ "authors": [ "aurickq", "gaow0007" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9574", "repo": "petuum/adaptdl", "url": "https://github.com/petuum/adaptdl/issues/104" }
gharchive/issue
The meaning of progress When I implement trace collection in your osdi21-artifact branch. I am a little confuse the meaning of progress in validation-2048.csv. I obtain progress information via get_progress. However, it seems not mismatch your released trace information. My collected batch=128 trace information. My collected batch=2048 trace information. May I ask some suggestions to mitigate the mismatch? Hi @gaow0007, apologies for the late reply. progress means the number of iterations discounted by the statistical efficiency. For example, 200 iterations at 50% statistical efficiency translates to "progress" of 100. This is based on theory so may not exactly match in practice. For your experiments, one thing I can think of is to make sure you are changing the scaled-up batch size (M) and not the baseline batch size (M_0). The batch_size parameter for the DataLoader class is the baseline batch size for Pollux. To set a fixed scaled-up batch size, you should set the TARGET_BATCH_SIZE environment variable instead. Sorry this part is not well documented. Thanks for your explanation. A further question: Where does TARGET_BATCH_SIZE work in the code? I fail to search such key word. @aurickq A further question: Where does TARGET_BATCH_SIZE work in the code? I fail to search such key word. @aurickq https://github.com/petuum/adaptdl/blob/osdi21-artifact/adaptdl/adaptdl/torch/data.py On line 266. Thanks for your reply.
2025-04-01T06:39:59.774923
2024-08-10T17:11:16
2459195423
{ "authors": [ "anujdeshpande", "petzval" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9575", "repo": "petzval/btferret", "url": "https://github.com/petzval/btferret/issues/45" }
gharchive/issue
Save bonding info to file after receiving it Hi I am trying to figure how to save the bonding info to the /etc/btferret.dat file right after pairing has succeeded. Right now, when I start the server from python3 btferret.py and then s and LE server - the pairing information gets saved to the file in etc when I close the server. The problem with that kind of an approach is that if my application closes unexpectedly, I might lose the information in memory. The btferret.dat file is written on program termination by this call in close_all in btlib.c: at LINE 5115 rwlinkey(1,0,NULL); There is some protection against close_all not being called explicitly by using atexit which specifies functions that are called when the program exits: at LINE 2017 atexit(close_all); This should work when the program terminates, but if not, to save the data at any other time, call rwlinkey(1,0,NULL). In C, this is pretty simple, just declare rwlinkey in btlib.h or in your code. It will then be callable from your C code. void rwlinkey(int rwflag,int ndevice,unsigned char *addr); In Python it is more complicated because function calls go via btfpython.c, and rwlinkey would need its own entry. An easier option is to piggy-back on an existing Python call. I would suggest Set_flags: at LINE 13062 in btlib.c modify as follows void set_flags(int flags,int onoff) { if(flags == 0) { rwlinkey(1,0,NULL); return; } if(onoff == FLAG_OFF) gpar.settings &= ~flags; else gpar.settings |= flags; } Then recompile the Python module with "python3 btfpy.py build". Call from Python code via: btfpy.Set_flags(0,0) I don't think there will be any problems with this, but I cannot be absolutely sure. @petzval thanks for the detailed info - I will try this out and report here. Are you open to accepting a PR where I add this as a separate API? I'd rather not maintain a fork if I can avoid it. I've had a careful look at rwlinkey and there is a problem with calling it before program termination. If the device disconnects, a re-connection during the same session will probably fail. Some minor changes will fix this, and a revised version of rwlinkey is below. Rather than maintaining a PR, I would rather just add a new function to save the pairing info and do a full test to make sure it works . I can also see the value of adding an empty user-defined function inside btlib.c. These will be in a new version within a week or so. void rwlinkey(int rwflag,int ndevice,unsigned char *addr) { int n,k,i,j,addcount,flag; unsigned char *badd,*key; struct devdata *dp; FILE *stream; static char *fname = "/etc/btferret.dat"; static int count = -1; static int delflag = 0; static int writeflag = 0; static unsigned char zero[6] = {0,0,0,0,0,0}; static unsigned char *table = NULL; if(rwflag == 0) { // read if(ndevice > 0 && (dev[ndevice]->linkflag & (KEY_NEW | KEY_FILE)) != 0) return; // dev[]->linkey is good if(count < 0) { count = 0; // in file stream = fopen(fname,"rb"); if(stream == NULL) return; n = 0; count = fgetc(stream); if(count != 0 && count != 0xFF) { k = count*22; table = (unsigned char *)malloc(k); if(table != NULL && fread(table,1,k,stream) == (unsigned int)k) n = 1; } fclose(stream); if(n == 0) { count = 0; NPRINT "Read key data failed\n"); return; } } for(k = 0 ; k < count && table != NULL ; ++k) { badd = table + k*22; key = badd+6; if(addr == NULL) n = devnfrombadd(badd,BTYPE_CL | BTYPE_LE | BTYPE_ME,DIRN_FOR); else { n = bincmp(badd,addr,6,DIRN_FOR); if(n != 0) n = ndevice; } if( (ndevice == 0 && n > 0) || (ndevice > 0 && n == ndevice) ) { // all on init (ndevice=0) or ndevice only dp = dev[n]; if((dp->linkflag & KEY_FILE) != 0 && bincmp(zero,key+10,6,DIRN_FOR) != 0) { for(n = 0 ; n < 10 ; ++n) dp->divrand[n] = key[n]; dp->linkflag |= PAIR_FILE; } else { for(n = 0 ; n < 16 ; ++n) dp->linkey[n] = key[n]; dp->linkflag |= KEY_FILE; } } } } else if(rwflag == 1) { // write // update table if(writeflag != 0) { NPRINT "WARNING 2nd close\n"); return; } writeflag = 1; flag = 0; // no changes to table if(count > 0 && table != NULL && delflag == 0) { for(k = 0 ; k < count ; ++k) { badd = table + k*22; key = badd+6; n = devnfrombadd(badd,BTYPE_CL | BTYPE_LE | BTYPE_ME,DIRN_FOR); if(n > 0) { dp = dev[n]; if(bincmp(zero,key+10,6,DIRN_FOR) != 0) { if((dp->linkflag & PAIR_NEW) != 0) { for(n = 0 ; n < 16 ; ++n) key[n] = dp->divrand[n]; dp->linkflag &= ~PAIR_NEW; dp->linkflag |= PAIR_FILE; } flag = 1; } else if((dp->linkflag & KEY_NEW) != 0) { // must be KEY_FILE also for(n = 0 ; n < 16 ; ++n) key[n] = dp->linkey[n]; dp->linkflag &= ~KEY_NEW; dp->linkflag |= KEY_FILE; flag = 1; } } } } // count NEW additions not in table addcount = 0; for(n = 1 ; devok(n) != 0 ; ++n) { if((dev[n]->linkflag & KEY_NEW) != 0) ++addcount; if((dev[n]->linkflag & PAIR_NEW) != 0) ++addcount; } if(flag == 0 && delflag == 0 && addcount == 0) return; // no changes if(count + addcount > 100) { NPRINT "/etc/btferret.dat file of paired devices is large\n"); NPRINT "Recommendation: delete it and re-pair devices\n"); } if(count + addcount > 254) { NPRINT "Too many paired devices - delete /etc/btferret.dat\n"); NPRINT "file to reset and then re-pair devices\n"); return; } stream = fopen(fname,"wb"); if(stream == NULL) return; fputc(count+addcount,stream); if(count > 0) fwrite(table,1,count*22,stream); k = 0; for(n = 1 ; k < addcount && devok(n) != 0 ; ++n) { dp = dev[n]; if((dp->linkflag & KEY_NEW) != 0) { fwrite(dp->baddr,1,6,stream); fwrite(dp->linkey,1,16,stream); dp->linkflag &= ~KEY_NEW; dp->linkflag |= KEY_FILE; ++k; } if((dp->linkflag & PAIR_NEW) != 0) { fwrite(dp->baddr,1,6,stream); fwrite(dp->divrand,1,16,stream); dp->linkflag &= ~PAIR_NEW; dp->linkflag |= PAIR_FILE; ++k; } } fclose(stream); } else if(rwflag == 2 && count > 0 && table != NULL && ndevice > 0) { // delete flag = 0; for(k = 0 ; k < count && flag == 0 ; ++k) { badd = table + k*22; n = devnfrombadd(badd,BTYPE_CL | BTYPE_LE | BTYPE_ME,DIRN_FOR); if(n == ndevice) { // found - remove flag = 1; delflag = 1; dev[n]->linkflag &= ~KEY_FILE; for(j = k ; j < count ; ++j) { badd = table + k*22; for(i = 0 ; i < 22 ; ++i) badd[i] = badd[i+22]; } --count; } } } else if(rwflag == 3 && count > 0) { for(k = 0 ; k < count && table != NULL ; ++k) { badd = table + k*22; n = devnfrombadd(badd,BTYPE_CL | BTYPE_LE | BTYPE_ME,DIRN_FOR); if(n >= 0) NPRINT "%s (%02X) =",dev[n]->name,dev[n]->linkflag >> 10); else NPRINT "Unknown ="); for(j = 0 ; j < 22 ; ++j) { NPRINT " %02X",badd[j]); if(j == 5) NPRINT " ="); } NPRINT "\n"); flushprint(); } } } I can now see other problems associated with calling rwlinkey twice, so I cannot recommend the changes I have suggested. A new version will have a fix. Thanks a lot @petzval - will try it out and get back! @petzval do you think it makes sense to have an event for that? Maybe LE_BONDED - so that user can take further action in the callback
2025-04-01T06:39:59.786789
2024-04-17T14:33:33
2248454112
{ "authors": [ "bogi1492" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9576", "repo": "peviitor-ro/ui.orase", "url": "https://github.com/peviitor-ro/ui.orase/issues/3097" }
gharchive/issue
”Brețcu”, ”Mărtănuș”, ”Oituz” are listed in the drop-down menu Description: When user writes ”Brețcu”, ”Mărtănuș”, ”Oituz” the locations appear in the drop-down menu next to the comune and the county witch is a part according to the law. Precondition: The website is up an running. Step 1 Write in the search bar "Brețcu" Expected results The location appears in the drop-down menu as "Sat Brețcu, COVASNA (Brețcu)". Step 2 Press "x" button. Expected results The location was deleted from the search bar. Step 3 Write in the search bar "Mărtănuș" Expected results The location appears in the drop-down menu as "Sat Mărtănuș, COVASNA (Brețcu)". Step 4 Press "x" button. Expected results The location was deleted from the search bar. Step 5 Write in the search bar "Oituz" Expected results The location appears in the drop-down menu as "Sat Oituz, COVASNA (Brețcu)".
2025-04-01T06:39:59.797445
2020-11-12T00:14:07
741167777
{ "authors": [ "ethanjli" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9577", "repo": "pez-globo/pufferfish-software", "url": "https://github.com/pez-globo/pufferfish-software/pull/240" }
gharchive/pull-request
Refactor CRC interfaces and implementations This PR: Creates a templated CRCChecker interface class, as well as CRC8 and CRC32 aliases which can be used as interfaces. Removes inclusion of the Pufferfish/HAL/STM32/CRC.h header from driver files so that they can compile without the STM32 libraries. Updates the CRC32C-based driver classes to take the CRC32 interface in their constructors, rather than an STM32-specific CRC32C class. Generalizes the CRC8 software implementation to work for CRCs of any size in the templated SoftCRC class in Pufferfish/HAL/CRCChecker.h. This class subclasses CRCChecker and is usable both for any CRC8 and for CRC32C. It uses a table lookup-based algorithm rather than the naive bit-by-bit algorithm. Because it implements the CRC32 interface, it should be possible to use the SoftCRC32 as a drop-in replacement for the hardware-based HALCRC32 in unit tests by dependency injection, though I haven't written a unit test confirming this. Defines a templated CRCParameters struct for initializing SoftCRC objects in the SFM3019 driver and other Sensirion drivers. Simplifies the interface of SensirionDevice by passing the CRC implementation in the constructor, rather than passing in CRC parameters with every method call. Creates Catch2 unit tests for CRC8 (with SDP3019 parameters) and CRC32C, matching the example inputs which were used in a unit test in the backend (though that test was deleted in #90 because we moved to a third-party library for CRC in python). Adds clang-tidy checks for the Catch2 unit tests to the Github Actions workflow. Fixes clang-tidy issues for the Catch2 Nonin OEM III unit tests written by Hemanth. As discussed with Renji on Slack, this PR appears to work, so I will merge this PR in first. For records-keeping: This project is licensed under Apache License v2.0 for any software, and Solderpad Hardware License v2.1 for any hardware - do you agree that your contributions to this project will be under these licenses, too? Yes Were any of these contributions also part of work you did for an employer or a client? No Does this work include, or is it based on, any third-party work which you did not create? No
2025-04-01T06:39:59.806297
2023-02-07T07:45:00
1573840015
{ "authors": [ "PatrickDDj", "fangli333", "soskek" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9578", "repo": "pfnet-research/distilled-feature-fields", "url": "https://github.com/pfnet-research/distilled-feature-fields/issues/8" }
gharchive/issue
!problems with training! I have done all preparations for training,including apex, tinycuda. My torch version is 1.13.1+cu116. Once I trained the first DFF, I came across such err as below: Traceback (most recent call last): File "train.py", line 21, in from models.networks import NGP File "/home/DFF/models/networks.py", line 3, in import tinycudann as tcnn File "/opt/conda/lib/python3.8/site-packages/tinycudann-1.7-py3.8-linux-x86_64.egg/tinycudann/init.py", line 9, in from tinycudann.modules import free_temporary_memory, NetworkWithInputEncoding, Network, Encoding File "/opt/conda/lib/python3.8/site-packages/tinycudann-1.7-py3.8-linux-x86_64.egg/tinycudann/modules.py", line 50, in C = importlib.import_module(f"tinycudann_bindings.{cc}_C") File "/opt/conda/lib/python3.8/importlib/init.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) ImportError: /opt/conda/lib/python3.8/site-packages/tinycudann-1.7-py3.8-linux-x86_64.egg/tinycudann_bindings/_86_C.cpython-38-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1010TensorImpl36is_contiguous_nondefault_policy_implENS_12MemoryFormatE It seems that you failed to install tinycudann itself. Did the tinycudann work by itself? If it fails, it would be an issue with your installation of tinycudann. I cannot solve it in this thread. Please see the tinycudann repository. And, note that DFF README uses 1.10.2+cu111. I didn't check if later versions work. Ok,I created a new conda environment and solved this problem. inycudann_bindings/_86_C Can you show how you solve this problem? My instance is CUDA 11.7. Can you show your torch, cuda and CUDA verison? Thank you.
2025-04-01T06:39:59.808273
2016-12-22T09:31:19
197130804
{ "authors": [ "okuta", "zori" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9579", "repo": "pfnet/chainer", "url": "https://github.com/pfnet/chainer/pull/2038" }
gharchive/pull-request
Use autospec in mock creation. Use 'autospec'. Related to https://github.com/pfnet/chainer/pull/2037 LGTM!
2025-04-01T06:39:59.809143
2021-09-06T10:09:07
989007392
{ "authors": [ "kmaehashi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9580", "repo": "pfnet/pytorch-pfn-extras", "url": "https://github.com/pfnet/pytorch-pfn-extras/pull/312" }
gharchive/pull-request
doc: Move type annotation to description from signature This is a resubmit of #241 to trainer-dev branch as it was out of synchronization. /test
2025-04-01T06:39:59.810503
2024-12-02T14:35:01
2712131306
{ "authors": [ "chenbihao", "pfrankov" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9581", "repo": "pfrankov/obsidian-colored-tags", "url": "https://github.com/pfrankov/obsidian-colored-tags/issues/27" }
gharchive/issue
Tags that start with an underscore may cause incorrect rendering eg. : #theme #_theme #the_me Maybe it has something to do with Chinese characters? #教程 #obsidian #theme #_theme #the_me #_教程 #_指南 #_参考 #_公司/阿里 #_公司/百度 Hi there! That's the limitation of the plugin as mentioned in #20
2025-04-01T06:39:59.817815
2015-04-30T03:07:23
72047902
{ "authors": [ "BBcan177", "rbgarga" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9582", "repo": "pfsense/pfsense", "url": "https://github.com/pfsense/pfsense/pull/1633" }
gharchive/pull-request
Independent Dashboard Theme With the current 'Theme' design, the chosen theme is used in all pages. This work-around will allow the setting of a 'theme' for the Dashboard, Independent of the Theme for the balance of the pfSense Pages. Use case would be setting the "Dashboard" to a Wide-Screen theme, and leaving all other pages in a Standard Width theme. When the index.php file is loaded, if the "Dashboard Theme" is defined with a theme that is different than the "Default Theme", the "Default Theme" setting is captured into $g['theme_revert'] . The new "Dashboard Theme" is now used for the Dashboard. All other pages will use the "Default Theme" setting. If the user applies any changes in any of the Dashboard Widgets, it will save the "Dashboard Theme" as the "Default Theme" and this will cause all future pages to use the "Dashboard Theme". To over come this issue, code is added to the php function write_config() to revert the "Theme" back to its "Default" before saving the changes to the config.xml file. Future designs of Themes should allow for independent Theme customization per page to better utilize Wide-Screen/Standard-Width-Screen settings. Thanks Phil, I have incorporated those changes. pfSense is moving to bootstrap, as you can see in https://github.com/SjonHortensius/pfsense. I'm not sure it's worth it to add this right now.
2025-04-01T06:39:59.819797
2020-05-08T13:21:05
614739350
{ "authors": [ "vktg" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9583", "repo": "pfsense/pfsense", "url": "https://github.com/pfsense/pfsense/pull/4309" }
gharchive/pull-request
DNS/Ping/Traceroute IDN support. Issue #10538 [X] Redmine Issue: https://redmine.pfsense.org/issues/10538 [X] Ready for review Add support for IDN hostnames on the DNS/Ping/Traceroute diagnostics pages. in the same way it's possible to add IDN support to DNS Resolver/IPsec/OpenVPN etc. but I think it's better to add idn_to_utf8() to Form_Input class
2025-04-01T06:39:59.833695
2023-10-28T16:21:21
1966655151
{ "authors": [ "lcheruka", "lfittl" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9584", "repo": "pganalyze/libpg_query", "url": "https://github.com/pganalyze/libpg_query/pull/216" }
gharchive/pull-request
Support passing parse mode to parse/fingerprint functions Fixes #215 Mainly commenting so I don't forget: For the newly introduced _opt functions, I wonder if we should also consider the use cases requested in https://github.com/pganalyze/libpg_query/issues/50 here, i.e. to set GUCs that affect parsing. Need to think through this more how it could be part of the API, but might be worth doing now so we don't have to break the _opt function API in the future. @lcheruka Thanks for the contribution! Merged with slight revisions to allow passing additional options beyond the parse mode.
2025-04-01T06:39:59.852970
2024-09-22T20:13:51
2541240304
{ "authors": [ "Jubilee101", "ashu3103", "jesperpedersen" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9585", "repo": "pgmoneta/pgmoneta", "url": "https://github.com/pgmoneta/pgmoneta/pull/379" }
gharchive/pull-request
[#366] client side translation of numeric values WIP Translations for :- Command Output Format Valid field Compression field Encryption field @jesperpedersen PTAL The idea is that all changes are in cli.c. Maybe we can extract some static functions into in the public API once we have a good idea @jesperpedersen PTAL We are close with this @jesperpedersen PTAL Looks good to me besides some format issues. Would be nice if we can have some output demos~ Okay attaching some demos for commands: Backup: List Backup: Status: Staus Details: @Jubilee101 Please have a look. You can also look at converting sizes to "B", "kB", "MB", "GB", "TB" and "PB", so something like UsedSpace: 15.3MB and so on You can also look at converting sizes to "B", "kB", "MB", "GB", "TB" and "PB", so something like UsedSpace: 15.3MB and so on Can we also do for timestamp I think timestamp is ok as is @jesperpedersen Do we need to limit the number of digits after decimal in the translated size? If so, I am thinking of adding another function pgmoneta_append_double_setprecision(char* s, double d, int32_t precision) in utils.c to limit the number of digits. Yes, you add a separate function for it - and set the default to 2 digits so XXX.YY @jesperpedersen @Jubilee101 PTAL Looks like you are having some problems with rebasing, other than that it looks good Looks like you are having some problems with rebasing, other than that it looks good Not sure what is wrong with the rebasing. Kindly elaborate Looks like you are having some problems with rebasing, other than that it looks good Not sure what is wrong with the rebasing. Kindly elaborate Nvm, it looks alright now. Probably a glitch on my side, apologies~ Looks like you are having some problems with rebasing, other than that it looks good Not sure what is wrong with the rebasing. Kindly elaborate Nvm, it looks alright now. Probably a glitch on my side, apologies~ No issues :) @jesperpedersen @Jubilee101 Review!! Comments like that doesn't work AT ALL. Your pull request is open, so we will get to it I am extremely sorry @jesperpedersen @Jubilee101 Never gonna happen again Hey, can you rebase? Hey, can you rebase? Done Merged, thanks for your contribution!
2025-04-01T06:39:59.860305
2016-01-23T09:27:12
128317299
{ "authors": [ "alekbarszczewski", "pgte" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9586", "repo": "pgte/nock", "url": "https://github.com/pgte/nock/pull/455" }
gharchive/pull-request
Fixed issue with header function evaluation + added 2 tests Fixes #454 Reply header functions were evaluated twice per request. Reply header functions were overridden by evaluated values. After fix each header function is evaluated only once per request. After fix each header function is evaluated on each and every request separately. Hmm, the problem with coverage is here: https://coveralls.io/builds/4824108/source?filename=lib%2Frequest_overrider.js#L461. I've coded it to be sure that both response.headers and response.rawHeaders will be evaluated. But it seems that interceptor.headers are identical to interceptor.rawHeaders, but in array (instead of object) representation. Is that true? If yes then I can change: if (typeof value === "function") { // Check if header has not been already evaluated. Evaluate it otherwise. if (evaluatedHeaders.hasOwnProperty(key)) { response.rawHeaders[rawHeaderIndex + 1] = evaluatedHeaders[key] } else { response.rawHeaders[rawHeaderIndex + 1] = value(req, response, responseBody); } } to if (typeof value === "function") { response.rawHeaders[rawHeaderIndex + 1] = evaluatedHeaders[key]; } @alekbarszczewski agree, that should be enough. @pgte Done, squashed additional commit, thanks. @alekbarszczewski thanks! Landed on v6.0.1.
2025-04-01T06:39:59.924947
2020-04-11T07:07:00
598207630
{ "authors": [ "CaioBRosa", "Juriy", "awongCM" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9587", "repo": "phalt/swapi", "url": "https://github.com/phalt/swapi/issues/148" }
gharchive/issue
Forked my own verson of SWAPI I forked your SWAPI. I'm curious to know when following the MakeFile, which python version do you used for this? I'm assuming the service has always been written in Python 2 for a while.. Also do you recommend using virtualenv when running locally? I'm running into a lot of versioning incompatibility noises with Django modules... @awongCM have you managed to run it locally or on the server? I tried to run it locally. It failed. When troubleshooting, I realised the entire django api platform is using Python 2, instead of Python 3. I have both Python versions on my Mac. But the app still doesn’t play nicely with it... Hence the reason for raising this github issue. I have managed to start the project: https://swapi.dev, it is deployed from my fork: https://github.com/juriy/swapi thank so much @Juriy Thanks for doing that @Juriy. I sincerely appreciated your help. Just out of my insatiable curiosity, how did you manage to get your forked project to work locally? Do you have the correct Python 2 dependencies that got everything working on the outset? I'm keen to know as I want to get it working for my forked project. Thanks again! @awongCM I was deploying it right away on the server, so I didn't try it locally yet, it is a next thing to do. I used Amazon Linux 2, which already has the appropriate version of Python - 2.7.16. I had to install the toolset and dependencies (gcc-c++, make, zlib-devel zip unzip bzip2-devel postgres, postgresql-devel, python-devel, libmemcached-devel, ). Not sure if everything is needed from the list, but most of it makes sense. Then I fixed the version of keen in requirements.txt to 0.3.0, to fix the incompatibility error. Then it started without an issue. You need to provide environment variables for DEBUG and DATABASE_URL to for the service to start. That's pretty much it. As for the local setup - I haven't tried it yet, but I assume the process is going to be similar. Except, maybe, for using sqlite instead of Postgres for development mode. @Juriy Awesome stuff. Thank you very much for putting all this together. I sincerely appreciate your effort to give this much detail. This would certainly be a useful guide for me to try installing myself locally to work when troubleshooting. Thanks again!
2025-04-01T06:39:59.955943
2022-10-06T04:45:29
1398776469
{ "authors": [ "thomas-neitmann" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9588", "repo": "pharmaverse/admiral", "url": "https://github.com/pharmaverse/admiral/pull/1492" }
gharchive/pull-request
Pre-release v0.8.3 Thank you for your Pull Request! We have developed this task checklist from the Development Process Guide to help with the final steps of the process. Completing the below tasks helps to ensure our reviewers can maximize their time on your code as well as making sure the admiral codebase remains robust and consistent. Please check off each taskbox as an acknowledgment that you completed the task or check off that it is not relevant to your Pull Request. This checklist is part of the Github Action workflows and the Pull Request will not be merged into the devel branch until you have checked off each task. [ ] Place Closes #<insert_issue_number> into the beginning of your Pull Request Title (Use Edit button in top-right if you need to update) [ ] Code is formatted according to the tidyverse style guide. Run styler::style_file() to style R and Rmd files [ ] Updated relevant unit tests or have written new unit tests - See Unit Test Guide [ ] If you removed/replaced any function and/or function parameters, did you fully follow the deprecation guidance? [ ] Update to all relevant roxygen headers and examples [ ] Run devtools::document() so all .Rd files in the man folder and the NAMESPACE file in the project root are updated appropriately [ ] Address any updates needed for vignettes and/or templates [ ] Update NEWS.md if the changes pertain to a user-facing function (i.e. it has an @export tag) or documentation aimed at users (rather than developers) [ ] Build admiral site pkgdown::build_site() and check that all affected examples are displayed correctly and that all new functions occur on the "Reference" page. [ ] Address or fix all lintr warnings and errors - lintr::lint_package() [ ] Run R CMD check locally and address all errors and warnings - devtools::check() [ ] Link the issue in the Development Section on the right hand side. [ ] Address all merge conflicts and resolve appropriately [ ] Pat yourself on the back for a job well done! Much love to your accomplishment! @bms63 We have to release today. As per CRAN (related to the encoding issue): Please fix before 2022-10-06 to safely retain your package on CRAN. @bms63 If you want to get updates relating to #1454 included, then please open a PR for that today. I'm out of office today but will login again this evening to create the release and upload to CRAN.
2025-04-01T06:40:00.011611
2022-09-30T09:24:58
1392141299
{ "authors": [ "benregn", "woylie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9589", "repo": "phenixdigital/phx_live_storybook", "url": "https://github.com/phenixdigital/phx_live_storybook/issues/117" }
gharchive/issue
Icons not rendering correctly The icons in the sidebar aren't rendering correctly for me. <div class="lsb lsb-flex lsb-items-center lsb-py-3 lg:lsb-py-1.5 -lsb-ml-2 lsb-group lsb-cursor-pointer lsb-group hover:lsb-text-indigo-600" phx-click="close-folder" phx-target="1" phx-value-path="/admin/storybook/components/feedback"> <i class="fa-solid fa-caret-down lsb lsb-pl-1 lsb-pr-2"></i> </div> From my investigation, it's because the lsb class defines a font-family property which cascades down to the icon element and overrides the font-family of the fa-* classes. When I added !important to the fa-* font-family property, the icons render properly. duplicate of #111 Posted as a comment on #111 instead.
2025-04-01T06:40:00.077666
2019-02-25T03:59:58
413922915
{ "authors": [ "jinfagang", "philip-huang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9590", "repo": "philip-huang/PIXOR", "url": "https://github.com/philip-huang/PIXOR/issues/2" }
gharchive/issue
Training multi class at the same time? Hi, how to training multi class at the same time? Sorry this is not supported currently.
2025-04-01T06:40:00.089712
2023-10-17T08:33:44
1946865239
{ "authors": [ "1171-jpg", "KarthikRaju391", "philipperemy" ], "license": "ISC", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9591", "repo": "philipperemy/stanford-openie-python", "url": "https://github.com/philipperemy/stanford-openie-python/issues/61" }
gharchive/issue
Can not get feedback from Sever I installed the environment and ran the demo code. However, I can not get the feedback from the client. Is there any issue with my set up? Do you have java installed properly? I have java installed: $ java -version openjdk version "24-loom" 2025-03-18 OpenJDK Runtime Environment (build 24-loom+7-60) OpenJDK 64-Bit Server VM (build 24-loom+7-60, mixed mo de, sharing) should I use an older java runtime? even after setting up java 8, the server isn't starting: Exception in thread "main" edu.stanford.nlp.io.RuntimeIOException: argsToProperties could not read properties f ile: corenlp_server-8090116e1a084988.props at edu.stanford.nlp.util.StringUtils.argsToProperties(StringUtils.java:1060) at edu.stanford.nlp.util.StringUtils.argsToProperties(StringUtils.java:973) at edu.stanford.nlp.pipeline.StanfordCoreNLPServer.<init>(StanfordCoreNLPServer.java:183) at edu.stanford.nlp.pipeline.StanfordCoreNLPServer.launchServer(StanfordCoreNLPServer.java:1590) at edu.stanford.nlp.pipeline.StanfordCoreNLPServer.main(StanfordCoreNLPServer.java:1644) Caused by: java.io.IOException: Unable to open "corenlp_server-8090116e1a084988.props" as class path, filename or URL at edu.stanford.nlp.io.IOUtils.getInputStreamFromURLOrClasspathOrFileSystem(IOUtils.java:501) at edu.stanford.nlp.io.IOUtils.readerFromString(IOUtils.java:634) at edu.stanford.nlp.util.StringUtils.argsToProperties(StringUtils.java:1051) ... 4 more [Thread-0] INFO CoreNLP - CoreNLP Server is shutting down. I am trying to start the server manually here because when the python library tries to start the server automatically, it doesn't do -cp "<filepath"> and instead just does -cp <filepath
2025-04-01T06:40:00.096820
2015-01-23T14:15:03
55285366
{ "authors": [ "storytime" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9592", "repo": "philipsorst/angular-rest-springsecurity", "url": "https://github.com/philipsorst/angular-rest-springsecurity/issues/17" }
gharchive/issue
Spring 4 Thank you for a very good tutorial. What about Spring 4 annotations config integration? Thank you. Check my fork for Spring4 + Annotations https://github.com/storytime/angular-rest-springsecurity
2025-04-01T06:40:00.113412
2024-02-13T13:00:37
2132274237
{ "authors": [ "phimado" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9593", "repo": "phimado/status", "url": "https://github.com/phimado/status/issues/66" }
gharchive/issue
⚠️ ECDC Shop has degraded performance In 394cb8c, ECDC Shop (https://shop.memmingen-indians.de) experienced degraded performance: HTTP code: 200 Response time: 7622 ms Resolved: ECDC Shop performance has improved in c5e3f04 after 21 minutes.
2025-04-01T06:40:00.131846
2024-05-31T03:44:44
2326903624
{ "authors": [ "chrismccord" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9594", "repo": "phoenixframework/flame", "url": "https://github.com/phoenixframework/flame/pull/40" }
gharchive/pull-request
Add dynamic syncing of code and applications support TODO [x] purge consolidated protocols [x] sync priv dirs of start_apps @josevalim ready for a new pass
2025-04-01T06:40:00.160555
2016-06-10T17:54:10
159687968
{ "authors": [ "mwbrooks", "surajpindoria" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9595", "repo": "phonegap/phonegap-cli", "url": "https://github.com/phonegap/phonegap-cli/issues/606" }
gharchive/issue
Add push template to package.json We should add the push template to the CLI: https://github.com/phonegap/phonegap-template-push Added this with above commit
2025-04-01T06:40:00.164843
2017-07-19T16:51:25
244106264
{ "authors": [ "jamesyangf", "jeanpi-gomez", "lsaballo", "macdonst" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9596", "repo": "phonegap/phonegap-plugin-barcodescanner", "url": "https://github.com/phonegap/phonegap-plugin-barcodescanner/issues/509" }
gharchive/issue
PDF_417 format not being read on iOS Issue I'm using this plugin to read the barcode on US drivers' licenses and extract some information, for android it worked like a charm, the problem is when you try to use it on iOS, it doesn't read the barcode, seems like it doesn't recognize the format. According to PR-431 (https://github.com/phonegap/phonegap-plugin-barcodescanner/pull/431) this problem was resolved, still it doesn't work on iOS. Platform and Version iOS 10.3.1 Android 6.2.3 Cordova CLI version and cordova platform version cordova --version 7.0.1 cordova platforms android 6.2.3 && ios 4.4.0 Plugin version cordova plugin version 6.0.8 Sample Code that illustrates the problem cordova.plugins.barcodeScanner.scan( scanSuccess, scanError, { formats: 'PDF_417,CODE_128' } ); } function scanSuccess (result) { console.log(result.format, result.text); } According to the documentation, I don't think they support PDF_417 format on iOS yet. We could ask them if they plan to support it soon. Please comment on #478 and give me an example image. Despite the documentation , PDF_417 works fine on ios for me , but never did the same on android, I've tried with some phonegap-plugin-barcodescanner plugins's version like 6.0.6 and 7.1.0 , didn't works for me on android. So frustrating, I can't find some post that really fix it.
2025-04-01T06:40:00.166880
2016-07-01T16:28:38
163430598
{ "authors": [ "dmaus", "imhotep", "mansimarkaur" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9597", "repo": "phonegap/phonegap-plugin-contentsync", "url": "https://github.com/phonegap/phonegap-plugin-contentsync/issues/136" }
gharchive/issue
sync.cancel(); not working propperly on second cancel() When canceling the sync process the first time, it works perfectly. Starting the same process again and cancel again, contentsync still downloads the file. It's easy to reproduce from the sample code. Just put sync.cancel() in the progress-callback and start it two times. sync.on('progress', function(progress) { sync.cancel(); console.log("Progress event", progress); app.setProgress(progress); }); Only tested on iOS. I'd like to work on this as my first contribution. Can you please help me get started with it? @mansimarkaur For sure! Fork this repository, make your changes in separate branch and submit a pull request please.
2025-04-01T06:40:00.169505
2016-03-29T16:48:22
144322364
{ "authors": [ "lucasabba", "macdonst" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9598", "repo": "phonegap/phonegap-plugin-push", "url": "https://github.com/phonegap/phonegap-plugin-push/issues/761" }
gharchive/issue
iOS 9.3 Hello, i reed this https://github.com/phonegap/phonegap-plugin-push/issues/752 about push problem in iOS 9.3. I have the same problem but i don't understand what i need to do to resolve the problem. What does it mean that I need to fill everything on the push object? Thanks to all! @lucasabba if you are using the same sample code as in #752 remove the call to unregister. You should only call that method if you don't want to get push messages anymore. Also, sorry to be a jerk but I'm now closing issues that don't follow the issue submission guidelines. This plugin takes a lot of work to maintain and I can't afford the time required to solicit the required info from everyone. I'm closing this issue but please feel free to re-open the issue if you provide the requested details from the issue submission guidelines.
2025-04-01T06:40:00.185582
2023-12-12T22:34:02
2038647591
{ "authors": [ "Peda1996", "mgineer85" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9599", "repo": "photobooth-app/photobooth-app", "url": "https://github.com/photobooth-app/photobooth-app/pull/135" }
gharchive/pull-request
Enhance Photobooth Experience with Distinct ISO and Shutter Speed Settings for Live View and Capture This pull request introduces an enhancement to the photobooth application by allowing separate ISO and shutter speed settings for the live preview mode and the photo capture mode. This feature is particularly useful for cameras that do not support Exposure Simulation and when using an external flash. Looks great, thank you for this contribution! ✨ I'll let the automated tests run, test on my system later also and merge this evening.
2025-04-01T06:40:00.193934
2016-05-21T22:46:46
156125175
{ "authors": [ "JimmyBoh", "clark-stevenson", "d3lm", "jamesgroat", "kikemx78", "photonstorm", "shinygruv3", "smks" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9600", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/issues/2492" }
gharchive/issue
Graphics improperly extends DisplayObject Hi there, Looks like the signature for Graphics.generateTexture doesn't match that of DisplayObject.generateTexture. This is an issue for typescript compilation (see pixi.d.ts). Thanks for the awesomeness y'all. Experienced the same problem when I tried to update to the latest version of Phaser. Suggested fix? Because PIXI.DisplayObject.generateTexture returns a RenderTexture, where-as PIXI.Graphics.generateTexture_does_ return Texture`. So they are actually different. Then is it possible to rename Graphics.generateTexture? The current implementation prevents typescript compilation from succeeding. I wish I was back on game work :( Reminds me of my old stack question. http://stackoverflow.com/questions/29593905/typescript-declaration-extending-class-with-static-method/29595798#29595798 declare class DisplayObject { static generateTexture(): RenderTexture; } Include the same signature from the primary above to the other class and overload it. declare class Graphics extends DisplayObject { static generateTexture(): RenderTexture; static generateTexture(): Texture; } Does it work? Overloading the function sounds like a reasonable solution to me. "Include the same signature from the primary above to the other class and overload it." Except that both signatures would be visible and the 1st signature (from base class) wouldn't ever actually work as described. Renaming the PIXI.Graphics.generateTexture() might be best. Yeah totally but that is a different issue. What you see today in JS is valid and correct JS. If TypeScript cannot define it, then you must work around it. If you want to change the implementation, you have the much bigger job of making a PR which is going to be bug free, and not break 1000 peoples existing games for the sake of a function overload. Hmm. Valid javascript or not, it's a flawed design. Either Graphics derives from DisplayObject or it does not. If it does, it must provide a working definition of DisplayObject.generateTexture. Typescript is simply enforcing proper design. I'm not sure what precipitated the original change to the signature of DisplayObject.generateTexture, but you might want to reconsider... Yeah I am having this issue even when rolling back to 2.4.6. I tried that function overloading recommendation and didn't seem to work. I am also wondering if this has anything to do with the misbehaving of the cursor when changing to a hand when over a graphic. @smks - i modified the Graphics signature to match that of DisplayObject and I'm working ok...though I'm exercising very little of the engine. This is kinda a blocker for me right now. @shinygruv3 How did you get it working exactly? @JimmyBoh - I modified ../phaser/typescript/pixi.d.ts locally and changed the signature of Graphics.generateTexture to match that of DisplayObject.generateTexture: generateTexture(resolution?: number, scaleMode?: number, renderer?: PixiRenderer): RenderTexture; Awesome, thanks @shinygruv3. I'm just starting a new project with Phaser. It's small, but I will share any issues or oddities I encounter. This is now fixed in the dev branch. Well, I say 'fixed', the defs are basically reverted back to how they were before. So the TS compilation errors are done, but the return type is technically incorrect. Then again, it has been for several years now and no-one complained, so such is life. @shinygruv3 Your turnaround is not working for me :(
2025-04-01T06:40:00.199440
2016-06-07T00:47:18
158806669
{ "authors": [ "MichelAlonso", "marcelodeassis", "photonstorm" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9601", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/issues/2540" }
gharchive/issue
useHandCursor not working properly in Phaser 2.4.8 The hard cursor only appears while pressing. Example: https://jsfiddle.net/snake/vg3Ljk26/ I've tested in other versions and works fine, only the 2.4.8 have this issue, I was going to report this, just to add more info: This Issue is about: A bug in the API When mouse over the button in version 2.4.7 works normally. 2.4.7 example 2.4.8 example In 2.4.8 detects only if you click outside button(inside canvas), hold and move over the button, and also after this don't dispatch mouseOut if you release inside button. The same occur with sprite with event.onInputOver(inputEnabled is enabled). After some research looks like something related with Phaser.Pointer parameter in inputHandler._pointerOverHandler. Thanks for spending time reporting this issue. However we've already fixed this in the dev branch of Phaser, and it will be part of the next official release. You can track what's already been fixed by looking at the Change Log part of the dev README file.
2025-04-01T06:40:00.202669
2018-02-28T18:58:01
301142076
{ "authors": [ "melissaelopez", "photonstorm" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9602", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/issues/3297" }
gharchive/issue
Documentation: Broken Link to Code of Conduct in phaser/.github/CONTRIBUTING.md This Issue is about: An error in the documentation Hello! I have noticed that in the first section of CONTRIBUTING.md, the link to the code of conduct is outdated, and thus results in a 404 error. See the image below to see the broken link highlighted in blue. The fix would be to edit the url of the link from "[...]/v2/CODE_OF_CONDUCT.md", to "[...]/.github/CODE_OF_CONDUCT.md". I'm happy to issue a pull request with this fix! Thanks for opening this issue, and for submitting a PR to fix it. We have merged your PR into the master branch and attributed the work to you in the Change Log. If you need to tweak the code for whatever reason please submit a new PR.
2025-04-01T06:40:00.207825
2018-08-18T15:28:44
351828841
{ "authors": [ "lgibson02", "photonstorm" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9603", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/issues/3948" }
gharchive/issue
Cropped sprites behave unexpectedly when recropped Phaser Version: 3.11 Operating System: Windows 7 Professional x64 Browser: Firefox Developer Edition 62.0b13 x64 I have noticed while working on a project of mine that if you use .setCrop on an already cropped sprite, all other cropped sprites will also be affected. Here's a small demo to try and show this: var config = { type: Phaser.AUTO, width: 800, height: 600, backgroundColor: '#2d2d2d', parent: 'phaser-example', scene: { preload: preload, create: create } }; var group; var game = new Phaser.Game(config); function preload() { this.load.spritesheet('diamonds', 'diamonds32x24x5.png', { frameWidth: 32, frameHeight: 24 }); } function create() { group = this.add.group(); group.createMultiple({ key: 'diamonds', frame: [0,1,2,3,4], frameQuantity: 2, repeat: 1 }); group.children.iterate(function(child) { child.setCrop(0, 0, 32, 32); }); group.children.entries[10].setCrop(0, 0, 32, 8); Phaser.Actions.SetXY(group.getChildren(), 32, 100, 32); } To start off with .setCrop(0, 0, 32, 32) is ran on all children of the group, I don't believe this should have any immediately visible effect. However, strangely the frame is set to four (the last of the sprite sheet). So this is what I would have expected: This is what I actually got: After this the 10th child of the group is cropped with .setCrop(0, 0, 16, 16) which I would think would make it so you could only see a quarter of that particular sprite. But for some reason all the sprites in the group are cropped this way instead of just the one. So in end what I would expect to see is this: What I actually end up with is this: If you need any more information for this issue, please let me know. Thank you for spending time reporting this issue. However, we've already fixed this in the master branch of Phaser and it will be part of the next official release. You can track what's already been fixed by searching closed issues, or by looking at the Change Log.
2025-04-01T06:40:00.211452
2018-10-29T06:53:36
374880805
{ "authors": [ "asukiaaa", "huangshengping2017", "photonstorm", "rexrainbow" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9604", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/issues/4133" }
gharchive/issue
is video module removed from phaser3? is phaser3 still has video module? Phaser 3 has never had video support (it wasn't removed, it was just never there in the first place) - it will be added in 2019, or if someone in the community wants to do it before then, I'd merge it in too. I have a video plugin, which can display video on DOM, or canvas, here is a demo. It took time to show fullscreen video on canvas for phaser so I leave my result. import 'phaser' const GameWidth = 1280 const GameHeight = 640 const FrameCenter = { x: GameWidth / 2, y: GameHeight / 2 } export class OpeningMovie extends Phaser.Scene { private movieFrame: Phaser.GameObjects.Image private movieTexture: Phaser.Textures.CanvasTexture private video: HTMLVideoElement constructor() { super({ key: 'OpeningMovie' }) } init(params) { } preload() { } create() { this.movieTexture = this.textures.createCanvas('movie', GameWidth, GameHeight) this.movieFrame = this.add.image(FrameCenter.x, FrameCenter.y, 'movie').setInteractive() this.video = document.createElement('video') this.video.src = '/some/movie.mp4' const game = this this.video.addEventListener('loadeddata', function() { this.play() const fps = 30 const loop = () => { if (!this.paused && !this.ended) { game.movieTexture.context.drawImage(this, 0, 0, GameWidth, GameHeight) game.movieTexture.refresh() setTimeout(loop, 1000/fps) } } loop() }) this.video.addEventListener('ended', function() { game.goNextScene() }) this.video.addEventListener('pause', function() { game.goNextScene() }) this.movieFrame.on('pointerdown', () => { this.video.pause() }) } goNextScene() { this.video.remove() this.movieTexture.destroy() this.scene.switch('Menu') } update() { } } Thank you. References: https://github.com/photonstorm/phaser/issues/3575#issuecomment-455912692 https://stackoverflow.com/questions/4429440/html5-display-video-inside-canvas/38711016
2025-04-01T06:40:00.221062
2021-07-08T01:41:44
939388742
{ "authors": [ "ivorcosta", "kootoopas", "photonstorm", "samme" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9605", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/issues/5773" }
gharchive/issue
Particle emitter does not stops, even though maxParticles is being set Version Phaser Version: 3.55.2 Operating system: Windows 10 Browser: Chrome 91 Description Particle emitter does not stop creating particles, even thought the maxParticles is being set. When changing the below codepen from maxParticles: 30 to maxParticles: 19, it stops. From 20 onwards it does not stops. It is a bit random, just keep playing with the numbers. It also have the same weird behavior if you change the lifespan or speed. Example Test Code https://codepen.io/ivorcosta/pen/qBmZGON I'm not sure what's the expected behavior for maxParticles. It changed from v3.15.1 to v3.16.1, but that may have been because of an unrelated bugfix. In the test code example, the emitter is never at its limit because it never creates 30 particles. It creates only 20 or so and then recycles them. If you change lifespan or frequency it may then hit the limit if no dead particles are available. It does seem peculiar to me that once an emitter reaches its limit (described as a creation limit) it stops emitting particles at all, even if dead ones are available. I thought maxParticles was the maximum total allowed amount of particles the emitter could fire. I want a small animation that fires a few particles and stops. In order to achieve that, I have to start a emitter and use setTimeout to stop it after a while? Wouldn't have a more elegant solution? maxParticles is the maximum amount it can fire in one single burst, not in total. There is no such feature for 'kill this emitter after X particles are emitted', although I may consider it for a PR. @photonstorm Is https://labs.phaser.io/view.html?src=src/game objects\particle emitter\fire max 10 particles.js&v=3.16.1 the expected behavior then? There is no such feature for 'kill this emitter after X particles are emitted', although I may consider it for a PR. @photonstorm The emitter does stop once the sum of the recycling-pending (dead) particle count and the alive count is greater than maxParticles. maxParticles is the maximum amount it can fire in one single burst, not in total. By changing maxParticles to 15 in this example, the sum of dead and alive ones never exceed 15 since the dead ones are recycled. Set to hard limit the amount of particle objects this emitter is allowed to create. Docs sentence communicates particle object amount limit. While this is technically true since total amount of js objects (dead + alive) is not greater than that, it is misleading since one expects the option to apply on what they see on screen. https://github.com/photonstorm/phaser/blob/aa5f54cfa268c86823bf1dd52c83099fa0ef9ac2/src/gameobjects/particles/ParticleEmitter.js#L462-L471 By changing the following on the aforementioned example page, one can validate the non-exceeding counts over time: // ... maxParticles: 15, // ... function update (time, delta) { fpsText.setText('FPS: ' + (1000/delta).toFixed(3) + '\n' + `Alive ${particles.emitters.first.alive.length}` + '\n' + `Dead ${particles.emitters.first.dead.length}`); }
2025-04-01T06:40:00.225327
2020-03-23T02:05:32
585876249
{ "authors": [ "photonstorm", "samme" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9606", "repo": "photonstorm/phaser", "url": "https://github.com/photonstorm/phaser/pull/5051" }
gharchive/pull-request
Add Tiled point object and change offset in createFromObjects() This PR Adds a new feature Fixes a bug? Current Behavior Tiled point objects are parsed to { rectangle: true, width: 0, height: 0, … } When creating a sprite from a Tiled object with `{ width: 0, height: 0 }, createFromObjects() maps the Tiled object origin onto the sprite origin using the sprite's dimensions (which are not zero). This is probably not what authors want. New Behavior Tiled point objects are parsed to { point: true, width: 0, height: 0, … } When creating a sprite from a zero-size object, createFromObjects() doesn't adjust the position at all. The new sprite is given coordinates identical to the object. I think this is more sensible, but it is a "breaking" change for projects using createFromObjects() with zero-dimensioned Tiled objects. Internal Changes I removed some redundant assignments in Tiled.ParseObject(). 👍
2025-04-01T06:40:00.263529
2018-06-18T11:28:59
333231207
{ "authors": [ "konradja100" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9607", "repo": "php-ai/php-ml", "url": "https://github.com/php-ai/php-ml/issues/283" }
gharchive/issue
Can't separate numbers from strings Hi I have trained a RandomForest algoritm and it won't separate numbers from strings. This is my class: <?php declare(strict_types=1); namespace PhpmlExamples; include '../../vendor/autoload.php'; use Phpml\Classification\Ensemble\RandomForest; use Phpml\Dataset\CsvDataset; use Phpml\Dataset\ArrayDataset; use Phpml\FeatureExtraction\TokenCountVectorizer; use Phpml\Tokenization\WordTokenizer; use Phpml\CrossValidation\StratifiedRandomSplit; use Phpml\FeatureExtraction\TfIdfTransformer; use Phpml\Metric\Accuracy; $dataset = new CsvDataset('../data/languages.csv', 1); $vectorizer = new TokenCountVectorizer(new WordTokenizer()); $tfIdfTransformer = new TfIdfTransformer(); $samples = []; foreach ($dataset->getSamples() as $sample) { $samples[] = $sample[0]; } $vectorizer->fit($samples); $vectorizer->transform($samples); $tfIdfTransformer->fit($samples); $tfIdfTransformer->transform($samples); $dataset = new ArrayDataset($samples, $dataset->getTargets()); $randomSplit = new StratifiedRandomSplit($dataset, 0.1); $classifier = new RandomForest(); $classifier->setFeatureSubsetRatio('log'); $classifier->train($randomSplit->getTrainSamples(), $randomSplit->getTrainLabels()); $predictedLabels = $classifier->predict($randomSplit->getTestSamples()); $examples = [ ['content' => 'I need to build this english sentence by myself'], ['content' => 'Italiano pizza pasta margeritta'], ['content' => 'Do I have to change?'], ['content' => 'Je voudrais une boîte de chocolates.'], ['content' => 'irewhgewruhgiee e retyrtyrt'], ['content' => '64564'], ['content' => '5476'], ['content' => '9925'], ['content' => '24'], ]; foreach ($examples as $example) { $newSample = [$example['content']]; $vectorizer->transform($newSample); $tfIdfTransformer->transform($newSample); echo $classifier->predict($newSample)[0] . '<br>'; } var_dump($examples); echo 'Accuracy: '.Accuracy::score($randomSplit->getTestLabels(), $predictedLabels); ?> And this is my training data: `"content","label" "123","number" "6546","number" "548","number" "5678","number" "583","number" "38","number" "454","number" "28655","number" "65","number" "3568","number" "5679","number" "5679","number" "96","number" "Where's the nearest railway station?","string" "The storms caused flooding.","string" "Where is the duty free shop?","string" "I witnessed it happening.","string" "I would like two postcards, please.","string" "How about going to the cinema?","string" "Où est la boulangerie?","string" "Je voudrais une boîte de chocolates.","string" "Y a-t-il un autre hôtel près d'ici?","string" "Vérifiez la batterie, s'il vous plaît.","string" "La banque ouvre à quelle heure?","string" "Est-ce que je peux l'écouter?","string" "Vous devez faire une déclaration de perte.","string" "Combien des élèves y a-t-il dans votre collège?","string"` Any ideas why RandomForest recognise all strings as numbers? moved to: https://github.com/php-ai/php-ml/issues/282
2025-04-01T06:40:00.276073
2022-10-03T15:55:54
1394964749
{ "authors": [ "cezarpopa", "szepeviktor" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:9608", "repo": "php-stubs/acf-pro-stubs", "url": "https://github.com/php-stubs/acf-pro-stubs/pull/7" }
gharchive/pull-request
Generate stubs for ACF PRO 6.0.2 Hi, ACF Pro had an update recently and thought of creating a PR to update the stubs for it. Thank you, Cezar Hello @cezarpopa! This repo has not so much life in it. So I'm merging your PR blindly. Thank you. Hello @cezarpopa! This repo has not so much life in it. So I'm merging your PR blindly. Thank you!