id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1128373194
New command: Ensure that the particular Azure AD app registration exists and updates its properties if necessary Usage m365 aad app ensure [options] Description Ensures that the particular Azure AD app registration exists and updates its properties if necessary Options Option Description --manifest <manifest> Azure AD app manifest as retrieved from the Azure Portal to configure the app registration from Additional Info The command is a combination of aad app get, aad app add and aad app set. This command checks if an Azure AD app registration with the specified ID exists. If no Azure AD app registration is found, this command will create one using the information from the manifest. If an Azure AD registration is found, this command will update its properties using the information from the manifest. @waldekmastykarz i'll try this one, please assign it to me All yours! Thank you! @waldekmastykarz aad app add command ignores both , 'id' and 'appId' if provided in the manifest https://github.com/pnp/cli-microsoft365/blob/55e5dfb910c3abcede4db24cb29b041052a61ada/src/m365/aad/commands/app/app-add.ts#L175-L180 So if we don't find the 'id' property in the provided manifest of the ensure command, do we first try to get the app using 'appId' property (if present) or straightaway consider the execution as a new app creation ? I'd say we consider the app non-existent and consider new app creation. As far as I know it's not possible to create an app with a predefined objectID or clientID so in that sense they're equal and checking either one is sufficient. ok @waldekmastykarz we will go with only 'id' as required in the manifest to determine existing app. Also, if multiple 'identifierUris' are provided in manifest.json, aad app add command configures all, however, aad app set command only accepts a single --uri and replaces all the existing identifierUris with the one specified. This way, the set operation of the ensure command will not configure multiple identifierUris (not sure what are the use cases could be). Should we stop the command execution if there are multiple identifierUris in manifest ? @vipulkelkar in the Azure Portal it seems like you can configure only one application ID URI. Do you mean redirect URI for auth by any chance? Good find! I think we should consider the manifest as the desired state, so after running the ensure command, the AAD app you get should match what's in the manifest. aad set only accepts a single --uri. So if we use set command in ensure, we will not be able to configure multiple identifieruris to ensure the manifest desired state. Should we first extend the set command to accept comma separated URI's ? (similar to --redirecturis parameter ). Using the set command as-is in ensure makes sense. My vote is to extend set command to accept the manifest. @waldekmastykarz Should I create a separate issue to extend the aad app set command to accept manifest ? Please do @vipulkelkar created #3153 , we can put this one on-hold for now till the other one is implemented. Hi @vipulkelkar, I see the issue this is waiting for has been closed as completed in v5.4 of the CLI. I'll remove the on hold here. Are you still planning on working on it? Hi @martinlingstuyl , this issue is waiting on #3153 as mentioned above, which is still in progress. . The mention of #3333 (which originated from this issue) might have caused the confusion. You're correct. I've put it on hold again 👍
gharchive/issue
2022-02-09T10:50:30
2025-04-01T04:35:31.686271
{ "authors": [ "martinlingstuyl", "vipulkelkar", "waldekmastykarz" ], "repo": "pnp/cli-microsoft365", "url": "https://github.com/pnp/cli-microsoft365/issues/3035", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
516626351
Create environment request form The user must be able to request an environment to the podcast API. Closed by #6
gharchive/issue
2019-11-02T15:18:38
2025-04-01T04:35:31.807563
{ "authors": [ "hericlesme" ], "repo": "pod-cast/podcast-client", "url": "https://github.com/pod-cast/podcast-client/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1201824110
Links at top of screen in Chrome The links in the black bar at the top of the POD screen are displaying correctly in Firefox. In Chrome, though, they're not correctly spaced, and overlap each other. Resizing the window doesn't make a difference. Screenshots attached. Pull Request: https://github.com/pod4lib/aggregator/pull/582 @bobpersing thank you for reporting this. The fix is applied to the main branch of the project and will be deployed to production next Monday (at the latest) with dependency updates.
gharchive/issue
2022-04-12T13:06:42
2025-04-01T04:35:31.810359
{ "authors": [ "bobpersing", "corylown", "ktamaral" ], "repo": "pod4lib/aggregator", "url": "https://github.com/pod4lib/aggregator/issues/579", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2309913683
feat(HMS-4156): confirmation checkbox in delete domain confirm modal Add confirmation checkbox that enables delete button to warn user that the action cannot be undone. In unit test, some jest calls were replaced by newer variants as the older were deprecated. I tested the change locally but the UI was unchanged. Maybe I missed something? Screenshots look great though!
gharchive/pull-request
2024-05-22T08:31:15
2025-04-01T04:35:31.812089
{ "authors": [ "frasertweedale", "pvoborni" ], "repo": "podengo-project/idmsvc-frontend", "url": "https://github.com/podengo-project/idmsvc-frontend/pull/68", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2259134916
Unnecessary "observe" warning Description Warning "An "observe" was called from another "observe" closure, which can lead to over-observation and unintended side effects." was added in version 1.9.3 and become at unnecessary cases I attach example project where you can see that bug BadObserveExample.zip Checklist [ ] I have determined whether this bug is also reproducible in a vanilla SwiftUI project. [X] If possible, I've reproduced the issue using the main branch of this package. [X] This issue hasn't been addressed in an existing GitHub issue or discussion. Expected behavior Warning doesn't appear Actual behavior Warning appears Steps to reproduce Just run app and wait. Screens will navigate by themselves The Composable Architecture version information 1.9.3 Destination operating system iOS 17.4.1 Xcode version information 15.3 Swift Compiler version information swift-driver version: 1.90.11.1 Apple Swift version 5.10 (swiftlang-5.10.0.13 clang-1500.3.9.4) @nmalevich We recently added this runtime warning because nested observes can have unintended consequences. In this case, a viewDidLoad is evaluated in the same stack as the presenting view controller's observe, which means all child state evaluated in the child controller's observe will be observed by the parent. To work around the problem, you can move observe into viewWillAppear instead, and manage the lifetime of the observation token, or you can dispatch the observation a tick in viewDidLoad: DispatchQueue.main.async { observe { [weak self] in … } } We're still deciding if this warning is too heavy-handed, but we don't consider it a bug at the moment. Because of this I'm going to convert this to a discussion.
gharchive/issue
2024-04-23T15:08:56
2025-04-01T04:35:31.912272
{ "authors": [ "nmalevich", "stephencelis" ], "repo": "pointfreeco/swift-composable-architecture", "url": "https://github.com/pointfreeco/swift-composable-architecture/issues/3005", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2101076297
Unable to use @DependencyEndpoint with a typealias Description Not sure if this is a dependencies bug, macros bug, or macros feature? works! extension MyFeature { @DependencyClient struct Client { var refreshRequest: @Sendable (MyFeaturesRequest) async throws -> MyFeaturesRequest.Content } } Doesnt work extension MyFeature { @DependencyClient struct Client { var refreshRequest: RequestOf<MyFeaturesRequest> } } typealias RequestOf<R: NetworkRequestProtocol> = @Sendable (R) async throws -> R.Content Checklist [X] I have determined whether this bug is also reproducible in a vanilla SwiftUI project. [ ] If possible, I've reproduced the issue using the main branch of this package. [X] This issue hasn't been addressed in an existing GitHub issue or discussion. Expected behavior I'd expect the DependencyEndpoint macro to work, when the typealias is representing a closure. Actual behavior The DependencyEndpoint macro doesn't apply when the typealias is representing a closure. Steps to reproduce No response Dependencies version information No response Destination operating system No response Xcode version information No response Swift Compiler version information No response Hi @JOyo246, this is a limitation of macros. The @DependencyClient macro can only see the syntax it is directly attached to, so it has no idea what RequestOf is. So unfortunately you cannot use type aliases in this way. Since this isn't an issue with the library I am going to convert it to a discussion.
gharchive/issue
2024-01-25T19:56:59
2025-04-01T04:35:31.918288
{ "authors": [ "JOyo246", "mbrandonw" ], "repo": "pointfreeco/swift-dependencies", "url": "https://github.com/pointfreeco/swift-dependencies/issues/178", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1233077490
Request header unexpectedly mutated In this example, I'm supplying an uppercased request header but it's getting lowercased somewhere within the parsing library. I couldn't easily figure out where this was happening to fix with a PR so opening as an issue! enum TestRoute { case testing(value: String) } let testingRouter = Route(.case(TestRoute.testing)) { Path { "testing" } Headers { Field("UPPERCASED-HEADER") } } func test() async throws { let request = try testingRouter.request( for: TestRoute.testing(value: "VALUE") ) print(request.allHTTPHeaderFields!) } printed: ["uppercased-header": "VALUE"] expected: ["UPPERCASED-HEADER": "VALUE"] @eappel HTTP header names are case insensitive, so the library currently lowercases things for consistency. Have you encountered an issue in the wild where this is a problem? Yeah the API I'm dealing with is case-sensitive 😞 @eappel I think we should probably change this behavior on our end, too! @eappel We have an idea that should make it easy to fix this in the future, but we haven't quite finalized the tools yet. We'll leave the issue open for now, but will hopefully get to close it in the next month or two!
gharchive/issue
2022-05-11T19:14:44
2025-04-01T04:35:31.921260
{ "authors": [ "eappel", "stephencelis" ], "repo": "pointfreeco/swift-url-routing", "url": "https://github.com/pointfreeco/swift-url-routing/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1487987090
🛑 Pokanop Apps Frontend is down In c47bbae, Pokanop Apps Frontend (https://pokanop.com) was down: HTTP code: 500 Response time: 3149 ms Resolved: Pokanop Apps Frontend is back up in 2106e87.
gharchive/issue
2022-12-10T05:22:13
2025-04-01T04:35:31.930224
{ "authors": [ "saheljalal" ], "repo": "pokanop/uptime", "url": "https://github.com/pokanop/uptime/issues/62", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1000975924
Support triple backtick Treats ``` as a single token when do you use triple backtactics? markdown
gharchive/pull-request
2021-09-20T13:31:12
2025-04-01T04:35:31.971677
{ "authors": [ "pokey" ], "repo": "pokey/cursorless-vscode", "url": "https://github.com/pokey/cursorless-vscode/pull/279", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2404367774
🛑 Pokko - auth service is down In b25cf78, Pokko - auth service (https://id.pokko.io) was down: HTTP code: 0 Response time: 0 ms Resolved: Pokko - auth service is back up in be605bd after 4 minutes.
gharchive/issue
2024-07-11T23:45:24
2025-04-01T04:35:31.974039
{ "authors": [ "brendanmckenzie" ], "repo": "pokkocms/pokko-status", "url": "https://github.com/pokkocms/pokko-status/issues/124", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1762030913
[Persistence] Update KVStore to use Badger wrapper functions Description This PR introduces unit tests to cover the KVStore's functionality as well as updating the KVStore logic to use the Badger wrapper functions update, view, etc. This fixes the issues around deleting keys from the KVStore. Issue Fixes N/A Type of change Please mark the relevant option(s): [ ] New feature, functionality or library [x] Bug fix [x] Code health or cleanup [ ] Major breaking change [ ] Documentation [ ] Other List of changes Update KVStore to use badger's wrapper functions Add KVStore unit tests Testing [x] make develop_test; if any code changes were made [x] make test_e2e on k8s LocalNet; if any code changes were made [x] e2e-devnet-test passes tests on DevNet; if any code was changed [x] Docker Compose LocalNet; if any major functionality was changed or introduced [x] k8s LocalNet; if any infrastructure or configuration changes were made Required Checklist [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have added, or updated, godoc format comments on touched members (see: tip.golang.org/doc/comment) [x] I have tested my changes using the available tooling [ ] I have updated the corresponding CHANGELOG If Applicable Checklist [ ] I have updated the corresponding README(s); local and/or global [ ] I have added tests that prove my fix is effective or that my feature works [ ] I have added, or updated, mermaid.js diagrams in the corresponding README(s) [ ] I have added, or updated, documentation and mermaid.js diagrams in shared/docs/* if I updated shared/*README(s) NOTE @Olshansk, @dylanlott is currently working on savepoints/rollbacks that touches the same code and this PR will probably only be used as a reference as he is changing the logic their. The new wrapper functions fix the bug with Delete() but need to be integrated with savepoints. Will wait on merging this until @dylanlott gives an update This PR is already complete and ready to be merged in, while savepoints & rollbacks PR is 1+ weeks away. IMO we should merge this in and iterate on top of it. However, I will defer to @dylanlott to decide. This PR is already complete and ready to be merged in, while savepoints & rollbacks PR is 1+ weeks away. IMO we should merge this in and iterate on top of it. However, I will defer to @dylanlott to decide. I agree with merging this and addressing the logic changes necessary for savepoints later. It shouldn't hold this up.
gharchive/pull-request
2023-06-17T23:29:38
2025-04-01T04:35:31.984598
{ "authors": [ "Olshansk", "dylanlott", "h5law" ], "repo": "pokt-network/pocket", "url": "https://github.com/pokt-network/pocket/pull/838", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
708159647
Ensure qform/sform copy doesn't modify zooms (or warn if it does) What version of fMRIPrep are you using? 20.2.0rc0 What kind of installation are you using? Containers (Singularity, Docker), or "bare-metal"? TACC Run ds000229 - sub-07 Did fMRIPrep generate the visual report for this particular subject? If yes, could you share it? Details BOLD looks significantly larger than T1w, and sform was copied to qform. This has the potential to change pixdim, since qform is built from quaternions, offsets and pixdim. This is a note to verify, but probably shouldn't block. This particular instance could be a visual effect of #2282.
gharchive/issue
2020-09-24T13:12:34
2025-04-01T04:35:32.012193
{ "authors": [ "effigies", "oesteban" ], "repo": "poldracklab/fmriprep", "url": "https://github.com/poldracklab/fmriprep/issues/2284", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
570909751
FIX: Amend some error types Trying to normalize a little between TypeErrors and ValueErrors. Let's check on all error types when merging into nibabel (#110).
gharchive/pull-request
2020-02-25T23:17:01
2025-04-01T04:35:32.013435
{ "authors": [ "oesteban" ], "repo": "poldracklab/nitransforms", "url": "https://github.com/poldracklab/nitransforms/pull/63", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2516507471
Tidying Tox runs ruff format -diff. Remove suggestion to run tox -e reformat from readme. GitHub action automatically applies linting (ruff, trailing-whitespace and end-of-file-fixer) to PRs. Note that a .pre-commit-config.yaml file was added to the repo for running pre-commit. Also, bots must be given read and write permissions to make commits: This seems to work nicely. Commit 9556263 was made automatically by the bot which applied ruff to a few files. The branch now passes the tox env quality (specifically ruff format --diff). Lint GitHub Action fails because it doesn't have read and write permission to make commits: Error: Error: To https://github.com/poliastro/czml3 ! refs/heads/tidying:refs/heads/tidying [remote rejected] (refusing to allow a GitHub App to create or update workflow `.github/workflows/workflow.yml` without `workflows` permission) See top comment to apply the required change to allow the action to make commits. Reading from the UI, these are the default permissions. "You can specify more granular permissions in the workflow using YAML." Can't we just change the YAML? Also while we're at it... is this essentially equivalent to https://pre-commit.ci/ ?
gharchive/pull-request
2024-09-10T13:36:42
2025-04-01T04:35:32.025898
{ "authors": [ "Stoops-ML", "astrojuanlu" ], "repo": "poliastro/czml3", "url": "https://github.com/poliastro/czml3/pull/153", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
267876932
Simplify 2D plotting After @Juanlu001 creating the sample function, simplification of 2d plotting got significantly easier. The changes are the following: Now set_frame works also when a frame already exists, by removing all the lines and patches (not the legend), and replotting again. a list of tuples (orbit, legend) is stored to achieve the former. sample function is used in plotting, reducing code complexity. Closely related to #218 Also, there is one mypy failure and one test failure in appveyor, check them out. Notebook execution is failing in all platforms. I'm trying to know why. The problem lies in sampling an hyperbolic orbit. Right now the .sample method assumes a closed orbit, and that's a bug. Previously, the logic for sampling hyperbolic orbits was here: https://github.com/anhiga/poliastro/blob/bde55913ae3d3315e57ff1f553c9bd63de214ebe/src/poliastro/plotting.py#L170-L185 Which, by the way, should go away as well. And if that code goes away, then num_points should go away too! I tried to take over but there are a handful of errors, let's see what we can do. Well! I made some key changes and things look better now. Please take a look, and be so kind to add some tests :grin: Oops, conflicts appeared after merging #271.
gharchive/pull-request
2017-10-24T02:29:45
2025-04-01T04:35:32.030483
{ "authors": [ "Juanlu001", "anhiga" ], "repo": "poliastro/poliastro", "url": "https://github.com/poliastro/poliastro/pull/269", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
316248778
Added an option to choose propagator in OrbitPlotter Now we can choose appropriate propagator in OrbitPlotter while calling plot(). Solves #347 Just for reference It's great that we have so much flexibility, but perhaps it's too much. We are already setting the number of sampling points in the OrbitPlotter constructor, so what do you think of keeping the method as a plotter property as well? I think it's fine as then issues like the one you faced won't happen and the user will be able to use various propagators in OrbitPlotter as well.. Should I add it to OrbitPlotter constructor? But if we're setting the method per-plot instead of per-frame, what about the number of points? Shouldn't it be set per-plot as well? My concern here is consistency. Four options: We leave things as they are in this pull request. I think it's inconsistent, unless someone convinces me of the contrary. We move num_points to each plot call. I think it's bad because it makes the signature too long and, in any case, if someone is so concerned about it they can already sample in any way they want and use plot_trajectory. We move method to the OrbitPlotter constructor. This prevents the user from selecting the method per-plot. I ask: is this so important? We do nothing, and tell the user "if you want to change the propagator, just use plot_trajectory". Thoughts? I don't have much idea about consistency, but num_points is something that affects the quality of plot. Changing it can either introduce steps in the plot. I think if seems inconsistent, we can move it to the constructor. We can leave it too :P I am not much clear about this. I just remembered that, in STK, you can indeed select a different propagator for each orbit in one scenario: So I've changed my mind to accept this. ~@shreyasbapat would you like to change the propagator in the last orbit of the Florence notebook to fix it? (See the errors here http://docs.poliastro.space/en/latest/examples/Catch that asteroid!.html)~ I remember you told me you were on exams, so I'm going to merge this and take care of the rest myself. Thanks a lot @Juanlu001 . I will be actively up after 1st May Again! Between this, whenever I get time, I will try to do something!
gharchive/pull-request
2018-04-20T12:31:08
2025-04-01T04:35:32.037336
{ "authors": [ "Juanlu001", "shreyasbapat" ], "repo": "poliastro/poliastro", "url": "https://github.com/poliastro/poliastro/pull/350", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
257242926
5.9 Adecuar el login al entorno web para el administrador Es necesario adecuar el formulario de login para que también pueda ser usado de manera web por el administrador Al ejecutar el issue #96 en la que se pretende crear eventos se hizo el inicio de sesión debido a que se considero necesario para continuar con dicha tarea. Por lo anterior esta tarea puede ser cerrada.
gharchive/issue
2017-09-13T02:59:49
2025-04-01T04:35:32.039364
{ "authors": [ "DanteVDrako", "drvelasquezq" ], "repo": "polieduco/practicaaplicada20172", "url": "https://github.com/polieduco/practicaaplicada20172/issues/102", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1511697173
Уточнение типов https://github.com/polina-nn/news/blob/805447a13cea1bb97fbf977c85472a0541ee8e2c/src/EndPoints/GetAuthorsNewsSearchList.hs#L81 Лишние уточнения типов. Присутствуют очень часто. Если не уверена что тип выводится, то лучше поставь hls и уточняй там где это действительно необходимо. Добрый день! Вроде больше лишних выводов типов нет
gharchive/issue
2022-12-27T11:07:35
2025-04-01T04:35:32.040749
{ "authors": [ "pavelzarubin", "polina-nn" ], "repo": "polina-nn/news", "url": "https://github.com/polina-nn/news/issues/16", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2531553298
[Polkadot WIki Migration] Set up a RPC node This page was migrated from https://wiki.polkadot.network/docs/maintain-rpc Scan for Vale false positives. Resolved convos = made it on the update list and should go away now. I left comments for the legit flags with suggested edits. Thank you! Alrighty, made a bunch of edits. Need a new content review from @0xLucca (especially pruned nodes part as I Googled it!) and format review from @eshaben This should be good to know. Tech review was dismissed as stale so we'll need one tech and one format to merge. Thanks!
gharchive/pull-request
2024-09-17T16:09:26
2025-04-01T04:35:32.045362
{ "authors": [ "CrackTheCode016", "dawnkelly09" ], "repo": "polkadot-developers/polkadot-docs", "url": "https://github.com/polkadot-developers/polkadot-docs/pull/32", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1046835730
[feat]: implement public/private key authentication This is an issue to track the implementation of public/private key authentication for Stormi Using HMAC with a private key seems a more practical way to do this. I may consider replacing RSA with HMAC hashing or removing hashing completely. I may consider removing hashing completely and just sticking with a base64 encoded master key that would be provided by the config.yaml file. Main advantages Easier configuration process Faster authentication processing times Here's a diagram I have made the final decision to remove private/public key authentication. Users will now be authenticated via a base64 encoded user/password string.
gharchive/issue
2021-11-07T19:58:22
2025-04-01T04:35:32.072000
{ "authors": [ "michaelgrigoryan25" ], "repo": "polygon-isecure/stormi", "url": "https://github.com/polygon-isecure/stormi/issues/1", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
454217158
floor level not right on alpha5 quest Context : oculus quest Version : Alpha5 Heya I can't seem to have the floor level set right in steamVR and its games. It's like i'm sitting while actually standing. I'm sure there's a way to set that somewhere but can't figure out where Great job on the software, it works mostly great ! I would highly recommend using SteamVR Advanced Settings. There is a floor fix option in there that works great with the quest. I use it everytime. I found that by doing standing Room Setup with correct height (don't put headset on ground and do 0 as that doesn't seem to stick?) it was still basically correct the next time, maybe that will help Use this and go to the floor fix settings, super easy and works (I've used it too). https://github.com/OpenVR-Advanced-Settings/OpenVR-AdvancedSettings/releases/tag/v3.1.0 As my computer and "play room" is not on the same level, I find it hard to do the SteamVR setting as you need to follow the steps on the PC. Is there another way to do this? Advanced settings floor fix nailed it I don't remember the exact tab but on one of them there is an x y z thing that you can change, in my case i was on the floor level and not in the circle so i change the y from 0 to 1.8 (change it until you feel like you the right high) and the z to 1.5 (you may be in another place so just play with the xyz until you in the circle in the right high) I don't remember the exact tab but on one of them there is an x y z thing that you can change, in my case i was on the floor level and not in the circle so i change the y from 0 to 1.8 (change it until you feel like you the right high) and the z to 1.5 (you may be in another place so just play with the xyz until you in the circle in the right high) Yea, but I live in a contry where we use comma as decimater, so my PC does as well, so when I enter "1.8" in ALVR it changes automatically to "1,8", which only raise me "1 meter", when I then switch tab and return, it changes to "18", which sends me to the skies. This is created as an issue a long time ago, but it was never fixed, making this function useless to me @Yoshidk In Steam VR Advanced Settings you go to room adjustment (I'm pretty sure), then floor fix. You don't enter in any number. You place a controller on the floor, then click on floor fix. It will set the floor to where your lowest controller is. Just make sure before you click on fix floor that you take a good step back so when you click the button, the cameras on the quest can still see the controller on the floor. You may have to do each session, but it takes only a second to do. @Yoshidk In Steam VR Advanced Settings you go to room adjustment (I'm pretty sure), then floor fix. You don't enter in any number. You place a controller on the floor, then click on floor fix. It will set the floor to where your lowest controller is. Just make sure before you click on fix floor that you take a good step back so when you click the button, the cameras on the quest can still see the controller on the floor. You may have to do each session, but it takes only a second to do. Ok, I will have to try this. Is this "fix floor" button available after starting the SteamVR through ALVR from within the headset? Or would I still have to click somewhere on the PC? All done within the headset. Open steamvr menu, then you should see advaned settings at the bottom. Advanced settings floor fix nailed it Close the issue here please, if you are all set.
gharchive/issue
2019-06-10T15:04:54
2025-04-01T04:35:32.078706
{ "authors": [ "Myhamof", "POEDaley", "Yoshidk", "falcoriss", "iamliamiam", "machenmusik", "zxsdklfhskyrvk" ], "repo": "polygraphene/ALVR", "url": "https://github.com/polygraphene/ALVR/issues/404", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
101906199
glucose starvation vs. glucose limitation In #1687, we discussed "glucose limitation" and whether it's the same as "glucose starvation" or not. We eventually added FYPO:0003743, decreased cell population growth during glucose limitation, with a comment explaining glucose limitation as below about 4.4 mM (0.08%), but above zero. Since then, we've added three other terms using "glucose limitation" in the names, with the same comment (FYPO:0003938, FYPO:0004168, FYPO:0004765). All four terms are used in annotations. Now the question has (re-)arisen whether we should continue to make a distinction between limitation and starvation for glucose. In #2298 Antonia said: "I don't think we need to distinguish between glucose starvation and limitation, I think it is the same response." I (Midori) added the new terms requested in #2298, and am transferring the follow-up discussion of the starvation/limitation question here. From the follow-up exchange: Midori: The starvation vs limitation split comes from #1687 -- I got the impression that Val wanted to make the distinction (so she probably still does). M: #1687 includes the question about glucose limitation vs. glucose starvation, and is the one where the first glucose limitation term was added. I asked Val whether "glucose starvation" would do for her request, and if not, how to define the difference between "glucose starvation" and "glucose limitation". I don't think there's a great answer in the ticket comments. I ended up going with starvation = no glucose at all and limitation = below 4.4 mM (0.08%) but not zero, but I've been aware all along that that information really belongs in conditions. Val: I think it is important to keep the distinction but I don't know how best to do it. Happy for any solution. V: Basically the limitation was not starvation, but 4.4 mM which would be considered low glucose. I can't see the original ticket but we had a long discussion among ourselves and it was based on suggestions from the authors, who wanted us to represent that the response is different from the normal level of glucose supplied (which is really excess glucose). More transferred from #2298: Val: This is the session the term was created for http://curation.pombase.org/pombe/curs/b5db93caaa51792d/ro/ there are other genes annotated to decreased cell population growth during glucose limitation which may not be 'low glucose' as defined here, namely for gene products gpa2 sds23 ssp1 trt1 Val: ssp1 and sds23 are OK, from hanyu et al so only gpa2 and trt1 will need checking (when we decide the best way to proceed). I'd like to keep a term to represent this in some way if it is possible, as it does seem to be a specific adaptation to low glucose (probably normal in the wild when glucose is unlikely to be in such excess) More transferred from #2298: Val: Is the "adaptation to low glucose" different from the response to "no glucose"? Antonia: I can't imagine it is (surely, in no glucose, the cells would fire the "low glucose" response? i.e. a starvation response) Val: Is the "adaptation to low glucose" different from the response to "no glucose"? Antonia: I can't imagine it is (surely, in no glucose, the cells would fire the "low glucose" response? i.e. a starvation response) Val: Nope its different. See the old thread. IIRC the argument is that this low level of glucose is what fission yeast more than likely encounter in the wild. What is normally observed in laboratory setting with excess glucose is probably different from the pathways which are activated in the wild in low glucose. Val: http://www.ncbi.nlm.nih.gov/pubmed/24815688?dopt=Abstract Now commenting for myself at last ... I'm not convinced that "the old thread" (#1687) did satisfactorily answer the question of whether there's a difference between glucose limitation (aka low glucose, aka less than about 4.4 mM) and glucose starvation. I also don't see how it helps to know that laboratory media usually have lots more glucose than yeasties typically find in the wild. What is the difference in response to low glucose versus none at all? No glucose at all -> cells die after a few days "low" glucose -> cells reprogramme their gene expression profile compared to when in "high glucose". I don't like having the def in there about 4.4 mM glucose. This level is chosen because: "we previously reported that fission yeast can proliferate at essentially normal rates in media containing as little as 0.08% (4.4 mM) glucose, which is equivalent to the blood glucose level in a healthy human before breakfast" It doesn't really have anything to do with pombe in the wild (I'd guess pombe in the wild would grow on complex mixtures of different stuff, and the levels of each carbon source would fluctuate) I just want to make the distinction between "low enough glucose where the "low glucose" programme is active" & "high enough glucose so the "high glucose" programme is active" Maybe in reality there is even a grey-zone between these two states (some parts of each programme might overlap?) I don't like having the def in there about 4.4 mM glucose. I totally agree; that's a condition, not part of the phenotypes. And what I want to know is whether "low enough glucose where the "low glucose" programme is active" is the same as "glucose starvation". If there's no glucose (and no other carbon), do cells spend their remaining few days doing the same sort of gene expression as cells in low glucose? Or are there actually three situations: lots, a little, and none? It doesn't really have anything to do with pombe in the wild (I'd guess pombe in the wild would grow on complex mixtures of different stuff, and the levels of each carbon source would fluctuate) Agree. "The wild" isn't even remotely homogeneous or constant over time or space. Maybe in reality there is even a grey-zone between these two states I bet there is. This is biology, after all, so it's always safe to assume "it's more complicated than that". Umm, I'm only guessing now but I'd guess they would first go through with the normal changes (upregulate high affinity glucose (and other other carbon transporters?)), that would obviously not "work" (because there is no carbon added in the media) so maybe they would upregulate autophagy, and maybe try and mate more than cells kept at 'some' glucose level? phew, I don't think my head is getting any less melty ... In the abstract S. pombe arrests cell cycle progression when transferred from media containing 2.0% glucose to media containing 0.1%. After a delay, S. pombe resumes cell division at a surprisingly fast rate, comparable to that observed in 2% glucose. We found that a number of genes, including zinc-finger transcription factor Scr1, CaMKK-like protein kinase Ssp1, and glucose transporter Ght5, enable rapid cell division in low glucose. The suspect it is due to a shift from glycolysis to respiration to maintain ATP production despite limiting glucose I'm happy to capture this anyway, if you think that glucose limitation is inappropriate. I think it is worth capturing though. I don't think glucose limitation is inappropriate. I think that starvation = limitation. = there is no point in maintaining two separate terms Ok, it sounds like a merge OK, it sounds like Val & Antonia now agree that "glucose starvation" and "glucose limitation" are the same, at least for our phenotype annotation purposes. So what I'll do is: Standardize on "glucose starvation" in term names and "glucose limitation" in exact synonyms; Do any necessary merges; Make sure definitions are consistent; Delete comments with guff about glucose concentrations that belongs in conditions. Shout soon if that would cause problems — next FYPO session is tomorrow ;) didn't need any merges after all -- terms that used "limitation" up until a few minutes ago weren't otherwise identical to any "starvation" terms edit file: 60400ab363d2deef2a6a0f3c814fc2c7142c5d6b release: ad1a93ec3a6e446c7aa0632a50e74379931a8778
gharchive/issue
2015-08-19T14:24:24
2025-04-01T04:35:32.106409
{ "authors": [ "Antonialock", "ValWood", "mah11" ], "repo": "pombase/fypo", "url": "https://github.com/pombase/fypo/issues/2300", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
179428118
interphase and mitotic microtubule nucleation PMID:19001497 Interphase (cytoplasmic) microtubules, can be nucleated from the cytoplasm, or from the SPB. There are mutants where 'cytoplasmic microtubule nucleation" is abolished, BUT interphase microtubules are present in the cytoplasm, because they "These intranuclear MTs eventually escape into the cytoplasm, acting as progenitors of essentially all cytoplasmic MTs in the mutant cells and forming abnormal MT bundles that often curve around the cell tips" Also, cytoplasmic (astral) microtubules exist after during mitosis? so not all cytoplasmic microtubules are interphase microtubules. Soo....I think we need to make NTR abnormal interphase microtubule nucleation from cytoplasm --NTR abnormal interphase microtubule nucleation from iMTOC --NTR abnormal interphase astral microtubule nucleation (def include: occurs from cytoplasmic side of SPB) plus "abolished" and "decreased" for all and "NTR abnormal/abolished astral microtubule nucleation" --NTR abnormal interphase astral microtubule nucleation (def include: occurs from cytoplasmic side of SPB) --NTR abnormal mitotic astral microtubule nucleation (def include: occurs from cytoplasmic side of SPB) "interphase cytoplasmic microtubule nucleation" might be better wording? It's going to help a lot if you can define what is and isn't covered by "interphase microtubule". If some astral microtubules hang around after mitosis, then either (a) "interphase microtubule" includes those astral microtubules, or (b) "interphase microtubule" doesn't simply mean "a microtubule that exists during interphase". I will need to check what people refer to as interphase microtubules. @jvhayles do you know? I can ask Ken... I have altered the above request a little. We should probably stop calling them "interphase microtubules". This can be a related synonym. We should refer to "cytoplasmic microtubules during interphase" . But "Interphase cytoplasmic microtubules" can't NOT include the cytoplasmic microtubules generated from astral microtubules. Does this help: Only a small number of cytoplasmic MTs are present at any given time in fission yeast (Hoog et al., 2007), and these can be imaged by expressing GFP-tubulin at physiological concentrations. During interphase, MTs are nucleated not only from the spindle pole body but also from several interphase microtubule organizing centre ‘satellite’ particles (iMTOCs) present on the nuclear envelope, in the cytoplasm and on MTs themselves. During late mitosis, cytoplasmic astral MTs analogous to those seen in budding yeast are nucleated from the duplicated SPBs. OK, I think the existing terms can all be adjusted without much pain to "cytoplasmic microtubules [or microtubule nucleation] during interphase" - def text will need editing too, but that should be straightforward. I may have time to look at it properly tomorrow. and would this be OK for the new terms? abnormal cytoplasmic microtubule nucleation during mitotic interphase -- abnormal microtubule nucleation from iMTOC [is "during mitotic interphase" needed in the name for this one?] -- abnormal astral microtubule nucleation during mitotic interphase and abnormal astral microtubule nucleation -- abnormal astral microtubule nucleation during mitotic interphase -- abnormal astral microtubule nucleation during mitosis Don't worry if it needs to wait until you get back... Above looks perfect. No need for during interphase from iMTOC ..... Listing ALL the precise terms I need for annotation of this paper rename FYPO:0004619 cytoplasmic microtubules nucleated from equatorial microtubule organizing center absent from cell to abolished microtubule nucleation from eMTOC? NTR: abolished astral microtubule nucleation during mitosis NTR: normal cytoplasmic microtubule nucleation from the SPB during interphase NTR abolished microtubule nucleation from the iMTOC (these are interphase cytoplasmic but this should be implicit from therm) NTR: normal protein localzation to mitotic SPB NTR: abolished protein localization to eMTOC NTR: abolished protein localization to iMTOC NTR: decreased mitotic SPB oscillation NTR decreased cytoplasmic microtubule nucleation during interphase (not sure which MTOC) NTR: abolished cytoplasmic microtubule nucleation NTR: abolished cytoplasmic microtubule nucleation from the SPB during interphase I think that's it. Phew! new terms abnormal microtubule nucleation FYPO:0005692 abolished cytoplasmic microtubule nucleation FYPO:0005693 decreased cytoplasmic microtubule nucleation during mitotic interphase FYPO:0005694 abolished astral microtubule nucleation during mitosis FYPO:0005695 abolished cytoplasmic microtubule nucleation from interphase microtubule organizing center FYPO:0005696 abolished cytoplasmic microtubule nucleation from spindle pole body during mitotic interphase FYPO:0005697 abolished microtubule nucleation from equatorial microtubule organizing center FYPO:0005698 normal cytoplasmic microtubule nucleation from spindle pole body during mitotic interphase FYPO:0005699 abolished protein localization to equatorial microtubule organizing center FYPO:0005700 abolished protein localization to interphase microtubule organizing center FYPO:0005701 abnormal protein localization to microtubule organizing center FYPO:0005702 renamed existing terms from using "interphase microtubule" to "microtubule [x] during mitotic interphase"; adjusted defs accordingly edit file: 11877ed76fce9ae9583da5947f28b4c5e4ee6c6c release: db5d4e543f7f6dbe443a927feaedadd5e3158917
gharchive/issue
2016-09-27T08:29:08
2025-04-01T04:35:32.119487
{ "authors": [ "ValWood", "mah11" ], "repo": "pombase/fypo", "url": "https://github.com/pombase/fypo/issues/2833", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1330756024
This is getting slow The way I'm using Prejournal for the Ponder Source Books in practice, a localhost reset which imports 1300 movements from about 10 .pj files is already taking 2.5 minutes to execute. Maybe this is due to Postgres connections not being pooled in php -S localhost:8080 src/server.php, or maybe it's something more fundamental. In any case, this is OK if resets are rare, but if they're part of the standard workflow (make 1 small change in one .pj file, then run the reset script to see the effect) then it is already becoming unusable. :( I'll try to fix this using TigerBeetle!
gharchive/issue
2022-08-06T13:43:22
2025-04-01T04:35:32.133514
{ "authors": [ "michielbdejong" ], "repo": "pondersource/prejournal", "url": "https://github.com/pondersource/prejournal/issues/125", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
46808215
not able to setup bs grid Uncaught TypeError: Cannot read property 'error_message' of undefined jquery.bs_grid.js:1060 I am also getting this error. JS $("#demo_grid1").bs_grid({ ajaxFetchDataURL: '/Home/GetLabels', row_primary_key: "LabelTypeID", columns: [ { field: "LabelTypeID", header: "Id", visible: "no" }, { field: "Description", header: "Description" }, { field: "SizeDescription", header: "Size" }, { field: "ImageURL", header: "Image" , visible: "no" }, { field: "IsDeleted", header: "Deleted" , visible: "no" }, { field: "LabelTemplates", header: "Templates" , visible: "no" }, { field: "ChangeDate", header: "Date updated" } ], sorting: [ { sortingName: "Id", field: "LabelTypeID", order: "none" }, { sortingName: "Description", field: "Description", order: "ascending" } ], filterOptions: { filters: [] } }); Json {"total_rows":10,"page_data":[{"LabelTemplates":[],"LabelTypeID":1,"Description":"2" x 2" 7-Color DOW with Yellow-Stripe","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 2"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":2,"Description":"2" x 2" Yellow-Stripe Only","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 2"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":3,"Description":"2" x 1" 7-Color DOW with Yellow-Stripe","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":4,"Description":"2" x 1" Yellow-Stripe Only","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":5,"Description":"1" x 1" Yellow-Stripe Only","ChangeDate":"/Date(1421107200000)/","SizeDescription":"1" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":6,"Description":"1" x 1" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"1" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":7,"Description":"2" x 3" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 3"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":8,"Description":"2" x 4" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 4"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":9,"Description":"2" x 6" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 6"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":10,"Description":"2" x 8" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 8"","ImageURL":"","IsDeleted":false}]} I am also getting this error. why? Me too. No detail on the error. My JSON returned is valid. I am also having the same problem.. What can i do. Please give samples for Asp.net MVC I modify " if(filter_error["error_message"] != null) " to " if(false) ", then it is working now. But it is not a good solution, if anyone who has the right method, please share to us. Thx Hi yjd074 can you plz give the sample code.. i'll look into that
gharchive/issue
2014-10-25T10:21:15
2025-04-01T04:35:32.144097
{ "authors": [ "HarveyEV", "ManimaranUVI", "jcphlux", "juaby", "valaydesai3", "yjd074" ], "repo": "pontikis/bs_grid", "url": "https://github.com/pontikis/bs_grid/issues/13", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
460695536
Fix return type checking to allow aliasing for non-ephemeral return types. This PR fixes return type checking to allow aliasing for non-ephemeral return types. I discovered this type system issue after a discussion with @rkallos in which we discussed how to fix Map.insert (and others) to not have an unreachable error case. Looking into it further, I discovered that the type system had a simple bug that wasn't obvious due to the inconsistency in how Pony handles return type checking (the fact that using the ephemeral modifier is required to describe a unique return type, whereas on all other type specifications, the ephemeral modifier is useless). This PR doesn't address that inconsistency, because it would be a major breaking change for the language, but this does fix the small bug that went unnoticed because of it. The bug can be reproduced with the following code, in which get_1 and get_2 are valid and functionally identical, but the latter fails to compile: class iso Inner new iso create() => None class Container[A: Inner #any] var inner: A new create(inner': A) => inner = consume inner' fun get_1(): this->A => inner // works fun get_2(): this->A => let tmp = inner; consume tmp // fails actor Main new create(env: Env) => let o = Container[Inner iso](Inner) let i_1 : Inner tag = o.get_1() let i_2 : Inner tag = o.get_2() Upon investigation, it turned out to be a simple boolean logic bug, in which not (A or B) was expressed as !a || !b instead of the correct !a && !b. The code comment several lines above the change expresses the intended meaning, so I know that this fixed logic reflects the code author's original intent. Wow. Heck of a find. For posterity, I'll note that @SeanTAllen asked me to elaborate on the "inconsistency" in Zulip. You can find that conversation here: https://ponylang.zulipchat.com/#narrow/stream/189952-compiler-discussion/topic/return.20type.20ephemeral.20notation.20inconsistency The CI failure appears to be unrelated? @jemc yeah, i think so, it's been flakey lately. Okay, was going to kick off another CI build, but that'll take forever to finish - I'm just going to merge. But first, a changelog entry was added. Looks like the job was straight up cancelled. Eventually cancelled jobs will get rerun. Or should. But I've been seeing them not get rerun lately. Such is life with free stuff I guess.
gharchive/pull-request
2019-06-25T23:49:17
2025-04-01T04:35:32.149391
{ "authors": [ "SeanTAllen", "jemc" ], "repo": "ponylang/ponyc", "url": "https://github.com/ponylang/ponyc/pull/3201", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
1930280243
Build UI for Page 2 🚀 Task : Follow the UI given in Canva below and develop the UI for page 2 Use Futura Font Set for the UI GET STARTED button should hit the route /player/ @pooranjoyb hi I want to work on this issue @pooranjoyb hi I want to work on this issue Go Ahead @CS50X-RGB 🔥 Ohhk @pooranjoyb sure can u give some idea about the colors like pink and peach is it? 😅 okay sure @pooranjoyb @pooranjoyb how to get the images used there? @pooranjoyb how to get the images used there? can you provide me edit access too it would be helpful for me @CS50X-RGB Opps! Sorry that I missed that out, Here are the two gradients : #FF86C8 & #FBA277 And about the img just add a sample one, I'll co-author commits on your PR and add the original once you open it 👍 ohh thanks @pooranjoyb
gharchive/issue
2023-10-06T14:21:03
2025-04-01T04:35:32.157010
{ "authors": [ "CS50X-RGB", "pooranjoyb" ], "repo": "pooranjoyb/BeatBridge", "url": "https://github.com/pooranjoyb/BeatBridge/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1817256080
🛑 PrivilegedBA is down In 50c9b48, PrivilegedBA (https://privilegedba.com.ar) was down: HTTP code: 0 Response time: 0 ms Resolved: PrivilegedBA is back up in 70e4134.
gharchive/issue
2023-07-23T18:38:36
2025-04-01T04:35:32.235259
{ "authors": [ "porrale" ], "repo": "porrale/status", "url": "https://github.com/porrale/status/issues/262", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
993420066
[seq], opening a file while playing should restart when you open a file when reading an old file, it plays from the same spot as the old file was at. It should restart It should restart things don't really work well in MAX like that, but screw it, I'm doing this anyway, and I'm not documenting it :)
gharchive/issue
2021-09-10T16:58:08
2025-04-01T04:35:32.236455
{ "authors": [ "porres" ], "repo": "porres/pd-cyclone", "url": "https://github.com/porres/pd-cyclone/issues/559", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1373736466
Build image from uploaded file: button not enabled Bug description On the build image page the "Build the image button" button stays disabled after entering the image name, selecting Upload and uploading a file. Expected behavior The button should be enabled so the user can initiate the image build. Portainer Logs Provide the logs of your Portainer container or Service. You can see how here Steps to reproduce the issue: Go to Images > Build image Enter image name Click on "Upload" Click on "Select file", select a docker file Technical details: Portainer version: 2.15.0 Docker version (managed by Portainer): 20.10.17 Kubernetes version (managed by Portainer): n/a Platform (windows/linux): Linux Command used to start Portainer (docker run -p 9443:9443 portainer/portainer): Browser: Microsoft Edge, Chrome Use Case (delete as appropriate): Using Portainer in a Commercial setup. Have you reviewed our technical documentation and knowledge base? Yes Additional context This problem only surfaced after upgrading Portainer from v2.13.2 to v2.15.0. No problems building images using the Web editor or URL option. @duncanw I believe this issue is related to 7624 and is resolved in 2.15.1. Thanks! Thanks! Yes sorry looks like I created a duplicate issue.
gharchive/issue
2022-09-15T00:26:55
2025-04-01T04:35:32.243187
{ "authors": [ "duncanw", "tamarahenson" ], "repo": "portainer/portainer", "url": "https://github.com/portainer/portainer/issues/7666", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
71534166
feature request : Standalone Pod and white list Hello, I need : A way to configure a standalone pod which can't communicate with other pods. A way to create a white list of pods the pod is authorized to talk. Use case : I would like to install a pod for a school / a list of school. Wrong issue tracker? This isn't Diaspora* sorry, too much tabs opened...
gharchive/issue
2015-04-28T09:25:53
2025-04-01T04:35:32.253864
{ "authors": [ "Proxiweb", "posativ" ], "repo": "posativ/isso", "url": "https://github.com/posativ/isso/issues/188", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
282679145
403 Forbidden Error when posting new comments I am testing isso on a live server with a domain. After posting serval comments successfully, isso sends 403 403 FORBIDDEN status code. Here is how it looks: <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN"> <title>403 Forbidden</title> <h1>Forbidden</h1> <p>3 direct responses to /post/post-title/</p> What's the problem ? By default, isso allows only 3 direct comments to a thread. This is can be configured by setting direct-reply variable in the isso configuration.
gharchive/issue
2017-12-17T08:43:49
2025-04-01T04:35:32.255312
{ "authors": [ "noisytoken" ], "repo": "posativ/isso", "url": "https://github.com/posativ/isso/issues/364", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
528418847
minor grammar improvements to quickstart.rst Very minor changes to improve grammar and fix two misspelled word. Closing, this is being tracked in https://github.com/posativ/isso/pull/590
gharchive/pull-request
2019-11-26T00:42:07
2025-04-01T04:35:32.256440
{ "authors": [ "jasdeepgill", "jelmer" ], "repo": "posativ/isso", "url": "https://github.com/posativ/isso/pull/592", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
688318888
smtp only has 2 instances (should have 1) * smtp.2 is not running (Failed) postal I have recently installed postal mail server on my VPS. After that when I am trying to connect to my sever using smtp credentials i am not able to do so. I am getting error. After doing 'postal restart' I got the result given below. I checked my netstat -lnp and the result is also given below. please help me fix the issue. need help @willpower232 root@postal:~# postal restart Stopped smtp.2 Restarted web.1 Restarted worker.4 -> worker.5 Restarted cron.4 -> cron.5 Restarted smtp.3 Restarted requeuer.4 -> requeuer.5 root@postal:~# netstat -lnp Active Internet connections (only servers) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:80 0.0.0.0:* LISTEN 381/nginx: master p tcp 0 0 0.0.0.0:4369 0.0.0.0:* LISTEN 441/epmd tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN 360/sshd tcp 0 0 127.0.0.1:25 0.0.0.0:* LISTEN 1024/exim4 tcp 0 0 0.0.0.0:443 0.0.0.0:* LISTEN 381/nginx: master p tcp 0 0 127.0.0.1:5000 0.0.0.0:* LISTEN 1282/[postal] web.1 tcp 0 0 0.0.0.0:25672 0.0.0.0:* LISTEN 296/beam.smp tcp 0 0 127.0.0.1:3306 0.0.0.0:* LISTEN 575/mysqld tcp6 0 0 :::80 :::* LISTEN 381/nginx: master p tcp6 0 0 :::4369 :::* LISTEN 441/epmd tcp6 0 0 :::22 :::* LISTEN 360/sshd tcp6 0 0 ::1:25 :::* LISTEN 1024/exim4 tcp6 0 0 :::443 :::* LISTEN 381/nginx: master p tcp6 0 0 :::5672 :::* LISTEN 296/beam.smp Active UNIX domain sockets (only servers) Proto RefCnt Flags Type State I-Node PID/Program name Path unix 2 [ ACC ] STREAM LISTENING 692941580 1/init /run/systemd/journal/stdout unix 2 [ ACC ] STREAM LISTENING 692944182 276/saslauthd /var/run/saslauthd/mux unix 2 [ ACC ] STREAM LISTENING 693546224 1276/[procodile] Po /tmp/postal/pids/procodile.sock unix 2 [ ACC ] SEQPACKET LISTENING 692947316 1/init /run/udev/control unix 2 [ ACC ] STREAM LISTENING 692929142 1/init /var/run/dbus/system_bus_socket unix 2 [ ACC ] STREAM LISTENING 692949172 1/init /run/systemd/private unix 2 [ ACC ] STREAM LISTENING 692892378 575/mysqld /var/run/mysqld/mysqld.sock You are running another mail server software "Exim" using same SMTP port as Postal. You need to stop the Exim service and probably disable it, then issue a "postal restart". Thank you for your reply @yomiit @yomiit when i did #service exim stop it gives me the result below root@postal:~# service exim stop Failed to stop exim.service: Unit exim.service not loaded. I have installed postal using "curl https://raw.githubusercontent.com/atech/postal/master/script/install/ubuntu1604.sh | sh " this command. I am running ununtu 18.04. I have setup firewall using ufw. ufw status is given below root@postal:~# ufw status Status: active To Action From 25/tcp ALLOW Anywhere 465/tcp ALLOW Anywhere 587/tcp ALLOW Anywhere 22/tcp ALLOW Anywhere 80/tcp ALLOW Anywhere 443/tcp ALLOW Anywhere 25/tcp (v6) ALLOW Anywhere (v6) 465/tcp (v6) ALLOW Anywhere (v6) 587/tcp (v6) ALLOW Anywhere (v6) 22/tcp (v6) ALLOW Anywhere (v6) 80/tcp (v6) ALLOW Anywhere (v6) 443/tcp (v6) ALLOW Anywhere (v6) The service should be exim4. systemctl stop exim4 systemctl disable exim4 hello. thanks for the help. Now postal is listening to port 25, and only listening to port 25. I can not connect to postal using 465 & 587, need help with that. plz @willpower232 @yomiit Connecting on other ports shouldn't be required but if you want it, you will have to use something like iptables to forward ports internally. Connecting on other ports shouldn't be required but if you want it, you will have to use something like iptables to forward ports internally. @willpower232 plz help me to do it, as this is my first time doing this.
gharchive/issue
2020-08-28T19:46:46
2025-04-01T04:35:32.279474
{ "authors": [ "samg8520", "willpower232", "yomiit" ], "repo": "postalhq/postal", "url": "https://github.com/postalhq/postal/issues/1193", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1999476558
Corrected a typo in Readme file Please review this typo correction. Sure let me do that! thank you 👍 Closes #59
gharchive/pull-request
2023-11-17T16:16:32
2025-04-01T04:35:32.296776
{ "authors": [ "LKoech", "loopDelicious" ], "repo": "postmanlabs/pmquickstarts", "url": "https://github.com/postmanlabs/pmquickstarts/pull/61", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
177254822
Installation: 301 on cookbooks installation Hi! I am trying to set up a development server of my own. I am now on step 5 and I get a following error: [nina@leafblade tomatoes]$ librarian-chef install Could not cache bluepill/1.0.6 <http://community.opscode.com/api/v1> from http://community.opscode.com/api/v1/cookbooks/bluepill because 301 Moved Permanently! Can I do something to bypass this error? @matteodepalo is the one who setup that process in the first place, let's see if he can help you solving this issue. @Liarra @potomak the Cheffile was updated 4 years ago so some of the libraries might have disappeared or they might have been moved to other places. To fix this we would need to look at the git urls in the Cheffile and update all of them to the latest versions. Also check if there are greater versions for site 'http://community.opscode.com/api/v1'. Try bumping that version to the latest one and see what happens. @matteodepalo can you take a quick look at it and remove cheffile + doc if it's not usable anymore or update it please? @potomak it would take a while to update the Cheffile and the recipes to the latest version. I would personally remove the files as I don't have the time to update them unfortunately. Ok, I'll add a task to remove them. Thanks.
gharchive/issue
2016-09-15T18:37:35
2025-04-01T04:35:32.366246
{ "authors": [ "Liarra", "matteodepalo", "potomak" ], "repo": "potomak/tomatoes", "url": "https://github.com/potomak/tomatoes/issues/143", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2521891478
Please consider fixing issues detected by repository checker Notification from ioBroker Check and Service Bot Dear adapter developer, I'm the ioBroker Check and Service Bot. I'm an automated tool processing routine tasks for the ioBroker infrastructure. I have recently checked the repository for your adapter warp for common errors and appropiate suggestions to keep this adapter up to date. This check is based the current head revisions (master / main branch) of the adapter repository Please see the result of the check below. ioBroker.warp - ERRORS: [ ] :heavy_exclamation_mark: [E026] "{'engines': {'node'>='16'}}" is required at package.json, "{'engines':{'node'>='18'}}" is recommended [ ] :heavy_exclamation_mark: [E033] @iobroker/adapter-core 2.6.0 specified. 3.1.4 is required as minimum, 3.1.6 is recommended. Please update dependencies at package.json [ ] :heavy_exclamation_mark: [E036] @iobroker/testing 3.0.2 specified. 4.1.3 is required as minimum, 4.1.3 is recommended. Please update devDependencies at package.json [ ] :heavy_exclamation_mark: [E157] js-controller 2.0.0 listed as dependency but 4.0.24 is required as minimum, 5.0.19 is recommended. Please update dependency at io-package.json. WARNINGS: [ ] :eyes: [W040] "keywords" within package.json should contain "ioBroker" [ ] :eyes: [W127] Missing suggested translation into uk of "common.titleLang" in io-package.json. [ ] :eyes: [W132] Many "common.news" found in io-package.json. Repository builder will truncate at 7 news. Please remove old news. [ ] :eyes: [W134] Missing suggested translation into uk of "common.desc" in io-package.json. [ ] :eyes: [W135] "common.tier" is required in io-package.json. Please check https://github.com/ioBroker/ioBroker.docs/blob/master/docs/en/dev/objectsschema.md#adapter. [ ] :eyes: [W154] Missing suggested translation into uk of some "common.news" in io-package.json. [ ] :eyes: [W181] "common.license" in io-package.json is deprecated. Please define object "common.licenseInformation" [ ] :eyes: [W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Use "main" at package.json instead. [ ] :eyes: [W184] "common.materialize" is deprecated for admin >= 5 at io-package.json. Please use property "adminUI". [ ] :eyes: [W184] "common.title" is deprecated and replaced by "common.titleLang". Please remove from io-package.json. SUGGESTIONS: [ ] :pushpin: [S522] Please consider migrating to admin 5 UI (jsonConfig). Please review issues reported and consider fixing them as soon as appropiate. Errors reported by repository checker should be fixed as soon as possible. Some of them require a new release to be considered as fixed. Please note that errors reported by checker might be considered as blocking point for future updates at stable repository. Warnings reported by repository checker should be reviewed. While some warnings can be ignored due to good reasons or a dedicated decision of the developer, most warnings should be fixed as soon as appropiate. Suggestions reported by repository checker should be reviewed. Suggestions can be ignored due to a decision of the developer but they are reported as a hint to use a configuration which might get required in future or at least is used be most adapters. Suggestions are always optional to follow. You may start a new check or force the creation of a new issue at any time by adding the following comment to this issue: @iobroker-bot recheck @iobroker-bot recreate Please note that I (and the server at GitHub) have always plenty of work to do. So it may last up to 30 minutes until you see a reaction. I will drop a comment here as soon as I start processing. Feel free to contact me (@iobroker-bot) if you have any questions or feel that an issue is incorrectly flagged. And THANKS A LOT for maintaining this adapter from me and all users. Let's work together for the best user experience. your ioBroker Check and Service Bot @mcm1957 for evidence Last update at Fri, 13 Sep 2024 14:28:45 GMT based on commit 0d1addd26d6ba16510076e1d09edb8258aa6df0f ioBroker.repochecker 3.0.5 This issue has been updated by ioBroker Check and Service Bot The following issues have been fixed [W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Use "main" at package.json instead. :thumbsup:Thanks for fixing the issues. The following issues are new and have been added [W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Executable is defined by entry "main" at package.json.
gharchive/issue
2024-09-12T09:43:29
2025-04-01T04:35:32.385451
{ "authors": [ "ioBroker-Bot" ], "repo": "pottio/ioBroker.warp", "url": "https://github.com/pottio/ioBroker.warp/issues/134", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2063163496
🛑 PowBot SDN is down In ca5117d, PowBot SDN (https://api.powbot.org/products) was down: HTTP code: 502 Response time: 658 ms Resolved: PowBot SDN is back up in 3d4e750 after 9 minutes.
gharchive/issue
2024-01-03T02:31:02
2025-04-01T04:35:32.406599
{ "authors": [ "constt" ], "repo": "powbot/status", "url": "https://github.com/powbot/status/issues/279", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
845088230
Update kubernetes/master/golang/master build This is an automated PR via build-bot [APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: ltccci To complete the pull request process, please assign after the PR has been reviewed. You can assign the PR to them by writing /assign in a comment when ready. The full list of commands accepted by this bot can be found here. Needs approval from an approver in each of these files: Approvers can indicate their approval by writing /approve in a comment Approvers can cancel approval by writing /approve cancel in a comment
gharchive/pull-request
2021-03-30T19:08:36
2025-04-01T04:35:32.443153
{ "authors": [ "ltccci" ], "repo": "ppc64le-cloud/builds", "url": "https://github.com/ppc64le-cloud/builds/pull/1857", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2279737843
How to find the number of records in a relation? I'm trying to use pagination to display product cards. However I have a connection Typeorm as: Products <-> ​accounting ProductsEntity: @OneToMany(() => AccountingEntity, (accounting) => accounting.product, {}) accounting: AccountingEntity[]; AccountingEntity: @ManyToOne(() => ProductEntity, (product) => product.accounting, { onDelete: 'CASCADE', }) @JoinColumn({ name: 'product_id' }) product: ProductEntity; I display my product cards, but I can’t do it to show only those records in which there is NO accounting! or vice versa there is only ACCOUNTING How can I do this correctly? My code: const result: any = await paginate(query, this.productRepository, { sortableColumns: ['uuid', 'name', 'inventory', 'series', 'count', 'accounting', 'author.last_name',], relations: ['accounting', 'author', 'user'], nullSort: 'last', defaultSortBy: [['uuid', 'DESC']], searchableColumns: ['uuid', 'name', 'count', 'author.last_name', 'author.first_name', 'author.surname', 'user.last_name', 'user.first_name', 'user.surname', 'series', 'inventory',], filterableColumns: { count: [FilterOperator.EQ, FilterOperator.BTW], accounting: true, author: true, user: true }, }); bump I'm trying to use pagination to display product cards. However I have a connection Typeorm as: Products <-> ​accounting ProductsEntity: @OneToMany(() => AccountingEntity, (accounting) => accounting.product, {}) accounting: AccountingEntity[]; AccountingEntity: @ManyToOne(() => ProductEntity, (product) => product.accounting, { onDelete: 'CASCADE', }) @JoinColumn({ name: 'product_id' }) product: ProductEntity; I display my product cards, but I can’t do it to show only those records in which there is NO accounting! or vice versa there is only ACCOUNTING How can I do this correctly? My code: const result: any = await paginate(query, this.productRepository, { sortableColumns: ['uuid', 'name', 'inventory', 'series', 'count', 'accounting', 'author.last_name',], relations: ['accounting', 'author', 'user'], nullSort: 'last', defaultSortBy: [['uuid', 'DESC']], searchableColumns: ['uuid', 'name', 'count', 'author.last_name', 'author.first_name', 'author.surname', 'user.last_name', 'user.first_name', 'user.surname', 'series', 'inventory',], filterableColumns: { count: [FilterOperator.EQ, FilterOperator.BTW], accounting: true, author: true, user: true }, }); You can achieve this using where in pagination config
gharchive/issue
2024-05-05T21:20:57
2025-04-01T04:35:32.447493
{ "authors": [ "MrSharpp", "kenvals", "ppetzold" ], "repo": "ppetzold/nestjs-paginate", "url": "https://github.com/ppetzold/nestjs-paginate/issues/916", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1338007729
是否有人曾提过类似的问题? 否(No) 你觉得APP有什么不足之处? no 你觉得该怎么去完善会比较好?【非必答】 No response 等下一个版本发布 或 下载每周构建版本
gharchive/issue
2022-08-13T16:53:04
2025-04-01T04:35:32.460972
{ "authors": [ "daidaojianke", "pppscn" ], "repo": "pppscn/SmsForwarder", "url": "https://github.com/pppscn/SmsForwarder/issues/204", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
651552045
Fix setting GridContainer content by index resolves #3657 The problem is that an array does not notify about changing its elements. This PR adds a GridContainerContent class which wraps a jagged array and provides notifications for its elements changes. Awesome, but this is a pretty large breaking change. Is it possible to have an implicit operator conversion from Drawable[][] to GridContainerContent? Is it possible to have an implicit operator conversion from Drawable[][] to GridContainerContent? AFAIK it is not possible to convert an array to a custom object. However, I can write an extension method in order to make code a little cleaner. Vice-versa would also be good to have (GridContainerContent -> Drawable[][], even if it requires allocs). We can expose the original array via method or property. See https://github.com/UselessToucan/osu-framework/commit/f5b3f9205a8beaa847e392563fc4342703eb2e3d. AFAIK it is not possible to convert an array to a custom object. However, I can write an extension method in order to make code a little cleaner. I think the suggestion was to add operators like public static implicit operator Drawable[][](GridContainerContent content); public static implicit operator GridContainerContent(Drawable[][] drawables); and it seems to at least not throw syntax errors and build? This way the changes to Content should no longer be breaking, or am I missing something? Unfortunately this is still a breaking change, albeit a slightly less breaking one - game-side there's this usage that needs adjusting for.
gharchive/pull-request
2020-07-06T13:44:53
2025-04-01T04:35:32.467703
{ "authors": [ "UselessToucan", "bdach", "smoogipoo" ], "repo": "ppy/osu-framework", "url": "https://github.com/ppy/osu-framework/pull/3692", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1345879912
Fix potential startup crash if resolution/display pair from configuration results in no valid resolution Closes https://github.com/ppy/osu/issues/19877#issuecomment-1221438234. I wasn't able to reproduce what the user was seeing even when trying my best to break it, but have tested the flow works by artifically adding an exception: diff --git a/osu.Framework/Platform/SDL2DesktopWindow.cs b/osu.Framework/Platform/SDL2DesktopWindow.cs index 8ca11367f..d68d8368f 100644 --- a/osu.Framework/Platform/SDL2DesktopWindow.cs +++ b/osu.Framework/Platform/SDL2DesktopWindow.cs @@ -1233,6 +1233,9 @@ private void updateWindowStateAndSize() { var closestMode = getClosestDisplayMode(sizeFullscreen.Value, currentDisplayMode.Value.RefreshRate, currentDisplay.Index); + if (!sizeFullscreen.IsDefault) + throw new InvalidOperationException("test"); + Size = new Size(closestMode.w, closestMode.h); SDL.SDL_SetWindowDisplayMode(SDLWindowHandle, ref closestMode); #5649
gharchive/pull-request
2022-08-22T06:21:53
2025-04-01T04:35:32.469538
{ "authors": [ "peppy" ], "repo": "ppy/osu-framework", "url": "https://github.com/ppy/osu-framework/pull/5371", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2588017357
tests script does not work outside of nix environment When I try to run tests func locally (in a python environment with click installed), I get: (venv) > mlkem-c-aarch64 % tests func INFO > Functional Test INFO > make CROSS_PREFIX= mlkem AUTO=1 OPT=1 make: Circular test/build/lib/libfips202.a <- test/build/lib/libfips202.a dependency dropped. make: Circular test/build/lib/librng.a <- test/build/lib/libfips202.a dependency dropped. make: Circular test/build/lib/librng.a <- test/build/lib/librng.a dependency dropped. make: Circular test/build/randombytes/randombytes.c.o <- test/build/lib/libfips202.a dependency dropped. make: Circular test/build/randombytes/randombytes.c.o <- test/build/lib/librng.a dependency dropped. ld: warning: search path 'test/build/lib' not found ld: library 'fips202' not found which hints at some issue with the Makefile. The same problem arises with tests kat and tests nistkat. After https://github.com/pq-code-package/mlkem-c-aarch64/pull/241, the following problems emerge: mlkem-c-aarch64 % tests kat INFO > Kat Test INFO > make CROSS_PREFIX= kat OPT=1 AUTO=1 INFO > ./test/build/mlkem512/bin/gen_KAT512 ERROR > ML-KEM-512 failed, expecting , but getting 323d0e9aefe34819f10cdce1f0d9e5c8a55193cebe984fb1718e779ebfbc0da8 INFO > ./test/build/mlkem768/bin/gen_KAT768 ERROR > ML-KEM-768 failed, expecting , but getting 99b497dcddfe418f44d30c7376fda09ae7cca2e9141143032d842508b4a1f438 INFO > ./test/build/mlkem1024/bin/gen_KAT1024 ERROR > ML-KEM-1024 failed, expecting , but getting 104058bab1fef70aa10606831faabef7053d44b1adac6b34d35e505c3085db78 and > ./test/build/mlkem512/bin/test_kyber512 ERROR > ML-KEM-512 failed, expecting CRYPTO_SECRETKEYBYTES: CRYPTO_PUBLICKEYBYTES: CRYPTO_CIPHERTEXTBYTES: , but getting CRYPTO_SECRETKEYBYTES: 1632 CRYPTO_PUBLICKEYBYTES: 800 CRYPTO_CIPHERTEXTBYTES: 768 INFO > ./test/build/mlkem768/bin/test_kyber768 ERROR > ML-KEM-768 failed, expecting CRYPTO_SECRETKEYBYTES: CRYPTO_PUBLICKEYBYTES: CRYPTO_CIPHERTEXTBYTES: , but getting CRYPTO_SECRETKEYBYTES: 2400 CRYPTO_PUBLICKEYBYTES: 1184 CRYPTO_CIPHERTEXTBYTES: 1088 INFO > ./test/build/mlkem1024/bin/test_kyber1024 ERROR > ML-KEM-1024 failed, expecting CRYPTO_SECRETKEYBYTES: CRYPTO_PUBLICKEYBYTES: CRYPTO_CIPHERTEXTBYTES: , but getting CRYPTO_SECRETKEYBYTES: 3168 CRYPTO_PUBLICKEYBYTES: 1568 CRYPTO_CIPHERTEXTBYTES: 1568 and % tests nistkat INFO > Nistkat Test INFO > make CROSS_PREFIX= nistkat AUTO=1 OPT=1 INFO > ./test/build/mlkem512/bin/gen_NISTKAT512 ERROR > ML-KEM-512 failed, expecting , but getting a30184edee53b3b009356e1e31d7f9e93ce82550e3c622d7192e387b0cc84f2e INFO > ./test/build/mlkem768/bin/gen_NISTKAT768 ERROR > ML-KEM-768 failed, expecting , but getting 729367b590637f4a93c68d5e4a4d2e2b4454842a52c9eec503e3a0d24cb66471 INFO > ./test/build/mlkem1024/bin/gen_NISTKAT1024 ERROR > ML-KEM-1024 failed, expecting , but getting 3fba7327d0320cb6134badf2a1bcb963a5b3c0026c7dece8f00d6a6155e47b33 Fixed by https://github.com/pq-code-package/mlkem-c-aarch64/pull/243
gharchive/issue
2024-10-15T08:23:44
2025-04-01T04:35:32.533841
{ "authors": [ "hanno-becker" ], "repo": "pq-code-package/mlkem-c-aarch64", "url": "https://github.com/pq-code-package/mlkem-c-aarch64/issues/240", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1562993866
fix: improve types While the files property can also be populated with new files, it is advised to add new files using only the addFile and addFiles methods. When using React or other cases, such as a quickstart example: const [files, setFiles] = useState<FilePondFile[]>([]); <FilePond files={files} onupdatefiles={(files) => setFiles(files)} /> This will throw a type error because: files?: Array<FilePondInitialFile | ActualFileObject | Blob | string>; onupdatefiles?: ((files: FilePondFile[]) => void) This is an issue we are dealing with too
gharchive/pull-request
2023-01-30T18:38:50
2025-04-01T04:35:32.537400
{ "authors": [ "CarelessCourage", "songhn233" ], "repo": "pqina/filepond", "url": "https://github.com/pqina/filepond/pull/887", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1119393251
Feature - Reverse Transliteration Is there any plan for including Hindi-Hinglish transliteration (the reverse of what the library has currently)? Hey, as of now we don't have any plans for the same. To my knowledge, I also think that datasets for reverse transliteration are extremely scarce. I wasn't able to find anything relevant on the internet either after skimming briefly.
gharchive/issue
2022-01-31T12:56:13
2025-04-01T04:35:32.541296
{ "authors": [ "hetarth18", "praatibhsurana" ], "repo": "praatibhsurana/Hinglish_Hindi_WSD", "url": "https://github.com/praatibhsurana/Hinglish_Hindi_WSD/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
653986503
Criar esboço técnico de uma sala de telepresença O documento pode ter uma única página que liste os itens de uma sala de telepresença e a suas interações entre si. Além disso, o documento deve conter uma ilustração de uma sala de aula e onde os elementos estarão posicionados. Os itens são os seguintes: Central (Raspberry PI + câmera embutida + microfone condensador apontado para frente) fixado no teto da sala de aula, no centro, com a câmera apontada para o quadro/docente. TV de 50 polegadas (fixada na parede do fundo da sala de aula, no meio dela, conecatada à central por um cabo HDMI de 10m). Cabo de rede que liga a central até o ponto de rede mais próximo. Eu estou pedindo pro setor de obras o layout das salas de aula (creio que sejam iguais nos 6 campi) para pensarmos já espacialmente de acordo com as dimensões das salas. O esquema está disponível em alta resolução no drive em programa#57. Abaixo uma ilustração:
gharchive/issue
2020-07-09T11:28:27
2025-04-01T04:35:32.566468
{ "authors": [ "Dovyski", "phdmauricio" ], "repo": "practice-uffs/programa", "url": "https://github.com/practice-uffs/programa/issues/57", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1055518622
Qual a estrutura do curso? Derivada da macro tarefa AUR-1.3. Esta micro tarefa depende da conclusão das micro tarefas de 41 a 63. Atentem para o fato que a pergunta é um norteador, o responsável pela tarefa deve pensar em variações pertinentes daquela pergunta. Limitar-se a colocar as informações da pergunta da tarefa. Outras perguntas (ou informações novas) serão feitas em outras fases do projeto. Na macro tarefa AUR-1.3 deve-se elencar e catalogar (na planilha da Aura) quais são as informações relevantes do item do conjunto: Cursos da UFFS (graduação e pós-graduação). ESSA ISSUE REQUER UMA RESPOSTA PARA CADA CURSO DE PÓS-GRADUAÇÃO DA UFFS Link para a planilha onde devem ser alocadas as informações AS RESPOSTAS DEVEM SER ALOCADAS NA PLANILHA DO LINK E NÃO EM PLANILHAS SUBSEQUENTES TENDO EM VISTA A ORGANIZAÇÃO DAS TAREFAS E A REVISÃO DAS MESMAS. Criei as pastas dessa issue no Google Drive: Pasta da issue 🔽 Entrada 🔼 Saída Essa issue não é mais válida!
gharchive/issue
2021-11-16T23:07:03
2025-04-01T04:35:32.570748
{ "authors": [ "Claudineia-VR", "PracticeUFFSBot", "morgana-mo" ], "repo": "practice-uffs/programa", "url": "https://github.com/practice-uffs/programa/issues/923", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
340145002
Optout confirmation This PR adds a confirmation message when the optout button is pressed asking the user to confirm to minimise mistakes. I also fixed the problem where you need to refresh to get the optout button to hide after a successful optout and I made the extra info on the address look better. I did everything in separate commits for easy reviewing. +1
gharchive/pull-request
2018-07-11T08:40:10
2025-04-01T04:35:32.575001
{ "authors": [ "DevChima", "erikh360" ], "repo": "praekelt/seed-control-interface", "url": "https://github.com/praekelt/seed-control-interface/pull/71", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
167587111
Only use decimal digits in MTN Nigeria XML over TCP transport session IDs. Currently we use hexadecimal digits, but non-decimal digits are not allowed. Anyone have any opinions on using: session_id = "".join(random.choice("0123456789") for i in range(SESSION_ID_LENGTH)) as a method for generating a session id? :+1: when travis is happy I've confirmed that all the Travis errors are covered by #1049 and that the tests pass locally, so landing this.
gharchive/pull-request
2016-07-26T11:43:31
2025-04-01T04:35:32.576689
{ "authors": [ "hodgestar", "justinvdm" ], "repo": "praekelt/vumi", "url": "https://github.com/praekelt/vumi/pull/1050", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1794997832
Adding a web development project: Word Counter [gssoc23] Title and Issue number Web development project: Word Counter [gssoc23] Please add the tag of gssoc23 Close #3065 Checklist: [X] I have mentioned the issue number in my Pull Request. [X] I have commented my code, particularly in hard-to-understand areas [X] I have created a helpful and easy to understand README.md @MohitGupta121 @pranjay-poddar Please review this PR.
gharchive/pull-request
2023-07-08T15:56:08
2025-04-01T04:35:32.598206
{ "authors": [ "thevirengarg" ], "repo": "pranjay-poddar/Dev-Geeks", "url": "https://github.com/pranjay-poddar/Dev-Geeks/pull/3263", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
265033453
[issue-32] Table Connector Change log description add FlinkPravegaTableSource add FlinkPravegaTableSink add FlinkTableITCase add JSON serialization schema add jackson as a 'provided' (by Flink) dependency Purpose of the change Introduces Flink Table API support, including a table source and append-only table sink. Closes #32 . How to verify it Execute integration test FlinkTableITCase. Note that the JsonRowDeserializationSchema and JsonRowSerializationSchema classes were copied from the Flink Kafka connector. I'll follow up with the original authors as to whether they could be moved to the flink-table library. Unsure as to whether any attribution is needed. Providing a JsonSerializationSchema makes sense here. We may want to take the next step and also provide a Pravega io.pravega.client.stream.Serializer that works with Json. That way someone can write a standalone app to write the data and read it using the table connector. Extending that, if we have a JsonSerializer for Pravega, perhaps we should use the PravegaDeserializationSchema that already exists and use the JsonSerializer in conjunction with that for the table connector. I think this serializer may be too specialized to be replaced with a generic Pravega serializer since it is coded for the Row class and not based on Jackson databinding. I agree completely that the Pravega client should provide a Json serializer based on data binding. We could provide something in the samples. Will file. On Oct 12, 2017 11:17 AM, "Chris Dail" notifications@github.com wrote: Providing a JsonSerializationSchema makes sense here. We may want to take the next step and also provide a Pravega io.pravega.client.stream. Serializer that works with Json. That way someone can write a standalone app to write the data and read it using the table connector. Extending that, if we have a JsonSerializer for Pravega, perhaps we should use the PravegaDeserializationSchema that already exists and use the JsonSerializer in conjunction with that for the table connector. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/pravega/flink-connectors/pull/61#issuecomment-336222358, or mute the thread https://github.com/notifications/unsubscribe-auth/ABsXno8w95j-xUBoqsC_4Ovx1tl_2Piiks5srlfVgaJpZM4P3Zds . @tzulitai I invited you as a collaborator, then I can request a review from you.
gharchive/pull-request
2017-10-12T18:01:23
2025-04-01T04:35:33.977955
{ "authors": [ "EronWright", "chrisdail" ], "repo": "pravega/flink-connectors", "url": "https://github.com/pravega/flink-connectors/pull/61", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
822027765
Document demand chain objects Documents https://github.com/prebid/Prebid.js/pull/6383 and https://github.com/prebid/Prebid.js/pull/6380 Will be part of 4.30
gharchive/pull-request
2021-03-04T11:11:16
2025-04-01T04:35:34.059533
{ "authors": [ "bretg", "patmmccann" ], "repo": "prebid/prebid.github.io", "url": "https://github.com/prebid/prebid.github.io/pull/2737", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
920688261
Farm HNZ Farm in NC it went down because gateway was knock over by something (presumably a person with a baseball) gateway needs to be replaced. gateway got full of water. Baseball is a fun plot twist! Anna Paulk Birnbaum Agricultural Specialist UGA-Department of Crop and Soil Sciences 1420 Experiment Station Rd Watkinsville, GA 30677 C: 404-889-1688 From: elhenriq @.> Sent: Monday, June 14, 2021 2:56 PM To: precision-sustainable-ag/On-farm-Protocols @.> Cc: Subscribed @.***> Subject: [precision-sustainable-ag/On-farm-Protocols] Farm HNZ (#105) [EXTERNAL SENDER - PROCEED CAUTIOUSLY] Farm in NC it went down because gateway was knock over by something (presumably a person with a baseball) gateway needs to be replaced. gateway got full of water. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHubhttps://github.com/precision-sustainable-ag/On-farm-Protocols/issues/105, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APL6HJ2KXK2AHRIHMWG34L3TSZGFZANCNFSM46VZCVJQ. There was an actual baseball next to it in the field WOW someone REALLY doesn’t like the sensors Anna Paulk Birnbaum Agricultural Specialist UGA-Department of Crop and Soil Sciences 1420 Experiment Station Rd Watkinsville, GA 30677 C: 404-889-1688 From: elhenriq @.> Sent: Monday, June 14, 2021 3:16 PM To: precision-sustainable-ag/On-farm-Protocols @.> Cc: Anna Paulk Birnbaum @.>; Comment @.> Subject: Re: [precision-sustainable-ag/On-farm-Protocols] Farm HNZ (#105) [EXTERNAL SENDER - PROCEED CAUTIOUSLY] There was an actual baseball next to it in the field [image] https://user-images.githubusercontent.com/52674578/121946979-82fb7880-cd23-11eb-9b8c-ed02782eafc0.jpeg [image] https://user-images.githubusercontent.com/52674578/121946982-84c53c00-cd23-11eb-83e3-bf0404519b7d.jpeg — You are receiving this because you commented. Reply to this email directly, view it on GitHubhttps://github.com/precision-sustainable-ag/On-farm-Protocols/issues/105#issuecomment-860929646, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APL6HJYQOGZ5OLQOVKUJMKTTSZISXANCNFSM46VZCVJQ.
gharchive/issue
2021-06-14T18:56:13
2025-04-01T04:35:34.072017
{ "authors": [ "annabirnbaum", "elhenriq" ], "repo": "precision-sustainable-ag/On-farm-Protocols", "url": "https://github.com/precision-sustainable-ag/On-farm-Protocols/issues/105", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2185684636
Want Lorax with newer version of TGI Feature request hello,our models are deploying with TGI(v1.4.3), and we alse want to use lorax. But I find that the tgi version lorax is based on is very different with TGI version v1.4.3。 We are trying to integrate lorax(v0.8) into TGI(v1.4.3)。Is there possible to upgrade TGI of lorax or contribute lorax to TGI? Motivation use new features of TGI together with lorax Your contribution We are trying to integrate lorax(v0.8) into TGI(v1.4.3), but both lorax and tgi are changing! Hi @yangelaboy, thanks for trying out LoRAX. I'd love to incorporate more upstream work from TGI, but since they changed their license last year, we can no longer pull their code into our repo. That said, we have implemented many of the same features recently (though in slightly different ways). Are there specific features you're using in TGI you want to see in LoRAX? If so, we can definitely prioritize getting those added. One thing in TGI we're working to add very soon is speculative decoding. We think our implementation will be particularly interesting, as we'll be able to handle multiple speculation models at once. Let me know if there are other features you're interested in. @tgaddair Thinks for detailed replies. We are using features such as speculative decoding(ngram&medusa), quantization, also we're interested in much optimizations of TGI. We also added functions in TGI like shared prefix prompt cache。 Finally, We want a framework which can support different adapter models and medusa models in same self-trained model with a shared prefix prompt cache. I will pay attention to Lorax. Hey @yangelaboy, thanks for this context! The good news is all of the things you listed are on our near-term roadmap. Speculative decoding adapters per request - this is what I'm currently working on and hope to have out next week Prefix caching - this is the next major item on the roadmap after speculative decoding, so hopefully a few weeks away at most Quantization - we support a number of quantization options currently, but let me know if there are specifics ones we don't support that you would be interested in. I'll definitely let you know when the speculative decoding is ready to test out! Thanks @tgaddair , we are also waiting for the Speculative decoding 👍 The license is back to Apache-2.0 https://github.com/huggingface/text-generation-inference/commit/ff42d33e9944832a19171967d2edd6c292bdb2d6 @tgaddair
gharchive/issue
2024-03-14T07:59:57
2025-04-01T04:35:34.078417
{ "authors": [ "abhibst", "giyaseddin", "tgaddair", "yangelaboy" ], "repo": "predibase/lorax", "url": "https://github.com/predibase/lorax/issues/329", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2398105992
Add pixi entry to tldr-pages Problem description Would be cool if pixi had an entry in tldr pages: https://github.com/tldr-pages/tldr/issues/13248 Sweet, this was already implemented!
gharchive/issue
2024-07-09T12:41:04
2025-04-01T04:35:34.082337
{ "authors": [ "corneliusroemer" ], "repo": "prefix-dev/pixi", "url": "https://github.com/prefix-dev/pixi/issues/1595", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1954222675
Make project.name and project.version optional Problem description Currently pixi can not install a project without these two values: [project] name = "myproject" version = "0.1.0" I suggest making them optional because in a project that does not only use pixi, the name and version may already be defined in another file like package.json. Cargo.toml, making these values in the file duplicate and pointless. Making these optional would be in-line with package.json where there are also no required fields. We're making the version optional :+1: The name is needed for the shell, and requires less house keeping to align with the project. Good call on version. Name won't change that often, but it could potentially fall back to the directory name. Only problem that I can think of is that some CI systems check out projects in generic directory names like /src which would break the fallback.
gharchive/issue
2023-10-20T12:48:56
2025-04-01T04:35:34.084781
{ "authors": [ "ruben-arts", "silverwind" ], "repo": "prefix-dev/pixi", "url": "https://github.com/prefix-dev/pixi/issues/399", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1578442626
Install as LaunchDaemons Closes https://github.com/preludeorg/libraries/issues/333 Install needs to run as sudo sudo -E -- ./install.sh ...
gharchive/pull-request
2023-02-09T19:08:42
2025-04-01T04:35:34.086017
{ "authors": [ "mireaulf" ], "repo": "preludeorg/libraries", "url": "https://github.com/preludeorg/libraries/pull/335", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
478900496
use tagged version instead of latest update the helm chart to use a tagged version instead of latest. Hi @cpanato, we set those fields at tag time using some scripts. This is helpful only when installing the chart from file. Is there a workflow that will help to have those fileds set? @AMecea ah thanks for the explanation. I don't have any workflow just following some best practices for the charts. however, I saw there are some tags and this chart are not updated. Or there are any other repo that is getting updated? thanks again Welcome, The charts are published here. The changes are not saved in the repo at all. Are only changed temporarily before publishing them in the chart repo. So when you run helm install presslabs/mysql-operator will install the latest tag with the images set accordingly.
gharchive/pull-request
2019-08-09T09:50:12
2025-04-01T04:35:34.110224
{ "authors": [ "AMecea", "cpanato" ], "repo": "presslabs/mysql-operator", "url": "https://github.com/presslabs/mysql-operator/pull/383", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
529672109
query the phoenix table , desc table , can't find the TIMESTAMP type columns,how can i do? when use phoenix i can see the timestamp type columns, but can't find in presto: the presto version is 324 Unfortunately, the type isn't supported for now. https://github.com/prestosql/presto/blob/master/presto-phoenix/src/main/java/io/prestosql/plugin/phoenix/TypeUtils.java
gharchive/issue
2019-11-28T03:26:45
2025-04-01T04:35:34.143453
{ "authors": [ "ebyhr", "rongyousu" ], "repo": "prestosql/presto", "url": "https://github.com/prestosql/presto/issues/2126", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
453442098
why discontinued? Hi, as you know firebase is only platform that support flutter but firebase will be too expensive in grow. Will you continue this library or finished? I don't want to call it discontinued. It is just that I am currently not able to provide enough time on this, that's all. But i am open to review and accept any pull request, fix any bugs. or if @stafyniaksacha is willing to mantain this in the long term, i can give him that access @prijindal, yes if you need, I can help maintain this project (I'm an old member of kuzzle' core team) I'm also planing to work on this project again to introduce some new features that are already implemented into the new javascript SDK. Don't hesitate to ask questions about kuzzle , the sdk, or even request for features ! @c2c2 Concerning this SDK we (the Kuzzle team) can not provide the same level of support than for our official project. But as @stafyniaksacha says, he is an old member of our team and we trust him to provide the best quality support for this SDK. Concerning the backend core (http://github.com/kuzzleio/kuzzle) is a long term project. We have a LTS development cycle, every version of Kuzzle is supported for at least 18 month. Kuzzle is a free and open source software but a company is behind it and we have a lot of clients who pay for support so even if I can not say that Kuzzle will last forever, it will stay for a while I think ;) Thanks for your responsibility but i ran an startup with dead time and i think maybe another project i'll use kuzzle for backend @c2c2 You are starting a new project using Flutter ? Yes i do @c2c2 What are missing for your project ? Normally all functionalities are implemented as javascript sdk v6.0.0. Don't hesitate to ask for questions about implementation with flutter here, I'll be happy to help you! You may find support for global kuzzle usage on the official gitter chat here: https://gitter.im/kuzzleio/kuzzle I saw that there are plan to officially do support for this sdk on official roadmap here: https://trello.com/c/b28UCNX3/41-official-dart-flutter-sdk
gharchive/issue
2019-06-07T10:21:01
2025-04-01T04:35:34.219748
{ "authors": [ "Aschen", "c2c2", "prijindal", "stafyniaksacha" ], "repo": "prijindal/kuzzle_dart", "url": "https://github.com/prijindal/kuzzle_dart/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
416382726
React does not recognize the maxWidth prop on a DOM element when using Box When using the Box component with the maxWidth prop, I receive an error saying: React does not recognize the maxWidth prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase maxwidth instead. If you accidentally passed it from a parent component, remove it from the DOM My code: import React from 'react' import { Box } from '@primer/components' export default class Home extends React.Component { public render() { return ( <div> <Box maxWidth={2}>Hello World</Box> </div> ) } } The emotion styles are still applied, but so is the html attribute to the dom. Am I doing something wrong or is this a bug? @Acidic9 Hi! the maxWidth prop needs a unit such as px. Every prop from styled-system (the library we use for our system props) is a CSS prop unless it says otherwise, so using any valid CSS unit will work here. Unsure if anything else in your setup would cause a problem though. The emotion styles are still applied, but so is the html attribute to the dom. FYI we switched to styled-components from emotion several versions ago (I think v8 or before) so you may want to update the version you're on. @Acidic9 I believe this is because some props are not getting filtered out of the remaining HTML attributes that get passed to the DOM by styled-components. I've been noticing this with a few other props and need to look into why this is still happening! It shouldn't affect anything besides having to see the ugly warning in your console. Thanks for the heads up :) This should have been cleared up with the 12.0.1 release! Please let me know if you're still seeing this :)
gharchive/issue
2019-03-02T10:19:43
2025-04-01T04:35:34.291812
{ "authors": [ "Acidic9", "broccolini", "emplums" ], "repo": "primer/components", "url": "https://github.com/primer/components/issues/416", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
926926531
is this should be distance? cosine(embeddings[0], embeddings[1]) is the cosine similarities of 0 and 1,then 1-cosine should be called distance of 0 an 1? ##Calculate cosine similarities ##Cosine similarities are in [-1, 1]. Higher means more similar cosine_sim_0_1 = 1 - cosine(embeddings[0], embeddings[1]) cosine_sim_0_2 = 1 - cosine(embeddings[0], embeddings[2]) print("Cosine similarity between "%s" and "%s" is: %.3f" % (texts[0], texts[1], cosine_sim_0_1)) print("Cosine similarity between "%s" and "%s" is: %.3f" % (texts[0], texts[2], cosine_sim_0_2)) Hi, cosine in scipy is 1 - cosine similarity, so we need to do 1 - cosine to get the cosine similarity. Hi, cosine in scipy is 1 - cosine similarity, so we need to do 1 - cosine to get the cosine similarity. 好的,了解了,谢谢~
gharchive/issue
2021-06-22T07:36:00
2025-04-01T04:35:34.297821
{ "authors": [ "gaotianyu1350", "qiuyu666" ], "repo": "princeton-nlp/SimCSE", "url": "https://github.com/princeton-nlp/SimCSE/issues/46", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1637195037
CoreData: warning: Unable to load class named '' for entity 'Entity'. Class not found, using default NSManagedObject instead. CoreData: warning: Unable to load class named '' for entity 'Entity'. Class not found, using default NSManagedObject instead. How to resolve this warning? Ipad mini 6(Ios 16.3) Mac Air 2020 M1 (MAC OS 13.3) XCode 14.2(14C18) Hello! please ignore these warnings - this is an internal log inside the CoreData framework and does not affect any functionality.
gharchive/issue
2023-03-23T09:54:01
2025-04-01T04:35:34.301916
{ "authors": [ "geor-kasapidi", "jidaojiuyou" ], "repo": "prisma-ai/Sworm", "url": "https://github.com/prisma-ai/Sworm/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
694040622
Associate with graphqls file extension Fixes #176 and here as well https://github.com/prisma-labs/vscode-graphql/blob/6f1e9252f9b5074333b02999d12cfc54f46a0859/src/extension.ts#L76 PR updated, thank you for the indications. I'll be cutting a new pre-release of the language server by monday, so once we cut that, we can bump the dependency in this PR and be good to go! @cailloumajor ok! the stable release of graphql-language-service-server is out, and we just merged a package bump, so this PR is good to go!
gharchive/pull-request
2020-09-05T09:04:03
2025-04-01T04:35:34.305240
{ "authors": [ "acao", "cailloumajor" ], "repo": "prisma-labs/vscode-graphql", "url": "https://github.com/prisma-labs/vscode-graphql/pull/211", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
366901131
Order in express middleware Hello, I currently use apollo-server-express with next.js. I REALLY want to replace apollo-server with graphql-yoga but it does not work. In graphql-yoga, middlewares are triggered before endpoint and playground createHttpServer next's router is a catch all that returns 404 if there is no matching route. run As next's router is triggered before yoga's endpoints, when I try to reach '/graphql' I get 404. With apollo-server-express, I don't have that problem because I have a way to order express middleware applyMiddleware Could you change middleware order so that endpoint, playground are before middlewares ? Thanks Here is my code with graphql-yoga : import next from 'next'; import routes from 'next-routes'; import { GraphQLServer } from 'graphql-yoga'; import schema from './models/schema'; const nextApp = next(); const handler = routes() .add('home', '/', 'Home') .getRequestHandler(nextApp); nextApp.prepare().then(() => { const server = new GraphQLServer({ schema }).use(handler); server.start({ endpoint: '/graphql', playground: '/graphql', getEndpoint: true }); }); @kevinya Wrap the next handler in a middleware that checks the request is not for the /graphql route. nextServer.prepare().then(() => { graphQLServer.use((req, res, next) => { if (req.path.startsWith("/graphql")) return next(); nextServerHandler(req, res, next); }); graphQLServer .start({ endpoint: '/graphql', playground: '/graphql', getEndpoint: true }) .then(() => { console.log("Next.js app is running on http://localhost:3000"); console.log("GraphQL API is running on http://localhost:3000/graphql"); }) .catch(() => { console.error("Server start failed", err); process.exit(1); }); }); I just removed getEndpoint: true and it works. Here is a server.js file `const next = require('next'); const { GraphQLServer } = require('graphql-yoga'); const dev = process.env.NODE_ENV !== 'production'; const port = process.env.PORT || 3000; const gqlEndpoint = '/gql'; const app = next({ dev }); const handle = app.getRequestHandler(); app.prepare().then(() => { const typeDefs = type Query { hello(name: String): String! } ; const resolvers = { Query: { hello: (_, { name }) => `Hello ${name || 'World'}`, }, }; const server = new GraphQLServer({ typeDefs, resolvers }); server.use((req, res, next) => { if (req.path.startsWith(gqlEndpoint)) return next(); handle(req, res, next); }); server.start( { endpoint: gqlEndpoint, playground: gqlEndpoint, subscriptions: gqlEndpoint, port, }).catch(err => console.log('ERROR:', err)) }); ` Just adding this to the pool. Below, working example that combines next and graphql-yoga (along with some integration of Prisma’s newly-released PhotonJS) // NextJS const nextJS = require('next') const nextRoutes = require('next-routes') // GraphQL Yoga const { GraphQLServer } = require('graphql-yoga') // Prisma Photon const Photon = require('@generated/photon') // Custom functions const { resolvers } = require('./server/resolvers') // Photon const photon = new Photon.default() // NextJS const nextJSApp = nextJS({ dev: process.env.NODE_ENV !== 'production' }) const nextJSRoutes = nextRoutes() .add({ pattern: '/', page: '/' }) .add({ pattern: '/signin', page: '/signin' }) .getRequestHandler(nextJSApp) // GraphQL Endpoint const gqlEndpoint = '/server' const gqlServer = new GraphQLServer({ typeDefs: 'server/schema.graphql', resolvers, context: data => ({ ...data, photon }) }) // Begin NextJS... nextJSApp.prepare().then(() => { // ...check if GraphQL endpoint is pinged... gqlServer.use((req, res, next) => { if (req.path.startsWith(gqlEndpoint)) return next() // ... if not, use NextJS routes. nextJSRoutes(req, res, next) }) // Start server. gqlServer .start( { endpoint: gqlEndpoint, playground: gqlEndpoint, subscriptions: gqlEndpoint, port: process.env.PORT || 3000 }, () => console.log(`\n🚀 GraphQL server ready at http://localhost:4000`) ) .then(httpServer => { async function cleanup() { console.log(`\n\nDisconnecting...`) await photon.disconnect() httpServer.close() console.log(`\nDone.\n`) } // process.on('SIGINT', cleanup) process.on('SIGTERM', cleanup) }) .catch(err => { console.error('Server start failed ', err) process.exit(1) }) })
gharchive/issue
2018-10-04T17:50:18
2025-04-01T04:35:34.313021
{ "authors": [ "djgrant", "gsanikidze", "heymartinadams", "kevinya" ], "repo": "prisma/graphql-yoga", "url": "https://github.com/prisma/graphql-yoga/issues/456", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1092571454
Full-text search doesn't work with filter operators/ conditions – OR Bug description Executing the following query throws an error when using the OR operator with full-text search: const query = 'Prisma rocks'.split(' ').join('&') const results = await prisma.book.findMany({ where: { OR: { title: { search: query }, content: { search: query } } }, }) The following error is thrown by Prisma Client: { query: 'Prisma & rocks' } { error: PrismaClientUnknownRequestError: Invalid `prisma.book.findMany()` invocation: Error occurred during query execution: ConnectorError(ConnectorError { user_facing_error: None, kind: QueryError(Error { kind: Db, cause: Some(DbError { severity: "ERROR", parsed_severity: Some(Error), code: SqlState("42601"), message: "syntax error in tsquery: \"david copperfield\"", detail: None, hint: None, position: None, where_: None, schema: None, table: None, column: None, datatype: None, constraint: None, file: Some("tsquery.c"), line: Some(689), routine: Some("makepol") }) }) }) at cb (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/@prisma/client/runtime/index.js:38692:17) at async handler (webpack-internal:///./pages/api/search.ts:40:29) at async Object.apiResolver (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/api-utils.js:102:9) at async DevServer.handleApiRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1064:9) at async Object.fn (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:951:37) at async Router.execute (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/router.js:222:32) at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1135:29) at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/dev/next-dev-server.js:445:20) at async DevServer.handleRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:325:20) { clientVersion: '3.7.0' } If filter operators are not supposed to work when paired with full-text search, we should provide some level of type-safety around it. How to reproduce Query const query = 'Prisma rocks'.split(' ').join(' & ') const results = await prisma.book.findMany({ where: { OR: { title: { search: query }, content: { search: query } } }, }) Expected behavior No error if the query is executed, or provide type safety that filter operators can't be used when using FTS Prisma information Schema: generator client { provider = "prisma-client-js" previewFeatures = ["fullTextSearch"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") } model Book { id Int @id @default(autoincrement()) title String content String url String cover String authors String[] } Migration Environment & setup OS: Mac OS Database: PostgreSQL Node.js version: 16.13 Prisma Version prisma : 3.7.0 @prisma/client : 3.7.0 Current platform : darwin Query Engine (Node-API) : libquery-engine 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node) Migration Engine : migration-engine-cli 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : 8746e055198f517658c08a0c426c7eec87f5a85f Studio : 0.445.0 Preview Features : fullTextSearch This was a mistake on my end. There's no bug here 🙂
gharchive/issue
2022-01-03T14:46:40
2025-04-01T04:35:34.323080
{ "authors": [ "ruheni" ], "repo": "prisma/prisma", "url": "https://github.com/prisma/prisma/issues/10930", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1796376010
PrismaClient is unable to be run in the browser. Bug description Got this error in the prismaClient configuration file. How to reproduce Expected behavior No response Prisma information generator client { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = env("DATABASE_URL") directUrl = env("DIRECT_URL") shadowDatabaseUrl = env("SHADOW_DATABASE_URL") } model Account { id String @id @default(cuid()) userId String type String provider String providerAccountId String refresh_token String? @db.Text access_token String? @db.Text expires_at Int? token_type String? scope String? id_token String? @db.Text session_state String? user User @relation(fields: [userId], references: [id], onDelete: Cascade) @@unique([provider, providerAccountId]) } model Session { id String @id @default(cuid()) userId String expires DateTime sessionToken String @unique accessToken String @unique createdAt DateTime @default(now()) updatedAt DateTime @updatedAt user User @relation(fields: [userId], references: [id]) } model User { id String @id @default(cuid()) name String? email String? @unique emailVerified DateTime? image String? createdAt DateTime @default(now()) updatedAt DateTime @updatedAt accounts Account[] sessions Session[] } model VerificationRequest { id String @id @default(cuid()) identifier String token String @unique expires DateTime createdAt DateTime @default(now()) updatedAt DateTime @updatedAt @@unique([identifier, token]) } import { PrismaClient } from '@prisma/client'; declare global { // eslint-disable-next-line no-var, no-unused-vars var cachedPrisma: PrismaClient; } let prisma: PrismaClient; if (process.env.NODE_ENV === 'production') { prisma = new PrismaClient(); } else { if (!global.cachedPrisma) { global.cachedPrisma = new PrismaClient(); } prisma = global.cachedPrisma; } export const db = prisma; Environment & setup OS: macOS Database: PostgreSQL Node.js version: v19.7.0 Prisma Version prisma : 4.16.2 @prisma/client : 4.16.2 Current platform : darwin-arm64 Query Engine (Node-API) : libquery-engine 4bc8b6e1b66cb932731fb1bdbbc550d1e010de81 (at node_modules/.pnpm/@prisma+engines@4.16.2/node_modules/@prisma/engines/libquery_engine-darwin-arm64.dylib.node) Migration Engine : migration-engine-cli 4bc8b6e1b66cb932731fb1bdbbc550d1e010de81 (at node_modules/.pnpm/@prisma+engines@4.16.2/node_modules/@prisma/engines/migration-engine-darwin-arm64) Format Wasm : @prisma/prisma-fmt-wasm 4.16.1-1.4bc8b6e1b66cb932731fb1bdbbc550d1e010de81 Default Engines Hash : 4bc8b6e1b66cb932731fb1bdbbc550d1e010de81 Studio : 0.484.0 I am implementing next auth, here's my auth file: import { db } from '@/lib/db'; import { PrismaAdapter } from '@next-auth/prisma-adapter'; import { NextAuthOptions } from 'next-auth'; import GoogleProvider from 'next-auth/providers/google'; function getGoogleCredentials(): { clientId: string; clientSecret: string } { const clientId = process.env.GOOGLE_CLIENT_ID; const clientSecret = process.env.GOOGLE_CLIENT_SECRET; if (!clientId || clientId.length === 0) { throw new Error('Missing GOOGLE_CLIENT_ID'); } if (!clientSecret || clientSecret.length === 0) { throw new Error('Missing GOOGLE_CLIENT_SECRET'); } return { clientId, clientSecret }; } export const authOptions: NextAuthOptions = { secret: process.env.NEXTAUTH_SECRET, adapter: PrismaAdapter(db), session: { strategy: 'jwt', }, pages: { signIn: '/', }, providers: [ GoogleProvider({ clientId: getGoogleCredentials().clientId, clientSecret: getGoogleCredentials().clientSecret, }), ], callbacks: { async session({ token, session }) { if (token) { session.user.id = token.id; session.user.name = token.name; session.user.email = token.email; session.user.image = token.picture; } return session; }, async jwt({ token, user }) { const dbUser = await db.user.findFirst({ where: { email: token.email, }, }); if (!dbUser) { token.id = user!.id; return token; } return { id: dbUser.id, name: dbUser.name, email: dbUser.email, picture: dbUser.image, }; }, redirect() { return '/'; }, }, }; Could you please provide us with a reproduction repository? Somehow it seems this is included in your frontend? The Prisma Client would then throw an error. I'm having a similar issue but mine is in the edge I believe. Any suggestions? For context, I'm also using nextauth with next13 with a similar auth setup as @millsp I'm having a similar issue but mine is in the edge I believe. Any suggestions? For context, I'm also using nextauth with next13 with a similar auth setup as @exosky12 Could you please provide us with a reproduction repository? Somehow it seems this is included in your frontend? The Prisma Client would then throw an error. I'm just trying to initialize Prisma, not calling anything for now @millsp :/ This code suggests that at some point something called new PrismaClient either in the browser-side code or somehow it was bundled incorrectly. It would help us if you could share a small repro repository with what is happening. Yes, I think the problem comes from my getServerSession, how can I get the actual session (next auth) without causing troubles ? I will do a reproduction repository as fast as possible. This code suggests that at some point something called new PrismaClient either in the browser-side code or somehow it was bundled incorrectly. It would help us if you could share a small repro repository with what is happening. I know it's not a Repo, but this error is happening to me when I use Prisma from a Server Function (server action) that's being called from a Client Component in NextJS using the App Router. It would be really, really helpful if someone could take the time to put a case like this into a reproduction repository. We'll be able to take that and make it run any time, but setting it up from scratch will take us to properly reserve time for this - which we barely have any right now. Sorry - so please someone help us out. Thanks! I had this same error and I was able to resolve it by updating the auth file. Instead of using Prisma Adapter directly in the authOptions, use it seperately Instead of this export const authOptions: NextAuthOptions = { secret: process.env.NEXTAUTH_SECRET, adapter: PrismaAdapter(db), session: { strategy: 'jwt', }, Do this const adapter = PrismaAdapter(prisma); export const authOptions: NextAuthOptions = { secret: process.env.NEXTAUTH_SECRET, adapter: adapter, session: { strategy: 'jwt', }, Update me if this works out for you, it did for me anyway! I got the same error because I was passing data as props that were called in the layout file I assigned as a client component, "use client". Hi, everything good? I had the same problem few minutes ago, after a some tries, the thing was everything the frontend and how i treated the client side button. I just made an AuthenticationButton recieving provider.id from getProviders method in Server Component: import { AuthenticationButton } from '@/components/AuthenticationButton'; import { getProviders } from 'next-auth/react'; import Image from 'next/image'; import { FcGoogle } from 'react-icons/fc'; export default async function SignUp() { const providers = await getProviders(); return ( <main> <div className='flex flex-col min-h-[calc(100vh/2)] bg-[#5528ff] z-10'></div> <div className='flex flex-col h-1/2'> <div className='p-8 min-w-[400px] mx-auto bg-white -mt-12 z-20 border border-gray-300 rounded-lg'> <div className='flex flex-col items-center pb-8'> <Image src='/promogate-logo.svg' alt='Promogate Logo' width={120} height={60} /> </div> { Object.values(providers!).map((provider, i) => { return ( <AuthenticationButton key={i} className='border border-gray-500 p-2 w-full text-sm text-center' providerId={provider.id}> <FcGoogle /> Cadastrar-se com {provider.name} </AuthenticationButton> ) }) } </div> </div> </main> ) 'use client' import { openSans } from '@/application/utils/fonts' import { signIn } from 'next-auth/react' import React, { ButtonHTMLAttributes } from 'react' interface ButtonProps extends ButtonHTMLAttributes<HTMLButtonElement> { children: React.ReactNode, className?: string, providerId: string } export function AuthenticationButton({ children, className, providerId, ...rest }: ButtonProps) { return ( <button {...rest} className={'p-4 rounded-lg rounded-tl-none flex justify-center items-center gap-x-4' + ' ' + className} style={openSans.style} onClick={() => signIn(providerId)}> {children} </button> ) } No prisma errors doing this. I had a similar issue due to me importing a prisma instance in my nextauth-file. I then exported the authOptions to be used with getServerSession, which I then had imported in a file shared between both the backend and the frontend The same error occurred in my monorepo repository as well. The cause was that I consolidated both frontend and backend functions into index.ts. As you all have pointed out, it was directly because I imported authOptions, but be careful as there seems to be a pattern where it's indirectly imported. index.ts (Bug occurs) export { default } from "next-auth"; export { signIn, signOut, useSession } from "next-auth/react"; export { authOptions } from "./src/auth-options"; export { getServerSession } from "./src/get-session"; 👇Fixed client.ts export { default } from "next-auth"; export { SessionProvider, signIn, signOut, useSession } from "next-auth/react"; server.ts export { authOptions } from "./src/auth-options"; export { getServerSession } from "./src/get-session"; package.json "files": [ "client.ts", "server.ts" ], Can someone please help us get a working reproduction of the problem? The code snippets above might be helpful if you know Next.js and next-auth, but it would take us quite some time to put that together. It would be amazing if one of you could create a new project and implement the minimal code to show the problem - and then put it on GitHub. Thank you! While we still don't have a reproduction of this issue, I don't think this is a Prisma issue per se. The problem to me is that Prisma Client should be used in a backend JS runtime (e.g., Node.js) only, but when developing isomorphic Next.js projects (in which the separation between frontend and backend isn't necessarily super clear at first glance), the Prisma Client instance can easily leak into the frontend. And, since these leaks cause Prisma Client to run in a browser (which is not supported), the runtime error above is thrown. In addition to what @jkomyno said, we recently fixed the fact that that we threw errors when doing new PrismaClient and now we moved these errors to only happen if you try emitting a query in your front-end (which obviously cannot work).
gharchive/issue
2023-07-10T09:39:32
2025-04-01T04:35:34.340500
{ "authors": [ "AbelR007", "Oliveira-86", "TakashiAihara", "alvesvaren", "evanrosa", "exosky12", "gutodiasdev", "janpio", "jkomyno", "millsp", "nukkerone" ], "repo": "prisma/prisma", "url": "https://github.com/prisma/prisma/issues/20147", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
811067103
POC using Prisma Migrate for CD Acceptance criteria: Generic steps to apply for CD A working POC of applying migrations to a cloud deployment via prisma migrate deploy with Heroku Release Rough documentation with the steps involved and any potential caveats This is done: Generic deployment guide Vercel deployment guide Heroku deployment guide
gharchive/issue
2021-02-18T12:45:11
2025-04-01T04:35:34.344789
{ "authors": [ "albertoperdomo" ], "repo": "prisma/prisma", "url": "https://github.com/prisma/prisma/issues/5725", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1466112693
ci(no-docker client tests): lower --max-old-space-size for linux and windows https://prisma-company.slack.com/archives/CUXLS0Z6K/p1669489504119379 the conditional works!
gharchive/pull-request
2022-11-28T09:45:52
2025-04-01T04:35:34.346352
{ "authors": [ "Jolg42" ], "repo": "prisma/prisma", "url": "https://github.com/prisma/prisma/pull/16485", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
539253246
Prisma migration engine on s3 has been deleted, cannot install on CI Problem When installing Prisma2, certain files from an s3 bucket (prisma-native) are downloaded. Just starting this morning, these files can now not be found. How To Reproduce Run npm i -g --unsafe-perm prisma2@2.0.0-preview014.2 in CircleCI Thanks for reporting, we will look into this. Until then, installing a recent version of prisma2 will probably solve that problem. See https://github.com/prisma/prisma2/releases for the most recent release. @janpio Thanks for the reply. We have plans to upgrade but we found it will take significant effort due to our migrations not being compatible.
gharchive/issue
2019-12-17T19:02:43
2025-04-01T04:35:34.348811
{ "authors": [ "aleccool213", "janpio" ], "repo": "prisma/prisma2", "url": "https://github.com/prisma/prisma2/issues/1172", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2144982408
Native node modules for downloading artifacts Related Issue(s) Closes #166 Checklist [x] My code follows the style guidelines of this project [x] I have performed a self-review of my code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [x] My changes generate no new warnings [x] I have run yarn prettier and yarn lint without getting any errors [x] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes [x] Any dependent changes have been merged and published in downstream modules LGTM. Just one thing, there is a duplication between eddsa-proof and poseidon-proof packages (the functions are the same). Wouldn't it make sense to have a shared utility? Yes, I think some code could be moved to a shared function (maybe in the utils package). Could you open another issue for this? LGTM. Just one thing, there is a duplication between eddsa-proof and poseidon-proof packages (the functions are the same). Wouldn't it make sense to have a shared utility? Yes, I think some code could be moved to a shared function (maybe in the utils package). Could you open another issue for this? Sure, tracked in https://github.com/privacy-scaling-explorations/zk-kit/issues/176
gharchive/pull-request
2024-02-20T18:10:13
2025-04-01T04:35:34.370009
{ "authors": [ "0xjei", "cedoor" ], "repo": "privacy-scaling-explorations/zk-kit", "url": "https://github.com/privacy-scaling-explorations/zk-kit/pull/173", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
932473105
🆕 Software Suggestion | Responding to reply on post being removed Basic Information Name: lazyweb.ai (redditor lazy-jem) Category: Search engine URL: lazyweb.ai Description Hi, I'm sorry. I posted earlier tonight seeking feedback about our alpha version of a new search engine app. It isn't a commercial app yet, so it wasn't intended as promotion, but the post was removed by /u/trai_dep with the following note: If you have a project that you want to promote here, open an issue on our GitHub repo so our entire team can advise and evaluate it first. I'm sorry I didn't appreciate that I should have cleared a post seeking feedback like this with the mods first. We aren't ready to promote it yet, as it isn't a commercial release. I thought the community would find it interesting and have useful feedback on how we could make what we're making more privacy-focused. We are mission driven and it wasn't our intention to be self promotional here as we are not at that stage. With 129 upvotes and 28 comments in the first few hours, it did seem interesting to the community, and I'm sorry we misunderstood about doing this the right way, and would very much like to work with you to correct the mistake. A search engine is obviously not typical privacy software, as it isn't run directly by the user. I built this out of frustration with ad-tech and privacy invasion. So it is a personal project that I though other folks might find interesting. Why I am making the suggestion Here are some comments from the post: Made a search web app that's anonymous, ad-free and non-tracking. It lets you read web content in a clean reader view anonymously through a proxy that strips ad-tech and tracking. Looking for feedback on our approach to privacy in the new alpha test version we just released. It's called LazyWeb. It uses a chat interface and gives you control of how you view search results. The chat interface means that searches stay within the anonymous chat session rather than going through the browser history. The alpha version is open to anyone to try here - https://lazyweb.ai We're a small bootstrapped two-person team. I'm the technical co-founder. We'd be grateful for thoughts, suggestions, and feedback on how we're approaching privacy. LazyWeb doesn't log searches, and it blocks tracking and ad-tech. We only collect and retain sufficient data to improve the service we provide and help our customers use the service effectively, or in future if customers want to create an account or be remembered between devices and sessions. We use limited in-app analytics solely to help improve the application for people using it. The metrics are anonymous, reported in aggregate, and it do not contain any personal data or searches. They are not shared with anyone. You can disable all in-app analytics on the settings page. One of the big challenge building a private, anonymous search engine is that nothing is logged, so we can't see what people search, or when things go wrong searching. So the only way to keep the results improving and fix problems is to get lots of feedback from people who are subject matter experts. Technical searches need a lot of work and we'd love any feedback on the search results in specialized areas (like privacy and security). My connection with the software Author [YES ] I will keep the issue up-to-date if something I have said changes or I remember a connection with the software. @jedwhite you mention that metrics are anonymous, so could you specify what exactly falls under 'metrics' for you and how do you ensure the anonymity hereof? What is the GEO IP service for and are IP's shared with third parties for this? Doesn't sound very privacy friendly. I also cannot find any real privacy policy, which you are required to have by the GDPR. I think this should be closed and maybe you could open a new issue when you have figured things out. Thank you for taking the time to respond and for your feedback. I think there may be some confusion about the request. I posted here at the suggestion of /u/trai_dep from the subreddit. We are not asking to be listed as suggested software or be recommended, and we are a long way from that point. The purpose of the subreddit post was explicitly to ask for help from the community to seek feedback on our approach at a very early stage, and to figure things out to do better. As noted, the web app is an alpha test of a prototype and proof-of-concept only, and not a commercial service. We are not promoting it for commercial or production use, and searches are currently limited to en-US only. I'm grateful for you taking to time to comment here, and this is exactly the type of feedback that we were seeking by posting on the subreddit. We definitely aren't asking for a recommendation or anything of the kind at this point, but we would like to work with you to be able to discuss with other people interested in privacy how we can build a better alternative to Google, DDG, Bing, Brave etc. Thanks for the chance to answer some of your questions too. While it is an early prototype, we have a plain language privacy policy here (https://lazyweb.ai/privacy/). It is one of the things we are seeking feedback on at an early stage. Our informal legal advice is that it does meet the GDPR requirements broadly given we do not store any personal information but that we need to add more detail (especially how we collect the in-app analytics and storage). With the GEOIP: We don't log or store IP addresses. It's used to lookup your approximate location (nearest city) for location searches only, then discarded. It is never passed to third-parties. We only use a GPS or detailed location for searches with a user's express permission, and then only to approximate the area (nearest city). Your GPS location details are not stored or passed to any third-parties. We're using the maxmind database. We don't log or store IP addresses or any other geolocation data. For the analytics, we collect the city location name, and then summarise the number of users for each city by day. We don't log or store searches including searches that incorporate location elements. IP and actual geo coordinates are not passed with searches, and retrieved only by the browser client side. So the geo lookup API is isolated from searches and not sent to the search back end. That's why we lookup geo data on the client as a separate process, and pass location city name with search requests, rather than doing it server side from the search engine. Originally, we completely disabled location until enabled. But people trying the app told us that was a terrible experience, because there is an expectation when someone asks the time, or weather, or the best coffee shop that the results will be at least nearby to them without them having to explictly enable location (we recognize that is a paradox). So we're trying to determine what the best trade-off is. Currently, we determine the broad city location from a separate independent client-side lookup, and a user can enable GPS (but it is still approximated to the closest city - just more accurately as IP/maxmind can be wildly inaccurate), and only the location city name is used for search localization. As far as we are aware, other major commercial search engines use IP geocoordinates with full fidelity server-side by default for location, so we think this is a better approach. But it is definitely one of the things we're seeking ideas and feedback on. With the in-app analytics, there is no personal information passed (no IP, no geocoords, no search request content). We record signals on the broad search intent (for example, a programming search intent or food place search intent) grouped into about 60 intent categories. No search terms are passed. And we record the types of actions performed (again only by broad category - for example "external link clicked" but not what any links clicked were. We don't record referrer or any other information. We don't record what items people read or navigate to. This data is reduced to summary figures by hour and day, and the original events discarded. There is a session client identifier. We don't pass user agent strings, usernames, or any PII. We pass whether it is a first time session or returning session, mobile or desktop, and top-level browser name and OS type (again by category, not user agent string). It's worth noting that searches are not even passed to the web browser history, as they remain within the chat session, and that is destroyed with each page load. So they are even masked from Google's typical Chrome-level tracking. We have tried to find the best way to approach understanding how to improve the app while keeping use anonymous, and it is one of the areas we are hoping to get feedback on and improve. Again, we aren't claiming to have solved this or that we have arrived at the right approach, but we are trying to talk with potential future users who care about privacy to work out how to build a good approach. Thanks again for your feedback and questions, and I hope that's helpful. @jedwhite thank you for clarification. Obviously you will need to get legal advice (which I cannot give you), but the thing you call a privacy policy does not seem to meet the requirements as defined in the GDPR. The things you mention about metrics would need to be audited, more then often I have found systems like these not to be implemented accurately and find options remaining to recombine data and being rather pseudonym instead of anonymous. Sessions are connect to a user and therefor likely a pseudonym. I don't think software suggestion as in your title is the right category. If you are looking for feedback this is likely not the right place to ask for help. Just to be clear, I like that people start up projects like this, but I see my role here to be hesitant and conservative against any new listings that yet not make the difference ;) I'm going to close this because of the privacy policy issue - but I'll reopen it if/when that is fixed.
gharchive/issue
2021-06-29T10:32:16
2025-04-01T04:35:34.405428
{ "authors": [ "freddy-m", "jedwhite", "ph00lt0" ], "repo": "privacytools/privacytools.io", "url": "https://github.com/privacytools/privacytools.io/issues/2360", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
916616092
typescript definition declare module 'html-to-docx' { interface Margins { /** * top <Number> distance between the top of the text margins for the main document and the top of the page for all pages in this section in TWIP. * Defaults to 1440. Supports equivalent measurement in pixel, cm or inch. */ top: number; /** * right <Number> distance between the right edge of the page and the right edge of the text extents for this document in TWIP. * Defaults to 1800. Supports equivalent measurement in pixel, cm or inch. */ right: number; /** * bottom <Number> distance between the bottom of text margins for the document and the bottom of the page in TWIP. * Defaults to 1440. Supports equivalent measurement in pixel, cm or inch. */ bottom: number; /** * left <Number> distance between the left edge of the page and the left edge of the text extents for this document in TWIP. * Defaults to 1800. Supports equivalent measurement in pixel, cm or inch. */ left: number; /** * header <Number> distance from the top edge of the page to the top edge of the header in TWIP. * Defaults to 720. Supports equivalent measurement in pixel, cm or inch. */ header: number; /** * footer <Number> distance from the bottom edge of the page to the bottom edge of the footer in TWIP. * Defaults to 720. Supports equivalent measurement in pixel, cm or inch. */ footer: number; /** * gutter <Number> amount of extra space added to the specified margin, above any existing margin values. This setting is typically used when a document is being created for binding in TWIP. * Defaults to 0. Supports equivalent measurement in pixel, cm or inch. */ gutter: number; } interface Row { /** * cantSplit <?Boolean> flag to allow table row to split across pages. Defaults to false. */ cantSplit?: boolean; } interface Table { row?: Row; } export interface DocumentOptions { /** * orientation <"portrait"|"landscape"> defines the general orientation of the document. Defaults to portrait. */ orientation?: "portrait" | "landscape"; margins?: Margins /** * title <?String> title of the document. */ title?: string; /** * subject <?String> subject of the document. */ subject?: string; /** * creator <?String> creator of the document. Defaults to html-to-docx */ creator?: string; /** * keywords <?Array<String>> keywords associated with the document. Defaults to ['html-to-docx']. */ keywords?: string[]; /** * description <?String> description of the document. */ description?: string; /** * lastModifiedBy <?String> last modifier of the document. Defaults to html-to-docx. */ lastModifiedBy?: string; /** * revision <?Number> revision of the document. Defaults to 1. */ revision?: number; /** * createdAt <?Date> time of creation of the document. Defaults to current time. */ createdAt?: Date; /** * modifiedAt <?Date> time of last modification of the document. Defaults to current time. */ modifiedAt?: Date; /** * headerType <"default"|"first"|"even"> type of header. Defaults to default. */ headerType?: "default" | "first" | "even"; /** * header <?Boolean> flag to enable header. Defaults to false. */ header?: boolean; /** * footerType <"default"|"first"|"even"> type of footer. Defaults to default. */ footerType?: "default" | "first" | "even"; /** * footer <?Boolean> flag to enable footer. Defaults to false. */ footer?: boolean; /** * font <?String> font name to be used. Defaults to Times New Roman. */ font?: string; /** * fontSize <?Number> size of font in HIP(Half of point). Defaults to 22. Supports equivalent measure in pt. */ fontSize?: number; /** * complexScriptFontSize <?Number> size of complex script font in HIP(Half of point). Defaults to 22. Supports equivalent measure in pt. */ complexScriptFontSize?: number; table?: Table; /** * pageNumber <?Boolean> flag to enable page number in footer. Defaults to false. Page number works only if footer flag is set as true. */ pageNumber?: boolean; /** * skipFirstHeaderFooter <?Boolean> flag to skip first page header and footer. Defaults to false. */ skipFirstHeaderFooter?: boolean; } /** * @param htmlString <String> clean html string equivalent of document content. * @param headerHTMLString <String> clean html string equivalent of header. Defaults to <p></p> if header flag is true. * @param documentOptions <DocumentOptions> * @param footerHTMLString <String> clean html string equivalent of footer. Defaults to <p></p> if footer flag is true. * @constructor * @private */ export default function HTMLtoDOCX(htmlString: string, headerHTMLString: string, documentOptions: DocumentOptions, footerHTMLString?: string): Promise<Buffer | Blob>; } @RoxnnyABarriosC Please raise a PR with this typescript definition, and also do we need any sort of typescript configuration to make use of this definition? Any updates on this ? @privateOmega RE: typescript. I've created a branch converted to typescript. It's working locally, if useful I can do up a PR. https://github.com/jafin/html-to-docx/tree/feat/typescript @jafin when will this PR be up ? it looks promising. any update on this?
gharchive/issue
2021-06-09T19:42:36
2025-04-01T04:35:34.421162
{ "authors": [ "RoxnnyABarriosC", "jafin", "lvndry", "privateOmega", "rossanodr", "whis20" ], "repo": "privateOmega/html-to-docx", "url": "https://github.com/privateOmega/html-to-docx/issues/79", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1806392404
feat: Add Theme Component for Toggling light/dark mode What feature? Want to add a Theme Component for Toggling light/dark mode in Next.js 13 ...Could you assign this issue to me as a GSSOC23 contributor? I want to make it this way when users click the moon icon UI will appear in dark mode & when it is the sun icon the UI will appear in a light mode Add screenshots Record [X] I agree to follow this project's Code of Conduct [X] I'm a GSSoC'23 contributor [X] I want to work on this issue @crocmons We decided to use only the dark mode, so we removed the option to switch between dark and light modes that we had before. ok
gharchive/issue
2023-07-16T01:30:46
2025-04-01T04:35:34.428869
{ "authors": [ "crocmons", "priyankarpal" ], "repo": "priyankarpal/ProjectsHut", "url": "https://github.com/priyankarpal/ProjectsHut/issues/1719", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2139916238
hoist change copyright date code from NeXus definitions close #5 Needs some unit tests and documentation. Pull Request Test Coverage Report for Build 7940777040 Details 0 of 0 changed or added relevant lines in 0 files are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 63.636% Totals Change from base Build 7937257643: 0.0% Covered Lines: 7 Relevant Lines: 11 💛 - Coveralls
gharchive/pull-request
2024-02-17T09:29:18
2025-04-01T04:35:34.433823
{ "authors": [ "coveralls", "prjemian" ], "repo": "prjemian/murky", "url": "https://github.com/prjemian/murky/pull/35", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
587316759
Double attach is possibly bad? The problem In most cases we perform a double attach. When the user creates a probe, we call attach on the probe: https://github.com/probe-rs/probe-rs/blob/master/probe-rs/src/probe/mod.rs#L123 And when the user attaches to a target, we call attach on the probe again: https://github.com/probe-rs/probe-rs/blob/master/probe-rs/src/probe/mod.rs#L210 This effect can be seen here: https://github.com/probe-rs/cargo-flash/blob/master/src/main.rs#L234-L268. The effect is possibly bad because if a probe for example does not support SWD and the default is SWD and it tries to use that, hell breaks loose. Solutions If we remove the first one, it is not guaranteed that when the user operates on the probe, for example via DapAccess: https://github.com/probe-rs/probe-rs/blob/master/probe-rs/src/probe/mod.rs#L250, that the probe is actually attached to the chip. If we remove the second one, all the settings like the protocol etc do not get applied. I see 2 solutions: a) use a typestate to track whether it's attached or not or b) use an internal bool and attach if the probe is not attached. ping: @Tiwalun @Disasm Fixed by #200 .
gharchive/issue
2020-03-24T22:25:19
2025-04-01T04:35:34.514163
{ "authors": [ "Yatekii" ], "repo": "probe-rs/probe-rs", "url": "https://github.com/probe-rs/probe-rs/issues/199", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1190496784
Book 2 (version 2022-04-01), Section 8.8, Page 373: Code listing uses wrong string quote The string quote before A needs to be corrected. It should be jnp.einsum('A ...') An alternative could be use LaTeX listings, that way the code would also be directly copy-pasteable (will not give quote problems). Example output LaTeX code used to generate the above listing. \documentclass{article} \usepackage{listings} \usepackage{xcolor} %New colors defined below \definecolor{codegreen}{rgb}{0,0.6,0} \definecolor{codegray}{rgb}{0.5,0.5,0.5} \definecolor{codepurple}{rgb}{0.58,0,0.82} \definecolor{backcolour}{rgb}{0.95,0.95,0.92} %Code listing style named "mystyle" \lstdefinestyle{mystyle}{ backgroundcolor=\color{backcolour}, commentstyle=\color{codegreen}, keywordstyle=\color{magenta}, numberstyle=\tiny\color{codegray}, stringstyle=\color{codepurple}, basicstyle=\ttfamily\footnotesize, breakatwhitespace=false, breaklines=true, captionpos=b, keepspaces=true, numbers=left, numbersep=5pt, showspaces=false, showstringspaces=false, showtabs=false, tabsize=2 } %"mystyle" code listing set \lstset{style=mystyle} \begin{document} %Python code highlighting \begin{lstlisting}[language=Python, caption=Python example] import jax.numpy as jnp from jax import grad logZ_fun = lambda logpots: np.log(jnp.einsum("A,B,C,AB,BC", *[jnp.exp(lp) for lp in logpots])) probs = grad(logZ_fun)(logpots) \end{lstlisting} \end{document} fixed, thanks
gharchive/issue
2022-04-02T04:45:29
2025-04-01T04:35:34.516652
{ "authors": [ "murphyk", "nipunbatra" ], "repo": "probml/pml-book", "url": "https://github.com/probml/pml-book/issues/299", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
559273871
docs: update ZEIT Now deployment instructions Fixes #985 View rendered docs/deployment.md nice docs! small comment (it's outside of your range of changes so couldn't comment on the code): you need to update the hyperlink to the Now docs, since you changed the section name (L17) - 1. [Now](#now) + 1. [Now](#zeit-now) @benjlevesque Nice catch! I updated that line 👍 Thanks you two :)
gharchive/pull-request
2020-02-03T19:06:00
2025-04-01T04:35:34.532403
{ "authors": [ "benjlevesque", "gr2m", "styfle" ], "repo": "probot/probot", "url": "https://github.com/probot/probot/pull/1130", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
544084625
Plugin not replacing img with Gatsby Img tag I've installed everything locally and I'm able to display content from my local wordpress site. The issue I'm having is that the images are all urls from the wordpress site, when I thought that they would be downloaded as static images to the gatsby site and sourced from the static folder. Currently none of that is working. I'm wondering if you can advise. my github of the demo install is here: https://github.com/spencersmb/wordpress-graphql-gatsby-demo I'm able to display my posts and show images, but it doesn't seem to make a difference when using dangerouslySetInnerHtml vs. ContentParser in my code. In my demo - I added the code to layout.js Thanks for any help or advice. Same issue here, maybe these warnings at build time can be useful: warn "createResolvers" passed resolvers for type "wpgraphql_page" that doesn't exist in the schema. Use "createTypes" to add the type before adding resolvers. Nevermind, with graphqlTypeName: "WPGraphQL" in this plugin configuration matching typeName: "WPGraphQL" in gatsby-source-graphql configuration, it works fine.
gharchive/issue
2019-12-31T04:27:48
2025-04-01T04:35:34.632746
{ "authors": [ "jenkin", "spencersmb" ], "repo": "progital/gatsby-wpgraphql-blog-example", "url": "https://github.com/progital/gatsby-wpgraphql-blog-example/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1996949892
Pods with unready Containers exist on this node, we can't clean the slots yet Describe the bug Akri agent daemonset keeps reporting following error whenever any of pod running on a cluster is not ready. 2023-11-16T13:44:46Z TRACE agent::util::slot_reconciliation] reconcile - Pods with unready Containers exist on this node, we can't clean the slots yet In my case failing POD doesn't use USB resources. Output of kubectl get pods,akrii,akric -o wide lpfe04@f1725b929a:~$ kubectl get pod,akrii,akric -n akri NAME READY STATUS RESTARTS AGE pod/akri-agent-daemonset-9gwl2 1/1 Running 0 10m pod/akri-controller-deployment-7c6455f79-zt779 1/1 Running 0 11m pod/akri-udev-discovery-daemonset-2d9hp 1/1 Running 0 10m pod/akri-webhook-configuration-7bf6656b45-mclth 1/1 Running 0 11m NAME CONFIG SHARED NODES AGE instance.akri.sh/gsm-dongle-6e977d gsm-dongle false ["f1725b929a"] 10m instance.akri.sh/wifi-dongle-254c38 wifi-dongle false ["f1725b929a"] 10m instance.akri.sh/wifi-dongle-ac917e wifi-dongle false ["f1725b929a"] 10m NAME CAPACITY AGE configuration.akri.sh/gsm-dongle 1 11h configuration.akri.sh/wifi-dongle 1 11h Kubernetes Version: [e.g. Native Kubernetes 1.19, MicroK8s 1.19, Minikube 1.19, K3s] kubernetes: v1.26.8+rke2r1" Expected behavior I would expect reconciliation process can be continue if failing pod is out of usb usage.management context. @rpieczon just to clarify, are you saying if any pod (even if unassociated with Akri) is unready, it causes this slot reconciliation error? From what i remember slot reconciliation should only check pods with an expected annotation. Exactly in my case I have failing Prometheus POD which has zero requirements related with USB allocation. Any update on it?
gharchive/issue
2023-11-16T14:12:00
2025-04-01T04:35:34.649637
{ "authors": [ "rpieczon" ], "repo": "project-akri/akri", "url": "https://github.com/project-akri/akri/issues/681", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2567649776
Pap 221 frontend create UI components Ticket ID: PAP-221 Link: https://project-ascend-io.atlassian.net/browse/PAP-221?atlOrigin=eyJpIjoiNTNjY2Q4YTY4YTIzNDI3ZTgxNWJiYzE5YjQzMGU1ZTIiLCJwIjoiaiJ9 Design File: https://www.figma.com/design/siiCloyWGa7e19lYRr8RP4/Messanging?node-id=0-1&node-type=canvas&t=3M0OEULDR89OcSz6-0 Problem The application needs a real-time direct messaging feature so users can converse with one another. The User Interface needed to be built to lay the foundation for the DM system. Solution Phase 1: I modified the existing codebase to enhance navigation, readability, and UI according to the design file. Full functional code and data persistence will be implemented in Phase 2. Type of change [X] Breaking change (fix or feature that would cause existing functionality not to work as expected) How Has This Been Tested? Visually tested to match the design file. Proper testing will be conducted during phase 4. Checklist: [X] I have performed a self-review of my code [X] I have commented my code, particularly in hard-to-understand areas [X] My changes generate no new warnings [X] I have added tests that prove my fix is effective or that my feature works [X] New and existing unit tests pass locally with my changes Great job, Robert! Everything looks good on I end. I approved. https://github.com/user-attachments/assets/7d7083aa-16ff-448f-bcdc-5034403eb5f4
gharchive/pull-request
2024-10-05T04:23:20
2025-04-01T04:35:34.654563
{ "authors": [ "bobbygrdn", "oestrada1001" ], "repo": "project-ascend-io/intracom-electron", "url": "https://github.com/project-ascend-io/intracom-electron/pull/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2515266769
service "backend" is not running Describe the bug This problem occurs when I execute the start script. How do I fix it?I tried upgrading and clearing, and then restarting the script, but it didn't work. Steps to reproduce the behavior 1、Clone the TH repository 2、Goto to TH folder 3、Install/configure the TH dependencies ./scripts/pi-setup/auto-install.sh Expected behavior No response Log files No response PICS file No response Screenshots No response Environment No response Additional Information No response The Raspberry PI system version is Ubuntu 24.04.1 LTS The branch is commit f52d40a16def63edecdabbc8edcb14eb76b12ba6 It appears that the installation steps are incomplete: Please try these: cd ~/certification-tool ./scripts/stop.sh cd .. rm -rf certification-tool git clone -b v2.11-beta3.1+fall2024 https://github.com/project-chip/certification-tool.git --recurse-submodules cd ~/certification-tool ./scripts/pi-setup/auto-install.sh Problem solved. Thank you for your help @fabiowmm @hiltonlima
gharchive/issue
2024-09-10T02:32:26
2025-04-01T04:35:34.659639
{ "authors": [ "hiltonlima", "song7788" ], "repo": "project-chip/certification-tool", "url": "https://github.com/project-chip/certification-tool/issues/412", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1960632125
subprocess.CalledProcessError: Command '['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']' returned non-zero exit status 1.[BUG] Reproduction steps I use this link to setup Matter Device https://community.arm.com/arm-community-blogs/b/internet-of-things-blog/posts/build-a-matter-home-automation-service-using-raspberry-pi-arm-virtual-hardware-and-python but i create the step at 8 & 9 always show error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: legacy-install-failure × Encountered error while trying to install package. ╰─> grpcio-tools note: This is an issue with the package mentioned above, not pip. hint: See above for output from the failure. ninja: build stopped: subcommand failed. ['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install'] Traceback (most recent call last): File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/virtualenv_setup/install.py", line 327, in install_packages subprocess.check_call(ninja_cmd, stdout=outs, stderr=outs) File "/usr/lib/python3.11/subprocess.py", line 413, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']' returned non-zero exit status 1. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 795, in sys.exit(main()) ^^^^^^ File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 787, in main return EnvSetup(**vars(parse())).setup() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 457, in setup result = step(spin) ^^^^^^^^^^ File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 606, in virtualenv if not virtualenv_setup.install( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/virtualenv_setup/install.py", line 342, in install install_packages(gn_target) File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/virtualenv_setup/install.py", line 330, in install_packages raise subprocess.CalledProcessError(err.returncode, err.cmd, subprocess.CalledProcessError: Command '['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']' returned non-zero exit status 1. So How can i fix this problem Bug prevalence everytimes GitHub hash of the SDK that was being used db49235c39635582ea522929f9905af03e3114c7 Platform python, raspi Platform Version(s) python 3.11.2 raspi Raspberry Pi OS Lite(64bit) 2023-10-10 Anything else? No response The error seems to be not being able to install grpcio-tools. Are you able to pip install grpcio-tools generally or in a virtualenv? Platform also seems to state python 3.11.2 raspi Raspberry Pi OS Lite(64bit). For our RPI compiles, we always use an ubuntu version. I would try with Ubuntu 24.04 to make sure you have the highest changes of success. Using RPiOS you are likely using something that not many (guessing nobody else?) are using.
gharchive/issue
2023-10-25T06:22:06
2025-04-01T04:35:34.671067
{ "authors": [ "andy31415", "patrickwang0821" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/29985", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1983978435
[Build] Build issue(s) I have stated my initial question here: https://github.com/project-chip/connectedhomeip/issues/30237 but it was closed before I could test the proposed solution so I opened it as a new issue. I have changed the lighting-app.zap as suggested to only have one slider in the HomeApp IOs for dimming rather than 6 for different colors by default in lighting-app): (enable color control: 0) that was my initial change which didn't help I then dowloaded ZAP GUI and opened lighting-app.zap using it and regenerated .zap then I run script to regenerate new .matter file. I put it in connectedhomeip/examples/lighting-app/light-common Device says Dimmible Light in ZAP file: After above changes I have deleted build files in the lighting-app and used idf.py build I get the following errors when building the file: I thought that should not be considered anymore as I disabled colourcontrol? Could you please advise what is wrong ? Platform esp32 Anything else? No response examples/lighting-app/esp32/main/CMakeLists.txt has: "${CMAKE_SOURCE_DIR}/third_party/connectedhomeip/src/app/clusters/color-control-server" which will unconditionally try to compile the color control server. This needs to be changed if you change the ZAP config, no?
gharchive/issue
2023-11-08T16:27:07
2025-04-01T04:35:34.676154
{ "authors": [ "BartJozwicki", "bzbarsky-apple" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/30324", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2047220145
[Android] The QR code payload parser does not support optional fields Reproduction steps On Android call parseQrCode("MT:SA6K48.E15R5WY5.D3P0SOCE0LDQJ1DK5N1K8SQ1RYCU1O0") in SetupPayloadParser class Examine returned payload Expected: An optional field for serialNumber should be present Actual: No optional fields are present Platform android Platform Version(s) No response Type Platform validated (Optional) If manually tested please explain why this is only manually tested No response Anything else? We were hoping to embed a serialNumber in the QR code but Android's payload parser does not support this yet. It is available on iOS as well as the chip tool ./chip-tool payload parse-setup-payload MT:SA6K48.E15R5WY5.D3P0SOCE0LDQJ1DK5N1K8SQ1RYCU1O0 [1702672725954] [5602:2751420] [SPL] Version: 0 [1702672725954] [5602:2751420] [SPL] VendorID: 1234 [1702672725954] [5602:2751420] [SPL] ProductID: 1234 [1702672725954] [5602:2751420] [SPL] Custom flow: 0 (STANDARD) [1702672725954] [5602:2751420] [SPL] Discovery Bitmask: 0x02 (BLE) [1702672725954] [5602:2751420] [SPL] Long discriminator: 717 (0x2cd) [1702672725954] [5602:2751420] [SPL] Passcode: 13360673 [1702672725954] [5602:2751420] [SPL] SerialNumber: CN0123456789 I noticed there is a TODO to add this to Android: https://github.com/project-chip/connectedhomeip/blob/35455fa577cf114b68016660118e3d408fca253c/src/controller/java/src/matter/onboardingpayload/QRCodeOnboardingPayloadParser.kt#L63 @stevePalmerin does this removed Java setuppayload support this optional field? https://github.com/project-chip/connectedhomeip/pull/27630/files?
gharchive/issue
2023-12-18T18:35:39
2025-04-01T04:35:34.681322
{ "authors": [ "stevePalmerin", "yunhanw-google" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/31079", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
837650639
Compile error when CHIP_BLE_END_POINT_DEBUG_LOGGING_ENABLED is set Problem When compiling with the CHIP_BLE_END_POINT_DEBUG_LOGGING_ENABLED enabled, I get compile errors in src/ble/BLEEndPoint.cpp Proposed Solution I propose to use the following patch: diff --git a/src/ble/BLEEndPoint.cpp b/src/ble/BLEEndPoint.cpp index aaf0d52c..ec95fc2c 100644 --- a/src/ble/BLEEndPoint.cpp +++ b/src/ble/BLEEndPoint.cpp @@ -632,12 +632,12 @@ void BLEEndPoint::QueueTx(PacketBufferHandle && data, PacketType_t type) if (mSendQueue.IsNull()) { mSendQueue = std::move(data); - ChipLogDebugBleEndPoint(Ble, "%s: Set data as new mSendQueue %p, type %d", __FUNCTION__, mSendQueue, type); + ChipLogDebugBleEndPoint(Ble, "%s: Set data as new mSendQueue %p, type %d", __FUNCTION__, &mSendQueue, type); } else { mSendQueue->AddToEnd(std::move(data)); - ChipLogDebugBleEndPoint(Ble, "%s: Append data to mSendQueue %p, type %d", __FUNCTION__, mSendQueue, type); + ChipLogDebugBleEndPoint(Ble, "%s: Append data to mSendQueue %p, type %d", __FUNCTION__, &mSendQueue, type); } QueueTxUnlock(); @@ -993,7 +993,7 @@ BLE_ERROR BLEEndPoint::DriveSending() { #ifdef CHIP_BLE_END_POINT_DEBUG_LOGGING_ENABLED if (mRemoteReceiveWindowSize <= BTP_WINDOW_NO_ACK_SEND_THRESHOLD && - !mTimerStateFlags.Has(TimerStateFlag::kSendAckTimerRunning) && mAckToSend == NULL) + !mTimerStateFlags.Has(TimerStateFlag::kSendAckTimerRunning) && mAckToSend.IsNull()) { ChipLogDebugBleEndPoint(Ble, "NO SEND: receive window almost closed, and no ack to send"); } Given this is an old issue, is this still present? If so, please re-open, thanks!
gharchive/issue
2021-03-22T12:04:01
2025-04-01T04:35:34.683529
{ "authors": [ "tlykkeberg-grundfos", "woody-apple" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/5534", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1066521117
Improve doxygen Docker Image Problem The Doxygen GitHub action is not using the connectedhomeip/chip-build-doxygen image and also executes a script which only requires two external tools (doxygen and graphviz). Change overview This change fixes the source code to pass the doxygen execution and includes a lighter Docker image (from 2.83GB to 41.9MB) for Doxygen CI execution. Testing These changes was tested using the act -j doxygen command Fast tracking, given this is updating the docker image.
gharchive/pull-request
2021-11-29T21:26:49
2025-04-01T04:35:34.685968
{ "authors": [ "electrocucaracha", "woody-apple" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/pull/12337", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
659411128
error: invalid conversion from 'const char*' to 'char*' [-fpermissive… …] in screen-framework Problem While building the ESP32 demo on my Mac. Adding @mlepage-google since I believe it may knows this code. The casts are necessary to compile because the TFT functions incorrectly take non-const arguments. The correct fix of course is to fix the upstream TFT repo so it doesn't take non-const arguments. I have already done the correct fix, so this workaround should not be necessary. If you still cannot build, it's likely your sub-repo needs to be updated? https://github.com/jeremyjh/ESP32_TFT_library/commit/f4e07be511d1bf1c54ae657d3248c352de6b93b1
gharchive/pull-request
2020-07-17T16:38:52
2025-04-01T04:35:34.688186
{ "authors": [ "mlepage-google", "vivien-apple" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/pull/1639", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1288120762
[BL602] Enable command line to flash chip Problem What is being fixed? Examples: Enable command line to flash chip Change overview Enable command line to flash chip Testing How was this tested? (at least one bullet point required) command line: cd third_party/bouffalolab/bl602_sdk/repo/tools/flash_tool ubuntu: ./bflb_iot_tool-ubuntu18 --chipname=BL602 --baudrate=115200 --port=/dev/ttyACM0 --pt=chips/bl602/partition/partition_cfg_4M.toml --dts=chips/bl602/device_tree/bl_factory_params_IoTKitA_40M.dts --firmware=../../../../../../out/bl602-light/chip-bl602-lighting-example.bin macos: ./bflb_iot_tool-macos --chipname=BL602 --baudrate=115200 --port=/dev/ttyACM0 --pt=chips/bl602/partition/partition_cfg_4M.toml --dts=chips/bl602/device_tree/bl_factory_params_IoTKitA_40M.dts --firmware=../../../../../../out/bl602-light/chip-bl602-lighting-example.bin Fast tracking platform changes.
gharchive/pull-request
2022-06-29T03:20:09
2025-04-01T04:35:34.691835
{ "authors": [ "jczhang777", "woody-apple" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/pull/20094", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
721638397
Drive the vscode esp32 wifi example through a separate script Problem The existing ESP32 script is the CI script, which has sideffects: clobbers the sdkconfig defaults on every run (would lose things like wifi configuration) builds two boards at once every time (extra work) every bloard clean is after a sdk config reset, so build is not incremental Summary of Changes Created a python script that can drive esp32 builds (used python to have prettier argument parsing, although in hindsight it required quite a few workarounds regarding environment) Changed tasks.json to reference this script instead. @mspang ? I believe there may be better ways of doing this (calling shell from python, knowing that the shell will call python in turn is odd), however maybe we could incrementally improve. For now this works better than before (is incremental, builds only what is needed). We could also convert it to just shell, for a cleaner feeling (but without as clean argument parsing). Either way, I would propose to do improvements in future PRs. @mspang ? I believe there may be better ways of doing this (calling shell from python, knowing that the shell will call python in turn is odd), however maybe we could incrementally improve. For now this works better than before (is incremental, builds only what is needed). We could also convert it to just shell, for a cleaner feeling (but without as clean argument parsing). Either way, I would propose to do improvements in future PRs. This looks very awkward to me, couldn't you split this two and have a activate_and_run_command.sh wrapper?
gharchive/pull-request
2020-10-14T17:14:50
2025-04-01T04:35:34.695776
{ "authors": [ "andy31415", "mspang" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/pull/3258", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1097645380
Dependency package not found I am trying to get the tool running by following the instructions the the main README.md and docs/instructions.md. In the root directory of the project I have tried executing npm ci and npm install, following from the documentation files listed above. Both of these fail to install the dependency https://nexus.tecnotree.com/repository/npm-public/listr2/-/listr2-3.13.5.tgz. Failing with the error npm ERR! code ENOTFOUND npm ERR! syscall getaddrinfo npm ERR! errno ENOTFOUND npm ERR! network request to https://nexus.tecnotree.com/repository/npm-public/listr2/-/listr2-3.13.5.tgz failed, reason: getaddrinfo ENOTFOUND nexus.tecnotree.com npm ERR! network This is a problem related to network connectivity. npm ERR! network In most cases you are behind a proxy or have bad network settings. npm ERR! network npm ERR! network If you are behind a proxy, please make sure that the npm ERR! network 'proxy' config is set properly. See: 'npm help config' I am running node version v16.13.1 and npm version 8.3.0. I have made sure that I am not behind a proxy and ran the following npm configuration commands npm config rm proxy npm config rm https-proxy npm config --global rm proxy npm config --global rm https-proxy This was fixed. There was some "tecnotree" nonsense that weasled it's way into the package.json at some point.
gharchive/issue
2022-01-10T09:23:25
2025-04-01T04:35:34.698307
{ "authors": [ "hicklin", "tecimovic" ], "repo": "project-chip/zap", "url": "https://github.com/project-chip/zap/issues/361", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2352519037
feat(erofs): initial commit for erofs support Fixes https://github.com/opencontainers/image-spec/issues/1190 What type of PR is this? Which issue does this PR fix: What does this PR do / Why do we need it: If an issue # is not available please add repro steps and logs showing the issue: Testing done on this change: Automation added to e2e: Will this break upgrades or downgrades? Does this PR introduce any user-facing change?: By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license. Hi @rchincha, I'm currently working on incremental builds and I plan to release erofs-utils v1.8 this month (with multi-threaded mkfs, incremental builds, Intel QPL, etc.), but anyway, I'd suggest use cgo for initial EROFS support (with formal liberofs APIs exported) for now (or even binary integration) since this go implemention seems somewhat incomplete (and maybe even broken.. I don't have enough time to look into that since other prioritied stuffs are on hands..) @hsiangkao I suspect that cgo path may be what ends up happening. Currently, just prototyping quickly to understand the interfaces and scope of changes/work. One thing I might need to mention here is that erofs supports external blobs or chunks since Linux 5.16 compared to Squashfs because Dragonfly Nydus once asked this feature to deduplicate data in chunks among different container images. I'm not sure if it's worthwhile to highlight this, you could just make a tiny metadata with external blobs (which can be used for multiple images) for reference. Currently only Nydus has userspace tools to generate chunk blobs, if that is interested in other use cases, I could seek more time to implement this in mkfs.erofs too. $ cat stacker.yaml build: from: type: docker url: docker://public.ecr.aws/docker/library/busybox:1.37-glibc $ /tmp/stacker build -layer-type erofs Creating new OCI Layout at "/data/hdd/rchincha/tmp/lxc/oci" preparing image build... imported file hashes (after substitutions): loading docker://public.ecr.aws/docker/library/busybox:1.37-glibc Copying blob a46fbb00284b done Copying config 27a71e19c9 done Writing manifest to image destination Storing signatures mkfs.erofs 1.4 c_version: [ 1.4] c_dbg_lvl: [ 2] c_dry_run: [ 0] filesystem build built successfully @hsiangkao Is there an equivalent of "unsquashfs" in erofs-utils? https://manpages.debian.org/testing/squashfs-tools/unsquashfs.1.en.html SYNOPSIS unsquashfs [OPTIONS] FILESYSTEM [files to extract or exclude (with -excludes) or cat (with -cat )] Hi! @hsiangkao Is there an equivalent of "unsquashfs" in erofs-utils? https://manpages.debian.org/testing/squashfs-tools/unsquashfs.1.en.html SYNOPSIS unsquashfs [OPTIONS] FILESYSTEM [files to extract or exclude (with -excludes) or cat (with -cat )] fsck.erofs --extract= FILESYSTEM to extract the whole filesystem... Currently, there is no exclude way but it can be added later.
gharchive/pull-request
2024-06-14T05:08:51
2025-04-01T04:35:34.750706
{ "authors": [ "hsiangkao", "rchincha" ], "repo": "project-stacker/stacker", "url": "https://github.com/project-stacker/stacker/pull/626", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
95047829
atomic-reactor fails on RHEL7 because of python-requests version The version of python-requests that is in RHEL7 (core OS package set), and derivative distros (CentOS, et al), is too old. I built an rpm from my git checkout with tito build --test --builder mock --arg mock=epel-7-x86_64 --rpm and then installed the resulting rpm. I get the following error: [root@losbs ~]# atomic-reactor create-build-image --reactor-tarball-path /usr/share/atomic-reactor/atomic-reactor.tar.gz /usr/share/atomic-reactor/images/dockerhost-builder/ buildroot Traceback (most recent call last): File "/usr/bin/atomic-reactor", line 5, in <module> from pkg_resources import load_entry_point File "/usr/lib/python2.7/site-packages/pkg_resources.py", line 3011, in <module> parse_requirements(__requires__), Environment() File "/usr/lib/python2.7/site-packages/pkg_resources.py", line 626, in resolve raise DistributionNotFound(req) pkg_resources.DistributionNotFound: requests>=2.2.1,<2.5.0 [root@losbs ~]# rpm -q python-requests python-requests-1.1.0-8.el7.noarch this is actually issue of docker-py; therefore we build python-requests and supply it from our own repo Yeah, I'm doing this currently with my COPR space but this isn't something we can to in EPEL because it will violate the policy on not overriding core RHEL packages. Is the plan to maintain this separately? I fear for what might potentially be broken by replacing a core RHEL package. I've discussed this with docker-py RHEL maintainer and am not sure what was the final statement. CC @lsm5 @arrfab CentOS SIGs can and do override base packages, I can add a new python-requests along with docker-py and atomic-reactor to the Virt SIG repositories. @maxamillion and @TomasTomecek would you like to be a part of CentOS virt SIG? RE: RHEL, @TomasTomecek @maxamillion I did patch docker-py earlier so that atomic could run well with an older python-requests. Looks like this would need another patch :\ . I vaguely remember a bug filed to update python-requests on RHEL 7, but can't find it atm :( @lsm5 I'm actually planning to patch python-docker-py in Fedora to don't depend pretty much on any version of python-requests: https://github.com/projectatomic/atomic-reactor/issues/226 @TomasTomecek please use @rhatdan's repo instead of original upstream moving forward. @lsm5 because of https://github.com/rhatdan/docker-py/commit/0fc135c6dbe994aae1530e96071ea9aa33d94c88? python-requests got an update in RHEL7 official to 2.6.0 -> python-requests-2.6.0-1.el7_1.noarch so now atomic-reactor is erroring the same way Fedora is in https://github.com/projectatomic/atomic-reactor/issues/226 @maxamillion I already reported this at https://bugzilla.redhat.com/show_bug.cgi?id=1249651 and am waiting for a fix there. Feel free to add yourself to CC of that bug to know when this gets fixed. BTW I believe the correct fix is pinning the dependencies in the specfile precisely the same way they are pinned in upstream. Then the package will fail to install if dependencies change (and even some compose tools may throw errors about unsatisfied dependencies). When the versions are pinned, then it's only a matter of patching upstream requirements (and the code) for dependencies that have different version in RHEL7/EPEL7. @bkabrda if i'm reading above comment right, if we have python-requests >= 2.5.3 in the python-docker-py specfile, it will make python-docker-py uninstallable on rhel7 which is just not an option, as there's atomic depending on python-docker-py can't we just upgrade python-requests on rhel7 and move on? ugh ..re-reading ...so we gotta have python-requests == 2.5.3, hmm :( ...will check on that :( :( @lsm5 the point is, that to do packaging correctly, you should do it this way: pin dependencies in the specfile to be precisely the same as dependencies in requirements.txt try to install the package see what breaks patch both requirements and the actual code to work with the version that is in RHEL 7 profit :) this appears to have been resolved, can others verify? if so, I'm good with closing Seems to be fixed. I can install python-docker-py from rhel7-server-extras-rpms and atomic-reactor from EPEL without problems.
gharchive/issue
2015-07-14T21:37:26
2025-04-01T04:35:34.824743
{ "authors": [ "TomasTomecek", "bkabrda", "lsm5", "maxamillion", "mmilata" ], "repo": "projectatomic/atomic-reactor", "url": "https://github.com/projectatomic/atomic-reactor/issues/225", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
255640385
Avoid traceback when pulp_pull fails Signed-off-by: Tim Waugh twaugh@redhat.com Release note: Bugfix: ** failure of the pulp_pull plugin no longer causes the store_metadata_in_osv3 plugin to fail
gharchive/pull-request
2017-09-06T15:12:39
2025-04-01T04:35:34.826524
{ "authors": [ "twaugh" ], "repo": "projectatomic/atomic-reactor", "url": "https://github.com/projectatomic/atomic-reactor/pull/817", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
190162644
We want to execute docker-storage-setup on the host system If docker-storage-setup is run from a Super Privileged Container (SPC) we want to execute the script on the host machine. $HOST indicates the mountpoint of the host root fs of the container. This patch will chroot to this rootfs and execute the script. @rhvgoyal PTAL This assumes that all the /usr/lib/docker-storage-setup files are installed on host. That means docker-storage-setup and all dependencies are installed on host. But AFAIK, we are not doing any of that? Yes you are right. I thought the script was all encompassing. Another version, this one just adds the $HOST/usr/bin and $HOST/usr/sbin to $PATH. Also changes a couple of other small parts of script. Probably need to look at other helper apps. We have decided not to do this for now, closing.
gharchive/pull-request
2016-11-17T21:00:10
2025-04-01T04:35:34.832233
{ "authors": [ "rhatdan", "rhvgoyal" ], "repo": "projectatomic/docker-storage-setup", "url": "https://github.com/projectatomic/docker-storage-setup/pull/173", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
86848529
docker-storage-setup.1: fix typos Signed-off-by: Alex Jia ajia@redhat.com :+1:
gharchive/pull-request
2015-06-10T05:40:34
2025-04-01T04:35:34.833264
{ "authors": [ "cgwalters", "chuanchang" ], "repo": "projectatomic/docker-storage-setup", "url": "https://github.com/projectatomic/docker-storage-setup/pull/41", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2726912147
V0.0.18 added ERC721 deployable and smart character token address to the artifacts as part of the deployer image output agreed! from v2 it can be written directly to json file from the scripts instead of console.logs
gharchive/pull-request
2024-12-09T12:31:38
2025-04-01T04:35:34.835132
{ "authors": [ "0xxlegolas" ], "repo": "projectawakening/world-chain-contracts", "url": "https://github.com/projectawakening/world-chain-contracts/pull/340", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1211855459
Is the -root-tld flag still working ? ./interactsh-server -domain example.com -hostmaster admin@example.com -ip X.X.X.X -listen-ip X.X.X.X -root-tld flag provided but not defined: -root-tld https://github.com/projectdiscovery/interactsh/issues/17 Is it now enabled by default ? @greckko flag name is updated with below: -wc, -wildcard enable wildcard interaction for interactsh domain (authenticated)
gharchive/issue
2022-04-22T05:54:51
2025-04-01T04:35:34.842649
{ "authors": [ "ehsandeep", "greckko" ], "repo": "projectdiscovery/interactsh", "url": "https://github.com/projectdiscovery/interactsh/issues/268", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
788500898
resolves infinite loop when received value was zero this commit solves the problem in projectdiscovery/naabu#123 hi, is there any plan to release a new version to fix this problems? @LazyMaple the latest release is after this commit - are you still having this issue?
gharchive/pull-request
2021-01-18T19:27:58
2025-04-01T04:35:34.843965
{ "authors": [ "LazyMaple", "nicolascb", "olearycrew" ], "repo": "projectdiscovery/ipranger", "url": "https://github.com/projectdiscovery/ipranger/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }