id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2 values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1128373194 | New command: Ensure that the particular Azure AD app registration exists and updates its properties if necessary
Usage
m365 aad app ensure [options]
Description
Ensures that the particular Azure AD app registration exists and updates its properties if necessary
Options
Option
Description
--manifest <manifest>
Azure AD app manifest as retrieved from the Azure Portal to configure the app registration from
Additional Info
The command is a combination of aad app get, aad app add and aad app set.
This command checks if an Azure AD app registration with the specified ID exists.
If no Azure AD app registration is found, this command will create one using the information from the manifest.
If an Azure AD registration is found, this command will update its properties using the information from the manifest.
@waldekmastykarz i'll try this one, please assign it to me
All yours! Thank you!
@waldekmastykarz
aad app add command ignores both , 'id' and 'appId' if provided in the manifest
https://github.com/pnp/cli-microsoft365/blob/55e5dfb910c3abcede4db24cb29b041052a61ada/src/m365/aad/commands/app/app-add.ts#L175-L180
So if we don't find the 'id' property in the provided manifest of the ensure command, do we first try to get the app using 'appId' property (if present) or straightaway consider the execution as a new app creation ?
I'd say we consider the app non-existent and consider new app creation. As far as I know it's not possible to create an app with a predefined objectID or clientID so in that sense they're equal and checking either one is sufficient.
ok @waldekmastykarz we will go with only 'id' as required in the manifest to determine existing app.
Also, if multiple 'identifierUris' are provided in manifest.json, aad app add command configures all, however, aad app set command only accepts a single --uri and replaces all the existing identifierUris with the one specified.
This way, the set operation of the ensure command will not configure multiple identifierUris (not sure what are the use cases could be). Should we stop the command execution if there are multiple identifierUris in manifest ?
@vipulkelkar in the Azure Portal it seems like you can configure only one application ID URI. Do you mean redirect URI for auth by any chance?
Good find! I think we should consider the manifest as the desired state, so after running the ensure command, the AAD app you get should match what's in the manifest.
aad set only accepts a single --uri. So if we use set command in ensure, we will not be able to configure multiple identifieruris to ensure the manifest desired state.
Should we first extend the set command to accept comma separated URI's ? (similar to --redirecturis parameter ).
Using the set command as-is in ensure makes sense. My vote is to extend set command to accept the manifest.
@waldekmastykarz Should I create a separate issue to extend the aad app set command to accept manifest ?
Please do @vipulkelkar
created #3153 , we can put this one on-hold for now till the other one is implemented.
Hi @vipulkelkar, I see the issue this is waiting for has been closed as completed in v5.4 of the CLI.
I'll remove the on hold here. Are you still planning on working on it?
Hi @martinlingstuyl , this issue is waiting on #3153 as mentioned above, which is still in progress.
.
The mention of #3333 (which originated from this issue) might have caused the confusion.
You're correct. I've put it on hold again 👍
| gharchive/issue | 2022-02-09T10:50:30 | 2025-04-01T04:35:31.686271 | {
"authors": [
"martinlingstuyl",
"vipulkelkar",
"waldekmastykarz"
],
"repo": "pnp/cli-microsoft365",
"url": "https://github.com/pnp/cli-microsoft365/issues/3035",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
516626351 | Create environment request form
The user must be able to request an environment to the podcast API.
Closed by #6
| gharchive/issue | 2019-11-02T15:18:38 | 2025-04-01T04:35:31.807563 | {
"authors": [
"hericlesme"
],
"repo": "pod-cast/podcast-client",
"url": "https://github.com/pod-cast/podcast-client/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1201824110 | Links at top of screen in Chrome
The links in the black bar at the top of the POD screen are displaying correctly in Firefox. In Chrome, though, they're not correctly spaced, and overlap each other. Resizing the window doesn't make a difference.
Screenshots attached.
Pull Request: https://github.com/pod4lib/aggregator/pull/582
@bobpersing thank you for reporting this. The fix is applied to the main branch of the project and will be deployed to production next Monday (at the latest) with dependency updates.
| gharchive/issue | 2022-04-12T13:06:42 | 2025-04-01T04:35:31.810359 | {
"authors": [
"bobpersing",
"corylown",
"ktamaral"
],
"repo": "pod4lib/aggregator",
"url": "https://github.com/pod4lib/aggregator/issues/579",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2309913683 | feat(HMS-4156): confirmation checkbox in delete domain confirm modal
Add confirmation checkbox that enables delete button to warn user that the action cannot be undone.
In unit test, some jest calls were replaced by newer variants as the older were deprecated.
I tested the change locally but the UI was unchanged. Maybe I missed something? Screenshots look great though!
| gharchive/pull-request | 2024-05-22T08:31:15 | 2025-04-01T04:35:31.812089 | {
"authors": [
"frasertweedale",
"pvoborni"
],
"repo": "podengo-project/idmsvc-frontend",
"url": "https://github.com/podengo-project/idmsvc-frontend/pull/68",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2259134916 | Unnecessary "observe" warning
Description
Warning "An "observe" was called from another "observe" closure, which can lead to over-observation and unintended side effects." was added in version 1.9.3 and become at unnecessary cases
I attach example project where you can see that bug
BadObserveExample.zip
Checklist
[ ] I have determined whether this bug is also reproducible in a vanilla SwiftUI project.
[X] If possible, I've reproduced the issue using the main branch of this package.
[X] This issue hasn't been addressed in an existing GitHub issue or discussion.
Expected behavior
Warning doesn't appear
Actual behavior
Warning appears
Steps to reproduce
Just run app and wait. Screens will navigate by themselves
The Composable Architecture version information
1.9.3
Destination operating system
iOS 17.4.1
Xcode version information
15.3
Swift Compiler version information
swift-driver version: 1.90.11.1 Apple Swift version 5.10 (swiftlang-5.10.0.13 clang-1500.3.9.4)
@nmalevich We recently added this runtime warning because nested observes can have unintended consequences. In this case, a viewDidLoad is evaluated in the same stack as the presenting view controller's observe, which means all child state evaluated in the child controller's observe will be observed by the parent.
To work around the problem, you can move observe into viewWillAppear instead, and manage the lifetime of the observation token, or you can dispatch the observation a tick in viewDidLoad:
DispatchQueue.main.async {
observe { [weak self] in
…
}
}
We're still deciding if this warning is too heavy-handed, but we don't consider it a bug at the moment. Because of this I'm going to convert this to a discussion.
| gharchive/issue | 2024-04-23T15:08:56 | 2025-04-01T04:35:31.912272 | {
"authors": [
"nmalevich",
"stephencelis"
],
"repo": "pointfreeco/swift-composable-architecture",
"url": "https://github.com/pointfreeco/swift-composable-architecture/issues/3005",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2101076297 | Unable to use @DependencyEndpoint with a typealias
Description
Not sure if this is a dependencies bug, macros bug, or macros feature?
works!
extension MyFeature {
@DependencyClient
struct Client {
var refreshRequest: @Sendable (MyFeaturesRequest) async throws -> MyFeaturesRequest.Content
}
}
Doesnt work
extension MyFeature {
@DependencyClient
struct Client {
var refreshRequest: RequestOf<MyFeaturesRequest>
}
}
typealias RequestOf<R: NetworkRequestProtocol> = @Sendable (R) async throws -> R.Content
Checklist
[X] I have determined whether this bug is also reproducible in a vanilla SwiftUI project.
[ ] If possible, I've reproduced the issue using the main branch of this package.
[X] This issue hasn't been addressed in an existing GitHub issue or discussion.
Expected behavior
I'd expect the DependencyEndpoint macro to work, when the typealias is representing a closure.
Actual behavior
The DependencyEndpoint macro doesn't apply when the typealias is representing a closure.
Steps to reproduce
No response
Dependencies version information
No response
Destination operating system
No response
Xcode version information
No response
Swift Compiler version information
No response
Hi @JOyo246, this is a limitation of macros. The @DependencyClient macro can only see the syntax it is directly attached to, so it has no idea what RequestOf is. So unfortunately you cannot use type aliases in this way.
Since this isn't an issue with the library I am going to convert it to a discussion.
| gharchive/issue | 2024-01-25T19:56:59 | 2025-04-01T04:35:31.918288 | {
"authors": [
"JOyo246",
"mbrandonw"
],
"repo": "pointfreeco/swift-dependencies",
"url": "https://github.com/pointfreeco/swift-dependencies/issues/178",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1233077490 | Request header unexpectedly mutated
In this example, I'm supplying an uppercased request header but it's getting lowercased somewhere within the parsing library. I couldn't easily figure out where this was happening to fix with a PR so opening as an issue!
enum TestRoute {
case testing(value: String)
}
let testingRouter = Route(.case(TestRoute.testing)) {
Path {
"testing"
}
Headers {
Field("UPPERCASED-HEADER")
}
}
func test() async throws {
let request = try testingRouter.request(
for: TestRoute.testing(value: "VALUE")
)
print(request.allHTTPHeaderFields!)
}
printed:
["uppercased-header": "VALUE"]
expected:
["UPPERCASED-HEADER": "VALUE"]
@eappel HTTP header names are case insensitive, so the library currently lowercases things for consistency. Have you encountered an issue in the wild where this is a problem?
Yeah the API I'm dealing with is case-sensitive 😞
@eappel I think we should probably change this behavior on our end, too!
@eappel We have an idea that should make it easy to fix this in the future, but we haven't quite finalized the tools yet. We'll leave the issue open for now, but will hopefully get to close it in the next month or two!
| gharchive/issue | 2022-05-11T19:14:44 | 2025-04-01T04:35:31.921260 | {
"authors": [
"eappel",
"stephencelis"
],
"repo": "pointfreeco/swift-url-routing",
"url": "https://github.com/pointfreeco/swift-url-routing/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1487987090 | 🛑 Pokanop Apps Frontend is down
In c47bbae, Pokanop Apps Frontend (https://pokanop.com) was down:
HTTP code: 500
Response time: 3149 ms
Resolved: Pokanop Apps Frontend is back up in 2106e87.
| gharchive/issue | 2022-12-10T05:22:13 | 2025-04-01T04:35:31.930224 | {
"authors": [
"saheljalal"
],
"repo": "pokanop/uptime",
"url": "https://github.com/pokanop/uptime/issues/62",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1000975924 | Support triple backtick
Treats ``` as a single token
when do you use triple backtactics?
markdown
| gharchive/pull-request | 2021-09-20T13:31:12 | 2025-04-01T04:35:31.971677 | {
"authors": [
"pokey"
],
"repo": "pokey/cursorless-vscode",
"url": "https://github.com/pokey/cursorless-vscode/pull/279",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2404367774 | 🛑 Pokko - auth service is down
In b25cf78, Pokko - auth service (https://id.pokko.io) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Pokko - auth service is back up in be605bd after 4 minutes.
| gharchive/issue | 2024-07-11T23:45:24 | 2025-04-01T04:35:31.974039 | {
"authors": [
"brendanmckenzie"
],
"repo": "pokkocms/pokko-status",
"url": "https://github.com/pokkocms/pokko-status/issues/124",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1762030913 | [Persistence] Update KVStore to use Badger wrapper functions
Description
This PR introduces unit tests to cover the KVStore's functionality as well as updating the KVStore logic to use the Badger wrapper functions update, view, etc.
This fixes the issues around deleting keys from the KVStore.
Issue
Fixes N/A
Type of change
Please mark the relevant option(s):
[ ] New feature, functionality or library
[x] Bug fix
[x] Code health or cleanup
[ ] Major breaking change
[ ] Documentation
[ ] Other
List of changes
Update KVStore to use badger's wrapper functions
Add KVStore unit tests
Testing
[x] make develop_test; if any code changes were made
[x] make test_e2e on k8s LocalNet; if any code changes were made
[x] e2e-devnet-test passes tests on DevNet; if any code was changed
[x] Docker Compose LocalNet; if any major functionality was changed or introduced
[x] k8s LocalNet; if any infrastructure or configuration changes were made
Required Checklist
[x] I have performed a self-review of my own code
[x] I have commented my code, particularly in hard-to-understand areas
[x] I have added, or updated, godoc format comments on touched members (see: tip.golang.org/doc/comment)
[x] I have tested my changes using the available tooling
[ ] I have updated the corresponding CHANGELOG
If Applicable Checklist
[ ] I have updated the corresponding README(s); local and/or global
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] I have added, or updated, mermaid.js diagrams in the corresponding README(s)
[ ] I have added, or updated, documentation and mermaid.js diagrams in shared/docs/* if I updated shared/*README(s)
NOTE @Olshansk, @dylanlott is currently working on savepoints/rollbacks that touches the same code and this PR will probably only be used as a reference as he is changing the logic their. The new wrapper functions fix the bug with Delete() but need to be integrated with savepoints. Will wait on merging this until @dylanlott gives an update
This PR is already complete and ready to be merged in, while savepoints & rollbacks PR is 1+ weeks away.
IMO we should merge this in and iterate on top of it. However, I will defer to @dylanlott to decide.
This PR is already complete and ready to be merged in, while savepoints & rollbacks PR is 1+ weeks away.
IMO we should merge this in and iterate on top of it. However, I will defer to @dylanlott to decide.
I agree with merging this and addressing the logic changes necessary for savepoints later. It shouldn't hold this up.
| gharchive/pull-request | 2023-06-17T23:29:38 | 2025-04-01T04:35:31.984598 | {
"authors": [
"Olshansk",
"dylanlott",
"h5law"
],
"repo": "pokt-network/pocket",
"url": "https://github.com/pokt-network/pocket/pull/838",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
708159647 | Ensure qform/sform copy doesn't modify zooms (or warn if it does)
What version of fMRIPrep are you using?
20.2.0rc0
What kind of installation are you using? Containers (Singularity, Docker), or "bare-metal"?
TACC Run
ds000229 - sub-07
Did fMRIPrep generate the visual report for this particular subject? If yes, could you share it?
Details
BOLD looks significantly larger than T1w, and sform was copied to qform. This has the potential to change pixdim, since qform is built from quaternions, offsets and pixdim.
This is a note to verify, but probably shouldn't block.
This particular instance could be a visual effect of #2282.
| gharchive/issue | 2020-09-24T13:12:34 | 2025-04-01T04:35:32.012193 | {
"authors": [
"effigies",
"oesteban"
],
"repo": "poldracklab/fmriprep",
"url": "https://github.com/poldracklab/fmriprep/issues/2284",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
570909751 | FIX: Amend some error types
Trying to normalize a little between TypeErrors and ValueErrors.
Let's check on all error types when merging into nibabel (#110).
| gharchive/pull-request | 2020-02-25T23:17:01 | 2025-04-01T04:35:32.013435 | {
"authors": [
"oesteban"
],
"repo": "poldracklab/nitransforms",
"url": "https://github.com/poldracklab/nitransforms/pull/63",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2516507471 | Tidying
Tox runs ruff format -diff.
Remove suggestion to run tox -e reformat from readme.
GitHub action automatically applies linting (ruff, trailing-whitespace and end-of-file-fixer) to PRs. Note that a .pre-commit-config.yaml file was added to the repo for running pre-commit. Also, bots must be given read and write permissions to make commits:
This seems to work nicely.
Commit 9556263 was made automatically by the bot which applied ruff to a few files. The branch now passes the tox env quality (specifically ruff format --diff).
Lint GitHub Action fails because it doesn't have read and write permission to make commits:
Error: Error: To https://github.com/poliastro/czml3
! refs/heads/tidying:refs/heads/tidying [remote rejected] (refusing to allow a GitHub App to create or update workflow `.github/workflows/workflow.yml` without `workflows` permission)
See top comment to apply the required change to allow the action to make commits.
Reading from the UI, these are the default permissions. "You can specify more granular permissions in the workflow using YAML."
Can't we just change the YAML?
Also while we're at it... is this essentially equivalent to https://pre-commit.ci/ ?
| gharchive/pull-request | 2024-09-10T13:36:42 | 2025-04-01T04:35:32.025898 | {
"authors": [
"Stoops-ML",
"astrojuanlu"
],
"repo": "poliastro/czml3",
"url": "https://github.com/poliastro/czml3/pull/153",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
267876932 | Simplify 2D plotting
After @Juanlu001 creating the sample function, simplification of 2d plotting got significantly easier.
The changes are the following:
Now set_frame works also when a frame already exists, by removing all the lines and patches (not the legend), and replotting again.
a list of tuples (orbit, legend) is stored to achieve the former.
sample function is used in plotting, reducing code complexity.
Closely related to #218
Also, there is one mypy failure and one test failure in appveyor, check them out.
Notebook execution is failing in all platforms. I'm trying to know why.
The problem lies in sampling an hyperbolic orbit. Right now the .sample method assumes a closed orbit, and that's a bug. Previously, the logic for sampling hyperbolic orbits was here:
https://github.com/anhiga/poliastro/blob/bde55913ae3d3315e57ff1f553c9bd63de214ebe/src/poliastro/plotting.py#L170-L185
Which, by the way, should go away as well.
And if that code goes away, then num_points should go away too!
I tried to take over but there are a handful of errors, let's see what we can do.
Well! I made some key changes and things look better now. Please take a look, and be so kind to add some tests :grin:
Oops, conflicts appeared after merging #271.
| gharchive/pull-request | 2017-10-24T02:29:45 | 2025-04-01T04:35:32.030483 | {
"authors": [
"Juanlu001",
"anhiga"
],
"repo": "poliastro/poliastro",
"url": "https://github.com/poliastro/poliastro/pull/269",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
316248778 | Added an option to choose propagator in OrbitPlotter
Now we can choose appropriate propagator in OrbitPlotter while calling plot(). Solves #347
Just for reference
It's great that we have so much flexibility, but perhaps it's too much. We are already setting the number of sampling points in the OrbitPlotter constructor, so what do you think of keeping the method as a plotter property as well?
I think it's fine as then issues like the one you faced won't happen and the user will be able to use various propagators in OrbitPlotter as well..
Should I add it to OrbitPlotter constructor?
But if we're setting the method per-plot instead of per-frame, what about the number of points? Shouldn't it be set per-plot as well? My concern here is consistency. Four options:
We leave things as they are in this pull request. I think it's inconsistent, unless someone convinces me of the contrary.
We move num_points to each plot call. I think it's bad because it makes the signature too long and, in any case, if someone is so concerned about it they can already sample in any way they want and use plot_trajectory.
We move method to the OrbitPlotter constructor. This prevents the user from selecting the method per-plot. I ask: is this so important?
We do nothing, and tell the user "if you want to change the propagator, just use plot_trajectory".
Thoughts?
I don't have much idea about consistency, but num_points is something that affects the quality of plot. Changing it can either introduce steps in the plot.
I think if seems inconsistent, we can move it to the constructor.
We can leave it too :P
I am not much clear about this.
I just remembered that, in STK, you can indeed select a different propagator for each orbit in one scenario:
So I've changed my mind to accept this. ~@shreyasbapat would you like to change the propagator in the last orbit of the Florence notebook to fix it? (See the errors here http://docs.poliastro.space/en/latest/examples/Catch that asteroid!.html)~ I remember you told me you were on exams, so I'm going to merge this and take care of the rest myself.
Thanks a lot @Juanlu001 . I will be actively up after 1st May Again! Between this, whenever I get time, I will try to do something!
| gharchive/pull-request | 2018-04-20T12:31:08 | 2025-04-01T04:35:32.037336 | {
"authors": [
"Juanlu001",
"shreyasbapat"
],
"repo": "poliastro/poliastro",
"url": "https://github.com/poliastro/poliastro/pull/350",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
257242926 | 5.9 Adecuar el login al entorno web para el administrador
Es necesario adecuar el formulario de login para que también pueda ser usado de manera web por el administrador
Al ejecutar el issue #96 en la que se pretende crear eventos se hizo el inicio de sesión debido a que se considero necesario para continuar con dicha tarea. Por lo anterior esta tarea puede ser cerrada.
| gharchive/issue | 2017-09-13T02:59:49 | 2025-04-01T04:35:32.039364 | {
"authors": [
"DanteVDrako",
"drvelasquezq"
],
"repo": "polieduco/practicaaplicada20172",
"url": "https://github.com/polieduco/practicaaplicada20172/issues/102",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1511697173 | Уточнение типов
https://github.com/polina-nn/news/blob/805447a13cea1bb97fbf977c85472a0541ee8e2c/src/EndPoints/GetAuthorsNewsSearchList.hs#L81
Лишние уточнения типов. Присутствуют очень часто. Если не уверена что тип выводится, то лучше поставь hls и уточняй там где это действительно необходимо.
Добрый день! Вроде больше лишних выводов типов нет
| gharchive/issue | 2022-12-27T11:07:35 | 2025-04-01T04:35:32.040749 | {
"authors": [
"pavelzarubin",
"polina-nn"
],
"repo": "polina-nn/news",
"url": "https://github.com/polina-nn/news/issues/16",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2531553298 | [Polkadot WIki Migration] Set up a RPC node
This page was migrated from https://wiki.polkadot.network/docs/maintain-rpc
Scan for Vale false positives. Resolved convos = made it on the update list and should go away now. I left comments for the legit flags with suggested edits. Thank you!
Alrighty, made a bunch of edits. Need a new content review from @0xLucca (especially pruned nodes part as I Googled it!) and format review from @eshaben
This should be good to know. Tech review was dismissed as stale so we'll need one tech and one format to merge. Thanks!
| gharchive/pull-request | 2024-09-17T16:09:26 | 2025-04-01T04:35:32.045362 | {
"authors": [
"CrackTheCode016",
"dawnkelly09"
],
"repo": "polkadot-developers/polkadot-docs",
"url": "https://github.com/polkadot-developers/polkadot-docs/pull/32",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1046835730 | [feat]: implement public/private key authentication
This is an issue to track the implementation of public/private key authentication for Stormi
Using HMAC with a private key seems a more practical way to do this. I may consider replacing RSA with HMAC hashing or removing hashing completely.
I may consider removing hashing completely and just sticking with a base64 encoded master key that would be provided by the config.yaml file.
Main advantages
Easier configuration process
Faster authentication processing times
Here's a diagram
I have made the final decision to remove private/public key authentication. Users will now be authenticated via a base64 encoded user/password string.
| gharchive/issue | 2021-11-07T19:58:22 | 2025-04-01T04:35:32.072000 | {
"authors": [
"michaelgrigoryan25"
],
"repo": "polygon-isecure/stormi",
"url": "https://github.com/polygon-isecure/stormi/issues/1",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
454217158 | floor level not right on alpha5 quest
Context : oculus quest
Version : Alpha5
Heya
I can't seem to have the floor level set right in steamVR and its games. It's like i'm sitting while actually standing. I'm sure there's a way to set that somewhere but can't figure out where
Great job on the software, it works mostly great !
I would highly recommend using SteamVR Advanced Settings. There is a floor fix option in there that works great with the quest. I use it everytime.
I found that by doing standing Room Setup with correct height (don't put headset on ground and do 0 as that doesn't seem to stick?) it was still basically correct the next time, maybe that will help
Use this and go to the floor fix settings, super easy and works (I've used it too).
https://github.com/OpenVR-Advanced-Settings/OpenVR-AdvancedSettings/releases/tag/v3.1.0
As my computer and "play room" is not on the same level, I find it hard to do the SteamVR setting as you need to follow the steps on the PC. Is there another way to do this?
Advanced settings floor fix nailed it
I don't remember the exact tab but on one of them there is an x y z thing that you can change, in my case i was on the floor level and not in the circle so i change the y from 0 to 1.8 (change it until you feel like you the right high) and the z to 1.5 (you may be in another place so just play with the xyz until you in the circle in the right high)
I don't remember the exact tab but on one of them there is an x y z thing that you can change, in my case i was on the floor level and not in the circle so i change the y from 0 to 1.8 (change it until you feel like you the right high) and the z to 1.5 (you may be in another place so just play with the xyz until you in the circle in the right high)
Yea, but I live in a contry where we use comma as decimater, so my PC does as well, so when I enter "1.8" in ALVR it changes automatically to "1,8", which only raise me "1 meter", when I then switch tab and return, it changes to "18", which sends me to the skies.
This is created as an issue a long time ago, but it was never fixed, making this function useless to me
@Yoshidk In Steam VR Advanced Settings you go to room adjustment (I'm pretty sure), then floor fix. You don't enter in any number. You place a controller on the floor, then click on floor fix. It will set the floor to where your lowest controller is. Just make sure before you click on fix floor that you take a good step back so when you click the button, the cameras on the quest can still see the controller on the floor. You may have to do each session, but it takes only a second to do.
@Yoshidk In Steam VR Advanced Settings you go to room adjustment (I'm pretty sure), then floor fix. You don't enter in any number. You place a controller on the floor, then click on floor fix. It will set the floor to where your lowest controller is. Just make sure before you click on fix floor that you take a good step back so when you click the button, the cameras on the quest can still see the controller on the floor. You may have to do each session, but it takes only a second to do.
Ok, I will have to try this. Is this "fix floor" button available after starting the SteamVR through ALVR from within the headset? Or would I still have to click somewhere on the PC?
All done within the headset. Open steamvr menu, then you should see advaned settings at the bottom.
Advanced settings floor fix nailed it
Close the issue here please, if you are all set.
| gharchive/issue | 2019-06-10T15:04:54 | 2025-04-01T04:35:32.078706 | {
"authors": [
"Myhamof",
"POEDaley",
"Yoshidk",
"falcoriss",
"iamliamiam",
"machenmusik",
"zxsdklfhskyrvk"
],
"repo": "polygraphene/ALVR",
"url": "https://github.com/polygraphene/ALVR/issues/404",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
101906199 | glucose starvation vs. glucose limitation
In #1687, we discussed "glucose limitation" and whether it's the same as "glucose starvation" or not. We eventually added FYPO:0003743, decreased cell population growth during glucose limitation, with a comment explaining glucose limitation as below about 4.4 mM (0.08%), but above zero.
Since then, we've added three other terms using "glucose limitation" in the names, with the same comment (FYPO:0003938, FYPO:0004168, FYPO:0004765). All four terms are used in annotations.
Now the question has (re-)arisen whether we should continue to make a distinction between limitation and starvation for glucose. In #2298 Antonia said:
"I don't think we need to distinguish between glucose starvation and limitation, I think it is the same response."
I (Midori) added the new terms requested in #2298, and am transferring the follow-up discussion of the starvation/limitation question here.
From the follow-up exchange:
Midori: The starvation vs limitation split comes from #1687 -- I got the impression that Val wanted to make the distinction (so she probably still does).
M: #1687 includes the question about glucose limitation vs. glucose starvation, and is the one where the first glucose limitation term was added. I asked Val whether "glucose starvation" would do for her request, and if not, how to define the difference between "glucose starvation" and "glucose limitation". I don't think there's a great answer in the ticket comments. I ended up going with starvation = no glucose at all and limitation = below 4.4 mM (0.08%) but not zero, but I've been aware all along that that information really belongs in conditions.
Val: I think it is important to keep the distinction but I don't know how best to do it. Happy for any solution.
V: Basically the limitation was not starvation, but 4.4 mM which would be considered low glucose.
I can't see the original ticket but we had a long discussion among ourselves and it was based on suggestions from the authors, who wanted us to represent that the response is different from the normal level of glucose supplied (which is really excess glucose).
More transferred from #2298:
Val:
This is the session the term was created for
http://curation.pombase.org/pombe/curs/b5db93caaa51792d/ro/
there are other genes annotated to
decreased cell population growth during glucose limitation
which may not be 'low glucose' as defined here, namely for gene products
gpa2
sds23
ssp1
trt1
Val:
ssp1 and sds23 are OK, from hanyu et al
so only gpa2 and trt1 will need checking (when we decide the best way to proceed).
I'd like to keep a term to represent this in some way if it is possible, as it does seem to be a specific adaptation to low glucose (probably normal in the wild when glucose is unlikely to be in such excess)
More transferred from #2298:
Val: Is the "adaptation to low glucose" different from the response to "no glucose"?
Antonia: I can't imagine it is (surely, in no glucose, the cells would fire the "low glucose" response? i.e. a starvation response)
Val: Is the "adaptation to low glucose" different from the response to "no glucose"?
Antonia: I can't imagine it is (surely, in no glucose, the cells would fire the "low glucose" response? i.e. a starvation response)
Val: Nope its different. See the old thread. IIRC the argument is that this low level of glucose is what fission yeast more than likely encounter in the wild. What is normally observed in laboratory setting with excess glucose is probably different from the pathways which are activated in the wild in low glucose.
Val: http://www.ncbi.nlm.nih.gov/pubmed/24815688?dopt=Abstract
Now commenting for myself at last ...
I'm not convinced that "the old thread" (#1687) did satisfactorily answer the question of whether there's a difference between glucose limitation (aka low glucose, aka less than about 4.4 mM) and glucose starvation. I also don't see how it helps to know that laboratory media usually have lots more glucose than yeasties typically find in the wild.
What is the difference in response to low glucose versus none at all?
No glucose at all -> cells die after a few days
"low" glucose -> cells reprogramme their gene expression profile compared to when in "high glucose".
I don't like having the def in there about 4.4 mM glucose. This level is chosen because:
"we previously reported that fission yeast can proliferate at essentially normal rates in media containing as little as 0.08% (4.4 mM) glucose, which is equivalent to the blood glucose level in a healthy human before breakfast"
It doesn't really have anything to do with pombe in the wild (I'd guess pombe in the wild would grow on complex mixtures of different stuff, and the levels of each carbon source would fluctuate)
I just want to make the distinction between "low enough glucose where the "low glucose" programme is active" & "high enough glucose so the "high glucose" programme is active"
Maybe in reality there is even a grey-zone between these two states (some parts of each programme might overlap?)
I don't like having the def in there about 4.4 mM glucose.
I totally agree; that's a condition, not part of the phenotypes.
And what I want to know is whether "low enough glucose where the "low glucose" programme is active" is the same as "glucose starvation". If there's no glucose (and no other carbon), do cells spend their remaining few days doing the same sort of gene expression as cells in low glucose? Or are there actually three situations: lots, a little, and none?
It doesn't really have anything to do with pombe in the wild (I'd guess pombe in the wild would grow on complex mixtures of different stuff, and the levels of each carbon source would fluctuate)
Agree. "The wild" isn't even remotely homogeneous or constant over time or space.
Maybe in reality there is even a grey-zone between these two states
I bet there is. This is biology, after all, so it's always safe to assume "it's more complicated than that".
Umm, I'm only guessing now but I'd guess they would first go through with the normal changes (upregulate high affinity glucose (and other other carbon transporters?)), that would obviously not "work" (because there is no carbon added in the media) so maybe they would upregulate autophagy, and maybe try and mate more than cells kept at 'some' glucose level?
phew, I don't think my head is getting any less melty ...
In the abstract
S. pombe arrests cell cycle progression when transferred from media containing 2.0% glucose to media containing 0.1%. After a delay, S. pombe resumes cell division at a surprisingly fast rate, comparable to that observed in 2% glucose. We found that a number of genes, including zinc-finger transcription factor Scr1, CaMKK-like protein kinase Ssp1, and glucose transporter Ght5, enable rapid cell division in low glucose.
The suspect it is due to a shift from glycolysis to respiration to maintain ATP production despite limiting glucose
I'm happy to capture this anyway, if you think that glucose limitation is inappropriate. I think it is worth capturing though.
I don't think glucose limitation is inappropriate.
I think that starvation = limitation.
= there is no point in maintaining two separate terms
Ok, it sounds like a merge
OK, it sounds like Val & Antonia now agree that "glucose starvation" and "glucose limitation" are the same, at least for our phenotype annotation purposes. So what I'll do is:
Standardize on "glucose starvation" in term names and "glucose limitation" in exact synonyms;
Do any necessary merges;
Make sure definitions are consistent;
Delete comments with guff about glucose concentrations that belongs in conditions.
Shout soon if that would cause problems — next FYPO session is tomorrow ;)
didn't need any merges after all -- terms that used "limitation" up until a few minutes ago weren't otherwise identical to any "starvation" terms
edit file: 60400ab363d2deef2a6a0f3c814fc2c7142c5d6b
release: ad1a93ec3a6e446c7aa0632a50e74379931a8778
| gharchive/issue | 2015-08-19T14:24:24 | 2025-04-01T04:35:32.106409 | {
"authors": [
"Antonialock",
"ValWood",
"mah11"
],
"repo": "pombase/fypo",
"url": "https://github.com/pombase/fypo/issues/2300",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
179428118 | interphase and mitotic microtubule nucleation PMID:19001497
Interphase (cytoplasmic) microtubules, can be nucleated from the cytoplasm, or from the SPB.
There are mutants where 'cytoplasmic microtubule nucleation" is abolished, BUT interphase microtubules are present in the cytoplasm, because they
"These intranuclear MTs eventually escape into the cytoplasm, acting as progenitors of essentially all cytoplasmic MTs in the mutant cells and forming abnormal MT bundles that often curve around the cell tips"
Also, cytoplasmic (astral) microtubules exist after during mitosis?
so not all cytoplasmic microtubules are interphase microtubules.
Soo....I think we need to make
NTR abnormal interphase microtubule nucleation from cytoplasm
--NTR abnormal interphase microtubule nucleation from iMTOC
--NTR abnormal interphase astral microtubule nucleation (def include: occurs from cytoplasmic side of SPB)
plus "abolished" and "decreased" for all
and
"NTR abnormal/abolished astral microtubule nucleation"
--NTR abnormal interphase astral microtubule nucleation (def include: occurs from cytoplasmic side of SPB)
--NTR abnormal mitotic astral microtubule nucleation (def include: occurs from cytoplasmic side of SPB)
"interphase cytoplasmic microtubule nucleation" might be better wording?
It's going to help a lot if you can define what is and isn't covered by "interphase microtubule".
If some astral microtubules hang around after mitosis, then either (a) "interphase microtubule" includes those astral microtubules, or (b) "interphase microtubule" doesn't simply mean "a microtubule that exists during interphase".
I will need to check what people refer to as interphase microtubules.
@jvhayles do you know? I can ask Ken...
I have altered the above request a little.
We should probably stop calling them "interphase microtubules". This can be a related synonym.
We should refer to "cytoplasmic microtubules during interphase" . But "Interphase cytoplasmic microtubules" can't NOT include the cytoplasmic microtubules generated from astral microtubules.
Does this help:
Only a small number of cytoplasmic MTs are present at any given time in fission yeast (Hoog et al., 2007), and these can be imaged by expressing GFP-tubulin at physiological concentrations. During interphase, MTs are nucleated not only from the spindle pole body but also from several interphase microtubule organizing centre ‘satellite’ particles (iMTOCs) present on the nuclear envelope, in the cytoplasm and on MTs themselves. During late mitosis, cytoplasmic astral MTs analogous to those seen in budding yeast are nucleated from the duplicated SPBs.
OK, I think the existing terms can all be adjusted without much pain to "cytoplasmic microtubules [or microtubule nucleation] during interphase" - def text will need editing too, but that should be straightforward. I may have time to look at it properly tomorrow.
and would this be OK for the new terms?
abnormal cytoplasmic microtubule nucleation during mitotic interphase
-- abnormal microtubule nucleation from iMTOC [is "during mitotic interphase" needed in the name for this one?]
-- abnormal astral microtubule nucleation during mitotic interphase
and
abnormal astral microtubule nucleation
-- abnormal astral microtubule nucleation during mitotic interphase
-- abnormal astral microtubule nucleation during mitosis
Don't worry if it needs to wait until you get back...
Above looks perfect. No need for during interphase from iMTOC .....
Listing ALL the precise terms I need for annotation of this paper
rename
FYPO:0004619 cytoplasmic microtubules nucleated from equatorial microtubule organizing center absent from cell
to
abolished microtubule nucleation from eMTOC?
NTR: abolished astral microtubule nucleation during mitosis
NTR: normal cytoplasmic microtubule nucleation from the SPB during interphase
NTR abolished microtubule nucleation from the iMTOC (these are interphase cytoplasmic but this should be implicit from therm)
NTR: normal protein localzation to mitotic SPB
NTR: abolished protein localization to eMTOC
NTR: abolished protein localization to iMTOC
NTR: decreased mitotic SPB oscillation
NTR decreased cytoplasmic microtubule nucleation during interphase (not sure which MTOC)
NTR: abolished cytoplasmic microtubule nucleation
NTR: abolished cytoplasmic microtubule nucleation from the SPB during interphase
I think that's it. Phew!
new terms
abnormal microtubule nucleation FYPO:0005692
abolished cytoplasmic microtubule nucleation FYPO:0005693
decreased cytoplasmic microtubule nucleation during mitotic interphase FYPO:0005694
abolished astral microtubule nucleation during mitosis FYPO:0005695
abolished cytoplasmic microtubule nucleation from interphase microtubule organizing center FYPO:0005696
abolished cytoplasmic microtubule nucleation from spindle pole body during mitotic interphase FYPO:0005697
abolished microtubule nucleation from equatorial microtubule organizing center FYPO:0005698
normal cytoplasmic microtubule nucleation from spindle pole body during mitotic interphase FYPO:0005699
abolished protein localization to equatorial microtubule organizing center FYPO:0005700
abolished protein localization to interphase microtubule organizing center FYPO:0005701
abnormal protein localization to microtubule organizing center FYPO:0005702
renamed existing terms from using "interphase microtubule" to "microtubule [x] during mitotic interphase"; adjusted defs accordingly
edit file: 11877ed76fce9ae9583da5947f28b4c5e4ee6c6c
release: db5d4e543f7f6dbe443a927feaedadd5e3158917
| gharchive/issue | 2016-09-27T08:29:08 | 2025-04-01T04:35:32.119487 | {
"authors": [
"ValWood",
"mah11"
],
"repo": "pombase/fypo",
"url": "https://github.com/pombase/fypo/issues/2833",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1330756024 | This is getting slow
The way I'm using Prejournal for the Ponder Source Books in practice, a localhost reset which imports 1300 movements from about 10 .pj files is already taking 2.5 minutes to execute.
Maybe this is due to Postgres connections not being pooled in php -S localhost:8080 src/server.php, or maybe it's something more fundamental. In any case, this is OK if resets are rare, but if they're part of the standard workflow (make 1 small change in one .pj file, then run the reset script to see the effect) then it is already becoming unusable. :(
I'll try to fix this using TigerBeetle!
| gharchive/issue | 2022-08-06T13:43:22 | 2025-04-01T04:35:32.133514 | {
"authors": [
"michielbdejong"
],
"repo": "pondersource/prejournal",
"url": "https://github.com/pondersource/prejournal/issues/125",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
46808215 | not able to setup bs grid
Uncaught TypeError: Cannot read property 'error_message' of undefined jquery.bs_grid.js:1060
I am also getting this error.
JS
$("#demo_grid1").bs_grid({
ajaxFetchDataURL: '/Home/GetLabels',
row_primary_key: "LabelTypeID",
columns: [
{ field: "LabelTypeID", header: "Id", visible: "no" },
{ field: "Description", header: "Description" },
{ field: "SizeDescription", header: "Size" },
{ field: "ImageURL", header: "Image" , visible: "no" },
{ field: "IsDeleted", header: "Deleted" , visible: "no" },
{ field: "LabelTemplates", header: "Templates" , visible: "no" },
{ field: "ChangeDate", header: "Date updated" }
],
sorting: [
{ sortingName: "Id", field: "LabelTypeID", order: "none" },
{ sortingName: "Description", field: "Description", order: "ascending" }
],
filterOptions: {
filters: []
}
});
Json
{"total_rows":10,"page_data":[{"LabelTemplates":[],"LabelTypeID":1,"Description":"2" x 2" 7-Color DOW with Yellow-Stripe","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 2"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":2,"Description":"2" x 2" Yellow-Stripe Only","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 2"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":3,"Description":"2" x 1" 7-Color DOW with Yellow-Stripe","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":4,"Description":"2" x 1" Yellow-Stripe Only","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":5,"Description":"1" x 1" Yellow-Stripe Only","ChangeDate":"/Date(1421107200000)/","SizeDescription":"1" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":6,"Description":"1" x 1" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"1" x 1"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":7,"Description":"2" x 3" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 3"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":8,"Description":"2" x 4" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 4"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":9,"Description":"2" x 6" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 6"","ImageURL":"","IsDeleted":false},{"LabelTemplates":[],"LabelTypeID":10,"Description":"2" x 8" Blank White","ChangeDate":"/Date(1421107200000)/","SizeDescription":"2" x 8"","ImageURL":"","IsDeleted":false}]}
I am also getting this error.
why?
Me too. No detail on the error. My JSON returned is valid.
I am also having the same problem.. What can i do. Please give samples for Asp.net MVC
I modify " if(filter_error["error_message"] != null) " to " if(false) ", then it is working now. But it is not a good solution, if anyone who has the right method, please share to us. Thx
Hi yjd074 can you plz give the sample code.. i'll look into that
| gharchive/issue | 2014-10-25T10:21:15 | 2025-04-01T04:35:32.144097 | {
"authors": [
"HarveyEV",
"ManimaranUVI",
"jcphlux",
"juaby",
"valaydesai3",
"yjd074"
],
"repo": "pontikis/bs_grid",
"url": "https://github.com/pontikis/bs_grid/issues/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
460695536 | Fix return type checking to allow aliasing for non-ephemeral return types.
This PR fixes return type checking to allow aliasing for non-ephemeral return types.
I discovered this type system issue after a discussion with @rkallos in which we discussed how to fix Map.insert (and others) to not have an unreachable error case.
Looking into it further, I discovered that the type system had a simple bug that wasn't obvious due to the inconsistency in how Pony handles return type checking (the fact that using the ephemeral modifier is required to describe a unique return type, whereas on all other type specifications, the ephemeral modifier is useless).
This PR doesn't address that inconsistency, because it would be a major breaking change for the language, but this does fix the small bug that went unnoticed because of it.
The bug can be reproduced with the following code, in which get_1 and get_2 are valid and functionally identical, but the latter fails to compile:
class iso Inner
new iso create() => None
class Container[A: Inner #any]
var inner: A
new create(inner': A) => inner = consume inner'
fun get_1(): this->A => inner // works
fun get_2(): this->A => let tmp = inner; consume tmp // fails
actor Main
new create(env: Env) =>
let o = Container[Inner iso](Inner)
let i_1 : Inner tag = o.get_1()
let i_2 : Inner tag = o.get_2()
Upon investigation, it turned out to be a simple boolean logic bug, in which not (A or B) was expressed as !a || !b instead of the correct !a && !b. The code comment several lines above the change expresses the intended meaning, so I know that this fixed logic reflects the code author's original intent.
Wow. Heck of a find.
For posterity, I'll note that @SeanTAllen asked me to elaborate on the "inconsistency" in Zulip. You can find that conversation here: https://ponylang.zulipchat.com/#narrow/stream/189952-compiler-discussion/topic/return.20type.20ephemeral.20notation.20inconsistency
The CI failure appears to be unrelated?
@jemc yeah, i think so, it's been flakey lately.
Okay, was going to kick off another CI build, but that'll take forever to finish - I'm just going to merge.
But first, a changelog entry was added.
Looks like the job was straight up cancelled. Eventually cancelled jobs will get rerun. Or should. But I've been seeing them not get rerun lately. Such is life with free stuff I guess.
| gharchive/pull-request | 2019-06-25T23:49:17 | 2025-04-01T04:35:32.149391 | {
"authors": [
"SeanTAllen",
"jemc"
],
"repo": "ponylang/ponyc",
"url": "https://github.com/ponylang/ponyc/pull/3201",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1930280243 | Build UI for Page 2 🚀
Task :
Follow the UI given in Canva below and develop the UI for page 2
Use Futura Font Set for the UI
GET STARTED button should hit the route /player/
@pooranjoyb hi I want to work on this issue
@pooranjoyb hi I want to work on this issue
Go Ahead @CS50X-RGB 🔥
Ohhk @pooranjoyb sure can u give some idea about the colors like pink and peach is it? 😅
okay sure @pooranjoyb
@pooranjoyb how to get the images used there?
@pooranjoyb how to get the images used there? can you provide me edit access too it would be helpful for me
@CS50X-RGB
Opps! Sorry that I missed that out,
Here are the two gradients : #FF86C8 & #FBA277
And about the img just add a sample one, I'll co-author commits on your PR and add the original once you open it 👍
ohh thanks @pooranjoyb
| gharchive/issue | 2023-10-06T14:21:03 | 2025-04-01T04:35:32.157010 | {
"authors": [
"CS50X-RGB",
"pooranjoyb"
],
"repo": "pooranjoyb/BeatBridge",
"url": "https://github.com/pooranjoyb/BeatBridge/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1817256080 | 🛑 PrivilegedBA is down
In 50c9b48, PrivilegedBA (https://privilegedba.com.ar) was down:
HTTP code: 0
Response time: 0 ms
Resolved: PrivilegedBA is back up in 70e4134.
| gharchive/issue | 2023-07-23T18:38:36 | 2025-04-01T04:35:32.235259 | {
"authors": [
"porrale"
],
"repo": "porrale/status",
"url": "https://github.com/porrale/status/issues/262",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
993420066 | [seq], opening a file while playing should restart
when you open a file when reading an old file, it plays from the same spot as the old file was at.
It should restart
It should restart
things don't really work well in MAX like that, but screw it, I'm doing this anyway, and I'm not documenting it :)
| gharchive/issue | 2021-09-10T16:58:08 | 2025-04-01T04:35:32.236455 | {
"authors": [
"porres"
],
"repo": "porres/pd-cyclone",
"url": "https://github.com/porres/pd-cyclone/issues/559",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1373736466 | Build image from uploaded file: button not enabled
Bug description
On the build image page the "Build the image button" button stays disabled after entering the image name, selecting Upload and uploading a file.
Expected behavior
The button should be enabled so the user can initiate the image build.
Portainer Logs
Provide the logs of your Portainer container or Service.
You can see how here
Steps to reproduce the issue:
Go to Images > Build image
Enter image name
Click on "Upload"
Click on "Select file", select a docker file
Technical details:
Portainer version: 2.15.0
Docker version (managed by Portainer): 20.10.17
Kubernetes version (managed by Portainer): n/a
Platform (windows/linux): Linux
Command used to start Portainer (docker run -p 9443:9443 portainer/portainer):
Browser: Microsoft Edge, Chrome
Use Case (delete as appropriate): Using Portainer in a Commercial setup.
Have you reviewed our technical documentation and knowledge base? Yes
Additional context
This problem only surfaced after upgrading Portainer from v2.13.2 to v2.15.0. No problems building images using the Web editor or URL option.
@duncanw
I believe this issue is related to 7624 and is resolved in 2.15.1.
Thanks!
Thanks! Yes sorry looks like I created a duplicate issue.
| gharchive/issue | 2022-09-15T00:26:55 | 2025-04-01T04:35:32.243187 | {
"authors": [
"duncanw",
"tamarahenson"
],
"repo": "portainer/portainer",
"url": "https://github.com/portainer/portainer/issues/7666",
"license": "Zlib",
"license_type": "permissive",
"license_source": "github-api"
} |
71534166 | feature request : Standalone Pod and white list
Hello,
I need :
A way to configure a standalone pod which can't communicate with other pods.
A way to create a white list of pods the pod is authorized to talk.
Use case : I would like to install a pod for a school / a list of school.
Wrong issue tracker? This isn't Diaspora*
sorry, too much tabs opened...
| gharchive/issue | 2015-04-28T09:25:53 | 2025-04-01T04:35:32.253864 | {
"authors": [
"Proxiweb",
"posativ"
],
"repo": "posativ/isso",
"url": "https://github.com/posativ/isso/issues/188",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
282679145 | 403 Forbidden Error when posting new comments
I am testing isso on a live server with a domain. After posting serval comments successfully, isso sends 403 403 FORBIDDEN status code. Here is how it looks:
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
<title>403 Forbidden</title>
<h1>Forbidden</h1>
<p>3 direct responses to /post/post-title/</p>
What's the problem ?
By default, isso allows only 3 direct comments to a thread. This is can be configured by setting direct-reply variable in the isso configuration.
| gharchive/issue | 2017-12-17T08:43:49 | 2025-04-01T04:35:32.255312 | {
"authors": [
"noisytoken"
],
"repo": "posativ/isso",
"url": "https://github.com/posativ/isso/issues/364",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
528418847 | minor grammar improvements to quickstart.rst
Very minor changes to improve grammar and fix two misspelled word.
Closing, this is being tracked in https://github.com/posativ/isso/pull/590
| gharchive/pull-request | 2019-11-26T00:42:07 | 2025-04-01T04:35:32.256440 | {
"authors": [
"jasdeepgill",
"jelmer"
],
"repo": "posativ/isso",
"url": "https://github.com/posativ/isso/pull/592",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
688318888 | smtp only has 2 instances (should have 1) * smtp.2 is not running (Failed) postal
I have recently installed postal mail server on my VPS. After that when I am trying to connect to my sever using smtp credentials i am not able to do so. I am getting error. After doing 'postal restart' I got the result given below. I checked my netstat -lnp and the result is also given below. please help me fix the issue. need help @willpower232
root@postal:~# postal restart
Stopped smtp.2
Restarted web.1
Restarted worker.4 -> worker.5
Restarted cron.4 -> cron.5
Restarted smtp.3
Restarted requeuer.4 -> requeuer.5
root@postal:~# netstat -lnp
Active Internet connections (only servers)
Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name
tcp 0 0 0.0.0.0:80 0.0.0.0:* LISTEN 381/nginx: master p
tcp 0 0 0.0.0.0:4369 0.0.0.0:* LISTEN 441/epmd
tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN 360/sshd
tcp 0 0 127.0.0.1:25 0.0.0.0:* LISTEN 1024/exim4
tcp 0 0 0.0.0.0:443 0.0.0.0:* LISTEN 381/nginx: master p
tcp 0 0 127.0.0.1:5000 0.0.0.0:* LISTEN 1282/[postal] web.1
tcp 0 0 0.0.0.0:25672 0.0.0.0:* LISTEN 296/beam.smp
tcp 0 0 127.0.0.1:3306 0.0.0.0:* LISTEN 575/mysqld
tcp6 0 0 :::80 :::* LISTEN 381/nginx: master p
tcp6 0 0 :::4369 :::* LISTEN 441/epmd
tcp6 0 0 :::22 :::* LISTEN 360/sshd
tcp6 0 0 ::1:25 :::* LISTEN 1024/exim4
tcp6 0 0 :::443 :::* LISTEN 381/nginx: master p
tcp6 0 0 :::5672 :::* LISTEN 296/beam.smp
Active UNIX domain sockets (only servers)
Proto RefCnt Flags Type State I-Node PID/Program name Path
unix 2 [ ACC ] STREAM LISTENING 692941580 1/init /run/systemd/journal/stdout
unix 2 [ ACC ] STREAM LISTENING 692944182 276/saslauthd /var/run/saslauthd/mux
unix 2 [ ACC ] STREAM LISTENING 693546224 1276/[procodile] Po /tmp/postal/pids/procodile.sock
unix 2 [ ACC ] SEQPACKET LISTENING 692947316 1/init /run/udev/control
unix 2 [ ACC ] STREAM LISTENING 692929142 1/init /var/run/dbus/system_bus_socket
unix 2 [ ACC ] STREAM LISTENING 692949172 1/init /run/systemd/private
unix 2 [ ACC ] STREAM LISTENING 692892378 575/mysqld /var/run/mysqld/mysqld.sock
You are running another mail server software "Exim" using same SMTP port as Postal.
You need to stop the Exim service and probably disable it, then issue a "postal restart".
Thank you for your reply @yomiit
@yomiit when i did #service exim stop it gives me the result below
root@postal:~# service exim stop
Failed to stop exim.service: Unit exim.service not loaded.
I have installed postal using "curl https://raw.githubusercontent.com/atech/postal/master/script/install/ubuntu1604.sh | sh
" this command. I am running ununtu 18.04. I have setup firewall using ufw. ufw status is given below
root@postal:~# ufw status
Status: active
To Action From
25/tcp ALLOW Anywhere
465/tcp ALLOW Anywhere
587/tcp ALLOW Anywhere
22/tcp ALLOW Anywhere
80/tcp ALLOW Anywhere
443/tcp ALLOW Anywhere
25/tcp (v6) ALLOW Anywhere (v6)
465/tcp (v6) ALLOW Anywhere (v6)
587/tcp (v6) ALLOW Anywhere (v6)
22/tcp (v6) ALLOW Anywhere (v6)
80/tcp (v6) ALLOW Anywhere (v6)
443/tcp (v6) ALLOW Anywhere (v6)
The service should be exim4.
systemctl stop exim4
systemctl disable exim4
hello. thanks for the help. Now postal is listening to port 25, and only listening to port 25. I can not connect to postal using 465 & 587, need help with that. plz @willpower232 @yomiit
Connecting on other ports shouldn't be required but if you want it, you will have to use something like iptables to forward ports internally.
Connecting on other ports shouldn't be required but if you want it, you will have to use something like iptables to forward ports internally.
@willpower232 plz help me to do it, as this is my first time doing this.
| gharchive/issue | 2020-08-28T19:46:46 | 2025-04-01T04:35:32.279474 | {
"authors": [
"samg8520",
"willpower232",
"yomiit"
],
"repo": "postalhq/postal",
"url": "https://github.com/postalhq/postal/issues/1193",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1999476558 | Corrected a typo in Readme file
Please review this typo correction.
Sure let me do that! thank you 👍
Closes #59
| gharchive/pull-request | 2023-11-17T16:16:32 | 2025-04-01T04:35:32.296776 | {
"authors": [
"LKoech",
"loopDelicious"
],
"repo": "postmanlabs/pmquickstarts",
"url": "https://github.com/postmanlabs/pmquickstarts/pull/61",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
177254822 | Installation: 301 on cookbooks installation
Hi! I am trying to set up a development server of my own. I am now on step 5 and I get a following error:
[nina@leafblade tomatoes]$ librarian-chef install Could not cache bluepill/1.0.6 <http://community.opscode.com/api/v1> from http://community.opscode.com/api/v1/cookbooks/bluepill because 301 Moved Permanently!
Can I do something to bypass this error?
@matteodepalo is the one who setup that process in the first place, let's see if he can help you solving this issue.
@Liarra @potomak the Cheffile was updated 4 years ago so some of the libraries might have disappeared or they might have been moved to other places. To fix this we would need to look at the git urls in the Cheffile and update all of them to the latest versions. Also check if there are greater versions for site 'http://community.opscode.com/api/v1'. Try bumping that version to the latest one and see what happens.
@matteodepalo can you take a quick look at it and remove cheffile + doc if it's not usable anymore or update it please?
@potomak it would take a while to update the Cheffile and the recipes to the latest version. I would personally remove the files as I don't have the time to update them unfortunately.
Ok, I'll add a task to remove them. Thanks.
| gharchive/issue | 2016-09-15T18:37:35 | 2025-04-01T04:35:32.366246 | {
"authors": [
"Liarra",
"matteodepalo",
"potomak"
],
"repo": "potomak/tomatoes",
"url": "https://github.com/potomak/tomatoes/issues/143",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2521891478 | Please consider fixing issues detected by repository checker
Notification from ioBroker Check and Service Bot
Dear adapter developer,
I'm the ioBroker Check and Service Bot. I'm an automated tool processing routine tasks for the ioBroker infrastructure. I have recently checked the repository for your adapter warp for common errors and appropiate suggestions to keep this adapter up to date.
This check is based the current head revisions (master / main branch) of the adapter repository
Please see the result of the check below.
ioBroker.warp
-
ERRORS:
[ ] :heavy_exclamation_mark: [E026] "{'engines': {'node'>='16'}}" is required at package.json, "{'engines':{'node'>='18'}}" is recommended
[ ] :heavy_exclamation_mark: [E033] @iobroker/adapter-core 2.6.0 specified. 3.1.4 is required as minimum, 3.1.6 is recommended. Please update dependencies at package.json
[ ] :heavy_exclamation_mark: [E036] @iobroker/testing 3.0.2 specified. 4.1.3 is required as minimum, 4.1.3 is recommended. Please update devDependencies at package.json
[ ] :heavy_exclamation_mark: [E157] js-controller 2.0.0 listed as dependency but 4.0.24 is required as minimum, 5.0.19 is recommended. Please update dependency at io-package.json.
WARNINGS:
[ ] :eyes: [W040] "keywords" within package.json should contain "ioBroker"
[ ] :eyes: [W127] Missing suggested translation into uk of "common.titleLang" in io-package.json.
[ ] :eyes: [W132] Many "common.news" found in io-package.json. Repository builder will truncate at 7 news. Please remove old news.
[ ] :eyes: [W134] Missing suggested translation into uk of "common.desc" in io-package.json.
[ ] :eyes: [W135] "common.tier" is required in io-package.json. Please check https://github.com/ioBroker/ioBroker.docs/blob/master/docs/en/dev/objectsschema.md#adapter.
[ ] :eyes: [W154] Missing suggested translation into uk of some "common.news" in io-package.json.
[ ] :eyes: [W181] "common.license" in io-package.json is deprecated. Please define object "common.licenseInformation"
[ ] :eyes: [W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Use "main" at package.json instead.
[ ] :eyes: [W184] "common.materialize" is deprecated for admin >= 5 at io-package.json. Please use property "adminUI".
[ ] :eyes: [W184] "common.title" is deprecated and replaced by "common.titleLang". Please remove from io-package.json.
SUGGESTIONS:
[ ] :pushpin: [S522] Please consider migrating to admin 5 UI (jsonConfig).
Please review issues reported and consider fixing them as soon as appropiate.
Errors reported by repository checker should be fixed as soon as possible. Some of them require a new release to be considered as fixed. Please note that errors reported by checker might be considered as blocking point for future updates at stable repository.
Warnings reported by repository checker should be reviewed. While some warnings can be ignored due to good reasons or a dedicated decision of the developer, most warnings should be fixed as soon as appropiate.
Suggestions reported by repository checker should be reviewed. Suggestions can be ignored due to a decision of the developer but they are reported as a hint to use a configuration which might get required in future or at least is used be most adapters. Suggestions are always optional to follow.
You may start a new check or force the creation of a new issue at any time by adding the following comment to this issue:
@iobroker-bot recheck
@iobroker-bot recreate
Please note that I (and the server at GitHub) have always plenty of work to do. So it may last up to 30 minutes until you see a reaction. I will drop a comment here as soon as I start processing.
Feel free to contact me (@iobroker-bot) if you have any questions or feel that an issue is incorrectly flagged.
And THANKS A LOT for maintaining this adapter from me and all users.
Let's work together for the best user experience.
your
ioBroker Check and Service Bot
@mcm1957 for evidence
Last update at Fri, 13 Sep 2024 14:28:45 GMT based on commit 0d1addd26d6ba16510076e1d09edb8258aa6df0f
ioBroker.repochecker 3.0.5
This issue has been updated by ioBroker Check and Service Bot
The following issues have been fixed
[W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Use "main" at package.json instead.
:thumbsup:Thanks for fixing the issues.
The following issues are new and have been added
[W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Executable is defined by entry "main" at package.json.
| gharchive/issue | 2024-09-12T09:43:29 | 2025-04-01T04:35:32.385451 | {
"authors": [
"ioBroker-Bot"
],
"repo": "pottio/ioBroker.warp",
"url": "https://github.com/pottio/ioBroker.warp/issues/134",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2063163496 | 🛑 PowBot SDN is down
In ca5117d, PowBot SDN (https://api.powbot.org/products) was down:
HTTP code: 502
Response time: 658 ms
Resolved: PowBot SDN is back up in 3d4e750 after 9 minutes.
| gharchive/issue | 2024-01-03T02:31:02 | 2025-04-01T04:35:32.406599 | {
"authors": [
"constt"
],
"repo": "powbot/status",
"url": "https://github.com/powbot/status/issues/279",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
845088230 | Update kubernetes/master/golang/master build
This is an automated PR via build-bot
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by: ltccci
To complete the pull request process, please assign after the PR has been reviewed.
You can assign the PR to them by writing /assign in a comment when ready.
The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
| gharchive/pull-request | 2021-03-30T19:08:36 | 2025-04-01T04:35:32.443153 | {
"authors": [
"ltccci"
],
"repo": "ppc64le-cloud/builds",
"url": "https://github.com/ppc64le-cloud/builds/pull/1857",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2279737843 | How to find the number of records in a relation?
I'm trying to use pagination to display product cards.
However I have a connection Typeorm as: Products <-> accounting
ProductsEntity:
@OneToMany(() => AccountingEntity, (accounting) => accounting.product, {})
accounting: AccountingEntity[];
AccountingEntity:
@ManyToOne(() => ProductEntity, (product) => product.accounting, {
onDelete: 'CASCADE',
})
@JoinColumn({ name: 'product_id' })
product: ProductEntity;
I display my product cards, but I can’t do it to show only those records in which there is NO accounting! or vice versa there is only ACCOUNTING
How can I do this correctly?
My code:
const result: any = await paginate(query, this.productRepository, {
sortableColumns: ['uuid', 'name', 'inventory', 'series', 'count', 'accounting', 'author.last_name',],
relations: ['accounting', 'author', 'user'],
nullSort: 'last',
defaultSortBy: [['uuid', 'DESC']],
searchableColumns: ['uuid', 'name', 'count', 'author.last_name', 'author.first_name', 'author.surname', 'user.last_name', 'user.first_name', 'user.surname', 'series', 'inventory',],
filterableColumns: {
count: [FilterOperator.EQ, FilterOperator.BTW],
accounting: true,
author: true,
user: true
},
});
bump
I'm trying to use pagination to display product cards.
However I have a connection Typeorm as: Products <-> accounting
ProductsEntity:
@OneToMany(() => AccountingEntity, (accounting) => accounting.product, {})
accounting: AccountingEntity[];
AccountingEntity:
@ManyToOne(() => ProductEntity, (product) => product.accounting, {
onDelete: 'CASCADE',
})
@JoinColumn({ name: 'product_id' })
product: ProductEntity;
I display my product cards, but I can’t do it to show only those records in which there is NO accounting! or vice versa there is only ACCOUNTING
How can I do this correctly?
My code:
const result: any = await paginate(query, this.productRepository, {
sortableColumns: ['uuid', 'name', 'inventory', 'series', 'count', 'accounting', 'author.last_name',],
relations: ['accounting', 'author', 'user'],
nullSort: 'last',
defaultSortBy: [['uuid', 'DESC']],
searchableColumns: ['uuid', 'name', 'count', 'author.last_name', 'author.first_name', 'author.surname', 'user.last_name', 'user.first_name', 'user.surname', 'series', 'inventory',],
filterableColumns: {
count: [FilterOperator.EQ, FilterOperator.BTW],
accounting: true,
author: true,
user: true
},
});
You can achieve this using where in pagination config
| gharchive/issue | 2024-05-05T21:20:57 | 2025-04-01T04:35:32.447493 | {
"authors": [
"MrSharpp",
"kenvals",
"ppetzold"
],
"repo": "ppetzold/nestjs-paginate",
"url": "https://github.com/ppetzold/nestjs-paginate/issues/916",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1338007729 | 是否有人曾提过类似的问题?
否(No)
你觉得APP有什么不足之处?
no
你觉得该怎么去完善会比较好?【非必答】
No response
等下一个版本发布 或 下载每周构建版本
| gharchive/issue | 2022-08-13T16:53:04 | 2025-04-01T04:35:32.460972 | {
"authors": [
"daidaojianke",
"pppscn"
],
"repo": "pppscn/SmsForwarder",
"url": "https://github.com/pppscn/SmsForwarder/issues/204",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
651552045 | Fix setting GridContainer content by index
resolves #3657
The problem is that an array does not notify about changing its elements.
This PR adds a GridContainerContent class which wraps a jagged array and provides notifications for its elements changes.
Awesome, but this is a pretty large breaking change. Is it possible to have an implicit operator conversion from Drawable[][] to GridContainerContent?
Is it possible to have an implicit operator conversion from Drawable[][] to GridContainerContent?
AFAIK it is not possible to convert an array to a custom object. However, I can write an extension method in order to make code a little cleaner.
Vice-versa would also be good to have (GridContainerContent -> Drawable[][], even if it requires allocs).
We can expose the original array via method or property. See https://github.com/UselessToucan/osu-framework/commit/f5b3f9205a8beaa847e392563fc4342703eb2e3d.
AFAIK it is not possible to convert an array to a custom object. However, I can write an extension method in order to make code a little cleaner.
I think the suggestion was to add operators like
public static implicit operator Drawable[][](GridContainerContent content);
public static implicit operator GridContainerContent(Drawable[][] drawables);
and it seems to at least not throw syntax errors and build? This way the changes to Content should no longer be breaking, or am I missing something?
Unfortunately this is still a breaking change, albeit a slightly less breaking one - game-side there's this usage that needs adjusting for.
| gharchive/pull-request | 2020-07-06T13:44:53 | 2025-04-01T04:35:32.467703 | {
"authors": [
"UselessToucan",
"bdach",
"smoogipoo"
],
"repo": "ppy/osu-framework",
"url": "https://github.com/ppy/osu-framework/pull/3692",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1345879912 | Fix potential startup crash if resolution/display pair from configuration results in no valid resolution
Closes https://github.com/ppy/osu/issues/19877#issuecomment-1221438234.
I wasn't able to reproduce what the user was seeing even when trying my best to break it, but have tested the flow works by artifically adding an exception:
diff --git a/osu.Framework/Platform/SDL2DesktopWindow.cs b/osu.Framework/Platform/SDL2DesktopWindow.cs
index 8ca11367f..d68d8368f 100644
--- a/osu.Framework/Platform/SDL2DesktopWindow.cs
+++ b/osu.Framework/Platform/SDL2DesktopWindow.cs
@@ -1233,6 +1233,9 @@ private void updateWindowStateAndSize()
{
var closestMode = getClosestDisplayMode(sizeFullscreen.Value, currentDisplayMode.Value.RefreshRate, currentDisplay.Index);
+ if (!sizeFullscreen.IsDefault)
+ throw new InvalidOperationException("test");
+
Size = new Size(closestMode.w, closestMode.h);
SDL.SDL_SetWindowDisplayMode(SDLWindowHandle, ref closestMode);
#5649
| gharchive/pull-request | 2022-08-22T06:21:53 | 2025-04-01T04:35:32.469538 | {
"authors": [
"peppy"
],
"repo": "ppy/osu-framework",
"url": "https://github.com/ppy/osu-framework/pull/5371",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2588017357 | tests script does not work outside of nix environment
When I try to run tests func locally (in a python environment with click installed), I get:
(venv) > mlkem-c-aarch64 % tests func
INFO > Functional Test
INFO > make CROSS_PREFIX= mlkem AUTO=1 OPT=1
make: Circular test/build/lib/libfips202.a <- test/build/lib/libfips202.a dependency dropped.
make: Circular test/build/lib/librng.a <- test/build/lib/libfips202.a dependency dropped.
make: Circular test/build/lib/librng.a <- test/build/lib/librng.a dependency dropped.
make: Circular test/build/randombytes/randombytes.c.o <- test/build/lib/libfips202.a dependency dropped.
make: Circular test/build/randombytes/randombytes.c.o <- test/build/lib/librng.a dependency dropped.
ld: warning: search path 'test/build/lib' not found
ld: library 'fips202' not found
which hints at some issue with the Makefile.
The same problem arises with tests kat and tests nistkat.
After https://github.com/pq-code-package/mlkem-c-aarch64/pull/241, the following problems emerge:
mlkem-c-aarch64 % tests kat
INFO > Kat Test
INFO > make CROSS_PREFIX= kat OPT=1 AUTO=1
INFO > ./test/build/mlkem512/bin/gen_KAT512
ERROR > ML-KEM-512 failed, expecting , but getting 323d0e9aefe34819f10cdce1f0d9e5c8a55193cebe984fb1718e779ebfbc0da8
INFO > ./test/build/mlkem768/bin/gen_KAT768
ERROR > ML-KEM-768 failed, expecting , but getting 99b497dcddfe418f44d30c7376fda09ae7cca2e9141143032d842508b4a1f438
INFO > ./test/build/mlkem1024/bin/gen_KAT1024
ERROR > ML-KEM-1024 failed, expecting , but getting 104058bab1fef70aa10606831faabef7053d44b1adac6b34d35e505c3085db78
and
> ./test/build/mlkem512/bin/test_kyber512
ERROR > ML-KEM-512 failed, expecting CRYPTO_SECRETKEYBYTES:
CRYPTO_PUBLICKEYBYTES:
CRYPTO_CIPHERTEXTBYTES:
, but getting CRYPTO_SECRETKEYBYTES: 1632
CRYPTO_PUBLICKEYBYTES: 800
CRYPTO_CIPHERTEXTBYTES: 768
INFO > ./test/build/mlkem768/bin/test_kyber768
ERROR > ML-KEM-768 failed, expecting CRYPTO_SECRETKEYBYTES:
CRYPTO_PUBLICKEYBYTES:
CRYPTO_CIPHERTEXTBYTES:
, but getting CRYPTO_SECRETKEYBYTES: 2400
CRYPTO_PUBLICKEYBYTES: 1184
CRYPTO_CIPHERTEXTBYTES: 1088
INFO > ./test/build/mlkem1024/bin/test_kyber1024
ERROR > ML-KEM-1024 failed, expecting CRYPTO_SECRETKEYBYTES:
CRYPTO_PUBLICKEYBYTES:
CRYPTO_CIPHERTEXTBYTES:
, but getting CRYPTO_SECRETKEYBYTES: 3168
CRYPTO_PUBLICKEYBYTES: 1568
CRYPTO_CIPHERTEXTBYTES: 1568
and
% tests nistkat
INFO > Nistkat Test
INFO > make CROSS_PREFIX= nistkat AUTO=1 OPT=1
INFO > ./test/build/mlkem512/bin/gen_NISTKAT512
ERROR > ML-KEM-512 failed, expecting , but getting a30184edee53b3b009356e1e31d7f9e93ce82550e3c622d7192e387b0cc84f2e
INFO > ./test/build/mlkem768/bin/gen_NISTKAT768
ERROR > ML-KEM-768 failed, expecting , but getting 729367b590637f4a93c68d5e4a4d2e2b4454842a52c9eec503e3a0d24cb66471
INFO > ./test/build/mlkem1024/bin/gen_NISTKAT1024
ERROR > ML-KEM-1024 failed, expecting , but getting 3fba7327d0320cb6134badf2a1bcb963a5b3c0026c7dece8f00d6a6155e47b33
Fixed by https://github.com/pq-code-package/mlkem-c-aarch64/pull/243
| gharchive/issue | 2024-10-15T08:23:44 | 2025-04-01T04:35:32.533841 | {
"authors": [
"hanno-becker"
],
"repo": "pq-code-package/mlkem-c-aarch64",
"url": "https://github.com/pq-code-package/mlkem-c-aarch64/issues/240",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1562993866 | fix: improve types
While the files property can also be populated with new files, it is advised to add new files using only the addFile and addFiles methods.
When using React or other cases, such as a quickstart example:
const [files, setFiles] = useState<FilePondFile[]>([]);
<FilePond
files={files}
onupdatefiles={(files) => setFiles(files)}
/>
This will throw a type error because:
files?: Array<FilePondInitialFile | ActualFileObject | Blob | string>;
onupdatefiles?: ((files: FilePondFile[]) => void)
This is an issue we are dealing with too
| gharchive/pull-request | 2023-01-30T18:38:50 | 2025-04-01T04:35:32.537400 | {
"authors": [
"CarelessCourage",
"songhn233"
],
"repo": "pqina/filepond",
"url": "https://github.com/pqina/filepond/pull/887",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1119393251 | Feature - Reverse Transliteration
Is there any plan for including Hindi-Hinglish transliteration (the reverse of what the library has currently)?
Hey, as of now we don't have any plans for the same. To my knowledge, I also think that datasets for reverse transliteration are extremely scarce. I wasn't able to find anything relevant on the internet either after skimming briefly.
| gharchive/issue | 2022-01-31T12:56:13 | 2025-04-01T04:35:32.541296 | {
"authors": [
"hetarth18",
"praatibhsurana"
],
"repo": "praatibhsurana/Hinglish_Hindi_WSD",
"url": "https://github.com/praatibhsurana/Hinglish_Hindi_WSD/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
653986503 | Criar esboço técnico de uma sala de telepresença
O documento pode ter uma única página que liste os itens de uma sala de telepresença e a suas interações entre si. Além disso, o documento deve conter uma ilustração de uma sala de aula e onde os elementos estarão posicionados.
Os itens são os seguintes:
Central (Raspberry PI + câmera embutida + microfone condensador apontado para frente) fixado no teto da sala de aula, no centro, com a câmera apontada para o quadro/docente.
TV de 50 polegadas (fixada na parede do fundo da sala de aula, no meio dela, conecatada à central por um cabo HDMI de 10m).
Cabo de rede que liga a central até o ponto de rede mais próximo.
Eu estou pedindo pro setor de obras o layout das salas de aula (creio que sejam iguais nos 6 campi) para pensarmos já espacialmente de acordo com as dimensões das salas.
O esquema está disponível em alta resolução no drive em programa#57. Abaixo uma ilustração:
| gharchive/issue | 2020-07-09T11:28:27 | 2025-04-01T04:35:32.566468 | {
"authors": [
"Dovyski",
"phdmauricio"
],
"repo": "practice-uffs/programa",
"url": "https://github.com/practice-uffs/programa/issues/57",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1055518622 | Qual a estrutura do curso?
Derivada da macro tarefa AUR-1.3. Esta micro tarefa depende da conclusão das micro tarefas de 41 a 63.
Atentem para o fato que a pergunta é um norteador, o responsável pela tarefa deve pensar em variações pertinentes daquela pergunta. Limitar-se a colocar as informações da pergunta da tarefa. Outras perguntas (ou informações novas) serão feitas em outras fases do projeto.
Na macro tarefa AUR-1.3 deve-se elencar e catalogar (na planilha da Aura) quais são as informações relevantes do item do conjunto: Cursos da UFFS (graduação e pós-graduação).
ESSA ISSUE REQUER UMA RESPOSTA PARA CADA CURSO DE PÓS-GRADUAÇÃO DA UFFS
Link para a planilha onde devem ser alocadas as informações
AS RESPOSTAS DEVEM SER ALOCADAS NA PLANILHA DO LINK E NÃO EM PLANILHAS SUBSEQUENTES TENDO EM VISTA A ORGANIZAÇÃO DAS TAREFAS E A REVISÃO DAS MESMAS.
Criei as pastas dessa issue no Google Drive:
Pasta da issue
🔽 Entrada
🔼 Saída
Essa issue não é mais válida!
| gharchive/issue | 2021-11-16T23:07:03 | 2025-04-01T04:35:32.570748 | {
"authors": [
"Claudineia-VR",
"PracticeUFFSBot",
"morgana-mo"
],
"repo": "practice-uffs/programa",
"url": "https://github.com/practice-uffs/programa/issues/923",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
340145002 | Optout confirmation
This PR adds a confirmation message when the optout button is pressed asking the user to confirm to minimise mistakes.
I also fixed the problem where you need to refresh to get the optout button to hide after a successful optout and I made the extra info on the address look better.
I did everything in separate commits for easy reviewing.
+1
| gharchive/pull-request | 2018-07-11T08:40:10 | 2025-04-01T04:35:32.575001 | {
"authors": [
"DevChima",
"erikh360"
],
"repo": "praekelt/seed-control-interface",
"url": "https://github.com/praekelt/seed-control-interface/pull/71",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
167587111 | Only use decimal digits in MTN Nigeria XML over TCP transport session IDs.
Currently we use hexadecimal digits, but non-decimal digits are not allowed.
Anyone have any opinions on using:
session_id = "".join(random.choice("0123456789") for i in range(SESSION_ID_LENGTH))
as a method for generating a session id?
:+1: when travis is happy
I've confirmed that all the Travis errors are covered by #1049 and that the tests pass locally, so landing this.
| gharchive/pull-request | 2016-07-26T11:43:31 | 2025-04-01T04:35:32.576689 | {
"authors": [
"hodgestar",
"justinvdm"
],
"repo": "praekelt/vumi",
"url": "https://github.com/praekelt/vumi/pull/1050",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1794997832 | Adding a web development project: Word Counter [gssoc23]
Title and Issue number
Web development project: Word Counter [gssoc23]
Please add the tag of gssoc23
Close #3065
Checklist:
[X] I have mentioned the issue number in my Pull Request.
[X] I have commented my code, particularly in hard-to-understand areas
[X] I have created a helpful and easy to understand README.md
@MohitGupta121 @pranjay-poddar Please review this PR.
| gharchive/pull-request | 2023-07-08T15:56:08 | 2025-04-01T04:35:32.598206 | {
"authors": [
"thevirengarg"
],
"repo": "pranjay-poddar/Dev-Geeks",
"url": "https://github.com/pranjay-poddar/Dev-Geeks/pull/3263",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
265033453 | [issue-32] Table Connector
Change log description
add FlinkPravegaTableSource
add FlinkPravegaTableSink
add FlinkTableITCase
add JSON serialization schema
add jackson as a 'provided' (by Flink) dependency
Purpose of the change
Introduces Flink Table API support, including a table source and append-only table sink. Closes #32 .
How to verify it
Execute integration test FlinkTableITCase.
Note that the JsonRowDeserializationSchema and JsonRowSerializationSchema classes were copied from the Flink Kafka connector. I'll follow up with the original authors as to whether they could be moved to the flink-table library. Unsure as to whether any attribution is needed.
Providing a JsonSerializationSchema makes sense here. We may want to take the next step and also provide a Pravega io.pravega.client.stream.Serializer that works with Json. That way someone can write a standalone app to write the data and read it using the table connector.
Extending that, if we have a JsonSerializer for Pravega, perhaps we should use the PravegaDeserializationSchema that already exists and use the JsonSerializer in conjunction with that for the table connector.
I think this serializer may be too specialized to be replaced with a
generic Pravega serializer since it is coded for the Row class and not
based on Jackson databinding.
I agree completely that the Pravega client should provide a Json serializer
based on data binding. We could provide something in the samples. Will
file.
On Oct 12, 2017 11:17 AM, "Chris Dail" notifications@github.com wrote:
Providing a JsonSerializationSchema makes sense here. We may want to take
the next step and also provide a Pravega io.pravega.client.stream.
Serializer that works with Json. That way someone can write a standalone
app to write the data and read it using the table connector.
Extending that, if we have a JsonSerializer for Pravega, perhaps we
should use the PravegaDeserializationSchema that already exists and use
the JsonSerializer in conjunction with that for the table connector.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/pravega/flink-connectors/pull/61#issuecomment-336222358,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABsXno8w95j-xUBoqsC_4Ovx1tl_2Piiks5srlfVgaJpZM4P3Zds
.
@tzulitai I invited you as a collaborator, then I can request a review from you.
| gharchive/pull-request | 2017-10-12T18:01:23 | 2025-04-01T04:35:33.977955 | {
"authors": [
"EronWright",
"chrisdail"
],
"repo": "pravega/flink-connectors",
"url": "https://github.com/pravega/flink-connectors/pull/61",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
822027765 | Document demand chain objects
Documents https://github.com/prebid/Prebid.js/pull/6383 and https://github.com/prebid/Prebid.js/pull/6380
Will be part of 4.30
| gharchive/pull-request | 2021-03-04T11:11:16 | 2025-04-01T04:35:34.059533 | {
"authors": [
"bretg",
"patmmccann"
],
"repo": "prebid/prebid.github.io",
"url": "https://github.com/prebid/prebid.github.io/pull/2737",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
920688261 | Farm HNZ
Farm in NC it went down because gateway was knock over by something (presumably a person with a baseball) gateway needs to be replaced. gateway got full of water.
Baseball is a fun plot twist!
Anna Paulk Birnbaum
Agricultural Specialist
UGA-Department of Crop and Soil Sciences
1420 Experiment Station Rd
Watkinsville, GA 30677
C: 404-889-1688
From: elhenriq @.>
Sent: Monday, June 14, 2021 2:56 PM
To: precision-sustainable-ag/On-farm-Protocols @.>
Cc: Subscribed @.***>
Subject: [precision-sustainable-ag/On-farm-Protocols] Farm HNZ (#105)
[EXTERNAL SENDER - PROCEED CAUTIOUSLY]
Farm in NC it went down because gateway was knock over by something (presumably a person with a baseball) gateway needs to be replaced. gateway got full of water.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHubhttps://github.com/precision-sustainable-ag/On-farm-Protocols/issues/105, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APL6HJ2KXK2AHRIHMWG34L3TSZGFZANCNFSM46VZCVJQ.
There was an actual baseball next to it in the field
WOW someone REALLY doesn’t like the sensors
Anna Paulk Birnbaum
Agricultural Specialist
UGA-Department of Crop and Soil Sciences
1420 Experiment Station Rd
Watkinsville, GA 30677
C: 404-889-1688
From: elhenriq @.>
Sent: Monday, June 14, 2021 3:16 PM
To: precision-sustainable-ag/On-farm-Protocols @.>
Cc: Anna Paulk Birnbaum @.>; Comment @.>
Subject: Re: [precision-sustainable-ag/On-farm-Protocols] Farm HNZ (#105)
[EXTERNAL SENDER - PROCEED CAUTIOUSLY]
There was an actual baseball next to it in the field [image] https://user-images.githubusercontent.com/52674578/121946979-82fb7880-cd23-11eb-9b8c-ed02782eafc0.jpeg [image] https://user-images.githubusercontent.com/52674578/121946982-84c53c00-cd23-11eb-83e3-bf0404519b7d.jpeg
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHubhttps://github.com/precision-sustainable-ag/On-farm-Protocols/issues/105#issuecomment-860929646, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APL6HJYQOGZ5OLQOVKUJMKTTSZISXANCNFSM46VZCVJQ.
| gharchive/issue | 2021-06-14T18:56:13 | 2025-04-01T04:35:34.072017 | {
"authors": [
"annabirnbaum",
"elhenriq"
],
"repo": "precision-sustainable-ag/On-farm-Protocols",
"url": "https://github.com/precision-sustainable-ag/On-farm-Protocols/issues/105",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2185684636 | Want Lorax with newer version of TGI
Feature request
hello,our models are deploying with TGI(v1.4.3), and we alse want to use lorax. But I find that the tgi version lorax is based on is very different with TGI version v1.4.3。
We are trying to integrate lorax(v0.8) into TGI(v1.4.3)。Is there possible to upgrade TGI of lorax or contribute lorax to TGI?
Motivation
use new features of TGI together with lorax
Your contribution
We are trying to integrate lorax(v0.8) into TGI(v1.4.3), but both lorax and tgi are changing!
Hi @yangelaboy, thanks for trying out LoRAX. I'd love to incorporate more upstream work from TGI, but since they changed their license last year, we can no longer pull their code into our repo.
That said, we have implemented many of the same features recently (though in slightly different ways). Are there specific features you're using in TGI you want to see in LoRAX? If so, we can definitely prioritize getting those added.
One thing in TGI we're working to add very soon is speculative decoding. We think our implementation will be particularly interesting, as we'll be able to handle multiple speculation models at once. Let me know if there are other features you're interested in.
@tgaddair Thinks for detailed replies. We are using features such as speculative decoding(ngram&medusa), quantization, also we're interested in much optimizations of TGI. We also added functions in TGI like shared prefix prompt cache。
Finally, We want a framework which can support different adapter models and medusa models in same self-trained model with a shared prefix prompt cache.
I will pay attention to Lorax.
Hey @yangelaboy, thanks for this context! The good news is all of the things you listed are on our near-term roadmap.
Speculative decoding adapters per request - this is what I'm currently working on and hope to have out next week
Prefix caching - this is the next major item on the roadmap after speculative decoding, so hopefully a few weeks away at most
Quantization - we support a number of quantization options currently, but let me know if there are specifics ones we don't support that you would be interested in.
I'll definitely let you know when the speculative decoding is ready to test out!
Thanks @tgaddair , we are also waiting for the Speculative decoding 👍
The license is back to Apache-2.0
https://github.com/huggingface/text-generation-inference/commit/ff42d33e9944832a19171967d2edd6c292bdb2d6
@tgaddair
| gharchive/issue | 2024-03-14T07:59:57 | 2025-04-01T04:35:34.078417 | {
"authors": [
"abhibst",
"giyaseddin",
"tgaddair",
"yangelaboy"
],
"repo": "predibase/lorax",
"url": "https://github.com/predibase/lorax/issues/329",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2398105992 | Add pixi entry to tldr-pages
Problem description
Would be cool if pixi had an entry in tldr pages: https://github.com/tldr-pages/tldr/issues/13248
Sweet, this was already implemented!
| gharchive/issue | 2024-07-09T12:41:04 | 2025-04-01T04:35:34.082337 | {
"authors": [
"corneliusroemer"
],
"repo": "prefix-dev/pixi",
"url": "https://github.com/prefix-dev/pixi/issues/1595",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1954222675 | Make project.name and project.version optional
Problem description
Currently pixi can not install a project without these two values:
[project]
name = "myproject"
version = "0.1.0"
I suggest making them optional because in a project that does not only use pixi, the name and version may already be defined in another file like package.json. Cargo.toml, making these values in the file duplicate and pointless.
Making these optional would be in-line with package.json where there are also no required fields.
We're making the version optional :+1: The name is needed for the shell, and requires less house keeping to align with the project.
Good call on version. Name won't change that often, but it could potentially fall back to the directory name. Only problem that I can think of is that some CI systems check out projects in generic directory names like /src which would break the fallback.
| gharchive/issue | 2023-10-20T12:48:56 | 2025-04-01T04:35:34.084781 | {
"authors": [
"ruben-arts",
"silverwind"
],
"repo": "prefix-dev/pixi",
"url": "https://github.com/prefix-dev/pixi/issues/399",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1578442626 | Install as LaunchDaemons
Closes https://github.com/preludeorg/libraries/issues/333
Install needs to run as sudo
sudo -E -- ./install.sh ...
| gharchive/pull-request | 2023-02-09T19:08:42 | 2025-04-01T04:35:34.086017 | {
"authors": [
"mireaulf"
],
"repo": "preludeorg/libraries",
"url": "https://github.com/preludeorg/libraries/pull/335",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
478900496 | use tagged version instead of latest
update the helm chart to use a tagged version instead of latest.
Hi @cpanato, we set those fields at tag time using some scripts.
This is helpful only when installing the chart from file. Is there a workflow that will help to have those fileds set?
@AMecea ah thanks for the explanation. I don't have any workflow just following some best practices for the charts.
however, I saw there are some tags and this chart are not updated. Or there are any other repo that is getting updated?
thanks again
Welcome,
The charts are published here. The changes are not saved in the repo at all. Are only changed temporarily before publishing them in the chart repo.
So when you run helm install presslabs/mysql-operator will install the latest tag with the images set accordingly.
| gharchive/pull-request | 2019-08-09T09:50:12 | 2025-04-01T04:35:34.110224 | {
"authors": [
"AMecea",
"cpanato"
],
"repo": "presslabs/mysql-operator",
"url": "https://github.com/presslabs/mysql-operator/pull/383",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
529672109 | query the phoenix table , desc table , can't find the TIMESTAMP type columns,how can i do?
when use phoenix i can see the timestamp type columns,
but can't find in presto:
the presto version is 324
Unfortunately, the type isn't supported for now.
https://github.com/prestosql/presto/blob/master/presto-phoenix/src/main/java/io/prestosql/plugin/phoenix/TypeUtils.java
| gharchive/issue | 2019-11-28T03:26:45 | 2025-04-01T04:35:34.143453 | {
"authors": [
"ebyhr",
"rongyousu"
],
"repo": "prestosql/presto",
"url": "https://github.com/prestosql/presto/issues/2126",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
453442098 | why discontinued?
Hi, as you know firebase is only platform that support flutter but firebase will be too expensive in grow. Will you continue this library or finished?
I don't want to call it discontinued.
It is just that I am currently not able to provide enough time on this, that's all.
But i am open to review and accept any pull request, fix any bugs.
or if @stafyniaksacha is willing to mantain this in the long term, i can give him that access
@prijindal, yes if you need, I can help maintain this project (I'm an old member of kuzzle' core team)
I'm also planing to work on this project again to introduce some new features that are already implemented into the new javascript SDK.
Don't hesitate to ask questions about kuzzle , the sdk, or even request for features !
@c2c2 Concerning this SDK we (the Kuzzle team) can not provide the same level of support than for our official project.
But as @stafyniaksacha says, he is an old member of our team and we trust him to provide the best quality support for this SDK.
Concerning the backend core (http://github.com/kuzzleio/kuzzle) is a long term project. We have a LTS development cycle, every version of Kuzzle is supported for at least 18 month.
Kuzzle is a free and open source software but a company is behind it and we have a lot of clients who pay for support so even if I can not say that Kuzzle will last forever, it will stay for a while I think ;)
Thanks for your responsibility but i ran an startup with dead time and i think maybe another project i'll use kuzzle for backend
@c2c2 You are starting a new project using Flutter ?
Yes i do
@c2c2 What are missing for your project ? Normally all functionalities are implemented as javascript sdk v6.0.0.
Don't hesitate to ask for questions about implementation with flutter here, I'll be happy to help you!
You may find support for global kuzzle usage on the official gitter chat here: https://gitter.im/kuzzleio/kuzzle
I saw that there are plan to officially do support for this sdk on official roadmap here: https://trello.com/c/b28UCNX3/41-official-dart-flutter-sdk
| gharchive/issue | 2019-06-07T10:21:01 | 2025-04-01T04:35:34.219748 | {
"authors": [
"Aschen",
"c2c2",
"prijindal",
"stafyniaksacha"
],
"repo": "prijindal/kuzzle_dart",
"url": "https://github.com/prijindal/kuzzle_dart/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
416382726 | React does not recognize the maxWidth prop on a DOM element when using Box
When using the Box component with the maxWidth prop, I receive an error saying:
React does not recognize the maxWidth prop on a DOM element. If you intentionally want it to appear in the DOM as a custom attribute, spell it as lowercase maxwidth instead. If you accidentally passed it from a parent component, remove it from the DOM
My code:
import React from 'react'
import { Box } from '@primer/components'
export default class Home extends React.Component {
public render() {
return (
<div>
<Box maxWidth={2}>Hello World</Box>
</div>
)
}
}
The emotion styles are still applied, but so is the html attribute to the dom.
Am I doing something wrong or is this a bug?
@Acidic9 Hi! the maxWidth prop needs a unit such as px. Every prop from styled-system (the library we use for our system props) is a CSS prop unless it says otherwise, so using any valid CSS unit will work here. Unsure if anything else in your setup would cause a problem though.
The emotion styles are still applied, but so is the html attribute to the dom.
FYI we switched to styled-components from emotion several versions ago (I think v8 or before) so you may want to update the version you're on.
@Acidic9 I believe this is because some props are not getting filtered out of the remaining HTML attributes that get passed to the DOM by styled-components. I've been noticing this with a few other props and need to look into why this is still happening! It shouldn't affect anything besides having to see the ugly warning in your console. Thanks for the heads up :)
This should have been cleared up with the 12.0.1 release! Please let me know if you're still seeing this :)
| gharchive/issue | 2019-03-02T10:19:43 | 2025-04-01T04:35:34.291812 | {
"authors": [
"Acidic9",
"broccolini",
"emplums"
],
"repo": "primer/components",
"url": "https://github.com/primer/components/issues/416",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
926926531 | is this should be distance?
cosine(embeddings[0], embeddings[1]) is the cosine similarities of 0 and 1,then 1-cosine should be called distance of 0 an 1?
##Calculate cosine similarities
##Cosine similarities are in [-1, 1]. Higher means more similar
cosine_sim_0_1 = 1 - cosine(embeddings[0], embeddings[1])
cosine_sim_0_2 = 1 - cosine(embeddings[0], embeddings[2])
print("Cosine similarity between "%s" and "%s" is: %.3f" % (texts[0], texts[1], cosine_sim_0_1))
print("Cosine similarity between "%s" and "%s" is: %.3f" % (texts[0], texts[2], cosine_sim_0_2))
Hi,
cosine in scipy is 1 - cosine similarity, so we need to do 1 - cosine to get the cosine similarity.
Hi,
cosine in scipy is 1 - cosine similarity, so we need to do 1 - cosine to get the cosine similarity.
好的,了解了,谢谢~
| gharchive/issue | 2021-06-22T07:36:00 | 2025-04-01T04:35:34.297821 | {
"authors": [
"gaotianyu1350",
"qiuyu666"
],
"repo": "princeton-nlp/SimCSE",
"url": "https://github.com/princeton-nlp/SimCSE/issues/46",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1637195037 | CoreData: warning: Unable to load class named '' for entity 'Entity'. Class not found, using default NSManagedObject instead.
CoreData: warning: Unable to load class named '' for entity 'Entity'. Class not found, using default NSManagedObject instead.
How to resolve this warning?
Ipad mini 6(Ios 16.3)
Mac Air 2020 M1 (MAC OS 13.3)
XCode 14.2(14C18)
Hello! please ignore these warnings - this is an internal log inside the CoreData framework and does not affect any functionality.
| gharchive/issue | 2023-03-23T09:54:01 | 2025-04-01T04:35:34.301916 | {
"authors": [
"geor-kasapidi",
"jidaojiuyou"
],
"repo": "prisma-ai/Sworm",
"url": "https://github.com/prisma-ai/Sworm/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
694040622 | Associate with graphqls file extension
Fixes #176
and here as well
https://github.com/prisma-labs/vscode-graphql/blob/6f1e9252f9b5074333b02999d12cfc54f46a0859/src/extension.ts#L76
PR updated, thank you for the indications.
I'll be cutting a new pre-release of the language server by monday, so once we cut that, we can bump the dependency in this PR and be good to go!
@cailloumajor ok! the stable release of graphql-language-service-server is out, and we just merged a package bump, so this PR is good to go!
| gharchive/pull-request | 2020-09-05T09:04:03 | 2025-04-01T04:35:34.305240 | {
"authors": [
"acao",
"cailloumajor"
],
"repo": "prisma-labs/vscode-graphql",
"url": "https://github.com/prisma-labs/vscode-graphql/pull/211",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
366901131 | Order in express middleware
Hello,
I currently use apollo-server-express with next.js. I REALLY want to replace apollo-server with graphql-yoga but it does not work.
In graphql-yoga, middlewares are triggered before endpoint and playground createHttpServer
next's router is a catch all that returns 404 if there is no matching route. run
As next's router is triggered before yoga's endpoints, when I try to reach '/graphql' I get 404.
With apollo-server-express, I don't have that problem because I have a way to order express middleware applyMiddleware
Could you change middleware order so that endpoint, playground are before middlewares ?
Thanks
Here is my code with graphql-yoga :
import next from 'next';
import routes from 'next-routes';
import { GraphQLServer } from 'graphql-yoga';
import schema from './models/schema';
const nextApp = next();
const handler = routes()
.add('home', '/', 'Home')
.getRequestHandler(nextApp);
nextApp.prepare().then(() => {
const server = new GraphQLServer({ schema }).use(handler);
server.start({
endpoint: '/graphql',
playground: '/graphql',
getEndpoint: true
});
});
@kevinya Wrap the next handler in a middleware that checks the request is not for the /graphql route.
nextServer.prepare().then(() => {
graphQLServer.use((req, res, next) => {
if (req.path.startsWith("/graphql")) return next();
nextServerHandler(req, res, next);
});
graphQLServer
.start({
endpoint: '/graphql',
playground: '/graphql',
getEndpoint: true
})
.then(() => {
console.log("Next.js app is running on http://localhost:3000");
console.log("GraphQL API is running on http://localhost:3000/graphql");
})
.catch(() => {
console.error("Server start failed", err);
process.exit(1);
});
});
I just removed getEndpoint: true and it works. Here is a server.js file
`const next = require('next');
const { GraphQLServer } = require('graphql-yoga');
const dev = process.env.NODE_ENV !== 'production';
const port = process.env.PORT || 3000;
const gqlEndpoint = '/gql';
const app = next({ dev });
const handle = app.getRequestHandler();
app.prepare().then(() => {
const typeDefs = type Query { hello(name: String): String! } ;
const resolvers = {
Query: {
hello: (_, { name }) => `Hello ${name || 'World'}`,
},
};
const server = new GraphQLServer({ typeDefs, resolvers });
server.use((req, res, next) => {
if (req.path.startsWith(gqlEndpoint)) return next();
handle(req, res, next);
});
server.start( {
endpoint: gqlEndpoint,
playground: gqlEndpoint,
subscriptions: gqlEndpoint,
port,
}).catch(err => console.log('ERROR:', err))
});
`
Just adding this to the pool. Below, working example that combines next and graphql-yoga (along with some integration of Prisma’s newly-released PhotonJS)
// NextJS
const nextJS = require('next')
const nextRoutes = require('next-routes')
// GraphQL Yoga
const { GraphQLServer } = require('graphql-yoga')
// Prisma Photon
const Photon = require('@generated/photon')
// Custom functions
const { resolvers } = require('./server/resolvers')
// Photon
const photon = new Photon.default()
// NextJS
const nextJSApp = nextJS({ dev: process.env.NODE_ENV !== 'production' })
const nextJSRoutes = nextRoutes()
.add({ pattern: '/', page: '/' })
.add({ pattern: '/signin', page: '/signin' })
.getRequestHandler(nextJSApp)
// GraphQL Endpoint
const gqlEndpoint = '/server'
const gqlServer = new GraphQLServer({
typeDefs: 'server/schema.graphql',
resolvers,
context: data => ({ ...data, photon })
})
// Begin NextJS...
nextJSApp.prepare().then(() => {
// ...check if GraphQL endpoint is pinged...
gqlServer.use((req, res, next) => {
if (req.path.startsWith(gqlEndpoint)) return next()
// ... if not, use NextJS routes.
nextJSRoutes(req, res, next)
})
// Start server.
gqlServer
.start(
{
endpoint: gqlEndpoint,
playground: gqlEndpoint,
subscriptions: gqlEndpoint,
port: process.env.PORT || 3000
},
() => console.log(`\n🚀 GraphQL server ready at http://localhost:4000`)
)
.then(httpServer => {
async function cleanup() {
console.log(`\n\nDisconnecting...`)
await photon.disconnect()
httpServer.close()
console.log(`\nDone.\n`)
}
// process.on('SIGINT', cleanup)
process.on('SIGTERM', cleanup)
})
.catch(err => {
console.error('Server start failed ', err)
process.exit(1)
})
})
| gharchive/issue | 2018-10-04T17:50:18 | 2025-04-01T04:35:34.313021 | {
"authors": [
"djgrant",
"gsanikidze",
"heymartinadams",
"kevinya"
],
"repo": "prisma/graphql-yoga",
"url": "https://github.com/prisma/graphql-yoga/issues/456",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1092571454 | Full-text search doesn't work with filter operators/ conditions – OR
Bug description
Executing the following query throws an error when using the OR operator with full-text search:
const query = 'Prisma rocks'.split(' ').join('&')
const results = await prisma.book.findMany({
where: {
OR: {
title: {
search: query
},
content: {
search: query
}
}
},
})
The following error is thrown by Prisma Client:
{ query: 'Prisma & rocks' }
{
error: PrismaClientUnknownRequestError:
Invalid `prisma.book.findMany()` invocation:
Error occurred during query execution:
ConnectorError(ConnectorError { user_facing_error: None, kind: QueryError(Error { kind: Db, cause: Some(DbError { severity: "ERROR", parsed_severity: Some(Error), code: SqlState("42601"), message: "syntax error in tsquery: \"david copperfield\"", detail: None, hint: None, position: None, where_: None, schema: None, table: None, column: None, datatype: None, constraint: None, file: Some("tsquery.c"), line: Some(689), routine: Some("makepol") }) }) })
at cb (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/@prisma/client/runtime/index.js:38692:17)
at async handler (webpack-internal:///./pages/api/search.ts:40:29)
at async Object.apiResolver (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/api-utils.js:102:9)
at async DevServer.handleApiRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1064:9)
at async Object.fn (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:951:37)
at async Router.execute (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/router.js:222:32)
at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1135:29)
at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/dev/next-dev-server.js:445:20)
at async DevServer.handleRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:325:20) {
clientVersion: '3.7.0'
}
If filter operators are not supposed to work when paired with full-text search, we should provide some level of type-safety around it.
How to reproduce
Query
const query = 'Prisma rocks'.split(' ').join(' & ')
const results = await prisma.book.findMany({
where: {
OR: {
title: {
search: query
},
content: {
search: query
}
}
},
})
Expected behavior
No error if the query is executed, or provide type safety that filter operators can't be used when using FTS
Prisma information
Schema:
generator client {
provider = "prisma-client-js"
previewFeatures = ["fullTextSearch"]
}
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
model Book {
id Int @id @default(autoincrement())
title String
content String
url String
cover String
authors String[]
}
Migration
Environment & setup
OS: Mac OS
Database: PostgreSQL
Node.js version: 16.13
Prisma Version
prisma : 3.7.0
@prisma/client : 3.7.0
Current platform : darwin
Query Engine (Node-API) : libquery-engine 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 8746e055198f517658c08a0c426c7eec87f5a85f
Studio : 0.445.0
Preview Features : fullTextSearch
This was a mistake on my end. There's no bug here 🙂
| gharchive/issue | 2022-01-03T14:46:40 | 2025-04-01T04:35:34.323080 | {
"authors": [
"ruheni"
],
"repo": "prisma/prisma",
"url": "https://github.com/prisma/prisma/issues/10930",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1796376010 | PrismaClient is unable to be run in the browser.
Bug description
Got this error in the prismaClient configuration file.
How to reproduce
Expected behavior
No response
Prisma information
generator client {
provider = "prisma-client-js"
}
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
directUrl = env("DIRECT_URL")
shadowDatabaseUrl = env("SHADOW_DATABASE_URL")
}
model Account {
id String @id @default(cuid())
userId String
type String
provider String
providerAccountId String
refresh_token String? @db.Text
access_token String? @db.Text
expires_at Int?
token_type String?
scope String?
id_token String? @db.Text
session_state String?
user User @relation(fields: [userId], references: [id], onDelete: Cascade)
@@unique([provider, providerAccountId])
}
model Session {
id String @id @default(cuid())
userId String
expires DateTime
sessionToken String @unique
accessToken String @unique
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
user User @relation(fields: [userId], references: [id])
}
model User {
id String @id @default(cuid())
name String?
email String? @unique
emailVerified DateTime?
image String?
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
accounts Account[]
sessions Session[]
}
model VerificationRequest {
id String @id @default(cuid())
identifier String
token String @unique
expires DateTime
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
@@unique([identifier, token])
}
import { PrismaClient } from '@prisma/client';
declare global {
// eslint-disable-next-line no-var, no-unused-vars
var cachedPrisma: PrismaClient;
}
let prisma: PrismaClient;
if (process.env.NODE_ENV === 'production') {
prisma = new PrismaClient();
} else {
if (!global.cachedPrisma) {
global.cachedPrisma = new PrismaClient();
}
prisma = global.cachedPrisma;
}
export const db = prisma;
Environment & setup
OS: macOS
Database: PostgreSQL
Node.js version: v19.7.0
Prisma Version
prisma : 4.16.2
@prisma/client : 4.16.2
Current platform : darwin-arm64
Query Engine (Node-API) : libquery-engine 4bc8b6e1b66cb932731fb1bdbbc550d1e010de81 (at node_modules/.pnpm/@prisma+engines@4.16.2/node_modules/@prisma/engines/libquery_engine-darwin-arm64.dylib.node)
Migration Engine : migration-engine-cli 4bc8b6e1b66cb932731fb1bdbbc550d1e010de81 (at node_modules/.pnpm/@prisma+engines@4.16.2/node_modules/@prisma/engines/migration-engine-darwin-arm64)
Format Wasm : @prisma/prisma-fmt-wasm 4.16.1-1.4bc8b6e1b66cb932731fb1bdbbc550d1e010de81
Default Engines Hash : 4bc8b6e1b66cb932731fb1bdbbc550d1e010de81
Studio : 0.484.0
I am implementing next auth, here's my auth file:
import { db } from '@/lib/db';
import { PrismaAdapter } from '@next-auth/prisma-adapter';
import { NextAuthOptions } from 'next-auth';
import GoogleProvider from 'next-auth/providers/google';
function getGoogleCredentials(): { clientId: string; clientSecret: string } {
const clientId = process.env.GOOGLE_CLIENT_ID;
const clientSecret = process.env.GOOGLE_CLIENT_SECRET;
if (!clientId || clientId.length === 0) {
throw new Error('Missing GOOGLE_CLIENT_ID');
}
if (!clientSecret || clientSecret.length === 0) {
throw new Error('Missing GOOGLE_CLIENT_SECRET');
}
return { clientId, clientSecret };
}
export const authOptions: NextAuthOptions = {
secret: process.env.NEXTAUTH_SECRET,
adapter: PrismaAdapter(db),
session: {
strategy: 'jwt',
},
pages: {
signIn: '/',
},
providers: [
GoogleProvider({
clientId: getGoogleCredentials().clientId,
clientSecret: getGoogleCredentials().clientSecret,
}),
],
callbacks: {
async session({ token, session }) {
if (token) {
session.user.id = token.id;
session.user.name = token.name;
session.user.email = token.email;
session.user.image = token.picture;
}
return session;
},
async jwt({ token, user }) {
const dbUser = await db.user.findFirst({
where: {
email: token.email,
},
});
if (!dbUser) {
token.id = user!.id;
return token;
}
return {
id: dbUser.id,
name: dbUser.name,
email: dbUser.email,
picture: dbUser.image,
};
},
redirect() {
return '/';
},
},
};
Could you please provide us with a reproduction repository?
Somehow it seems this is included in your frontend? The Prisma Client would then throw an error.
I'm having a similar issue but mine is in the edge I believe. Any suggestions? For context, I'm also using nextauth with next13 with a similar auth setup as @millsp
I'm having a similar issue but mine is in the edge I believe. Any suggestions? For context, I'm also using nextauth with next13 with a similar auth setup as @exosky12
Could you please provide us with a reproduction repository?
Somehow it seems this is included in your frontend? The Prisma Client would then throw an error.
I'm just trying to initialize Prisma, not calling anything for now @millsp :/
This code suggests that at some point something called new PrismaClient either in the browser-side code or somehow it was bundled incorrectly. It would help us if you could share a small repro repository with what is happening.
Yes, I think the problem comes from my getServerSession, how can I get the actual session (next auth) without causing troubles ?
I will do a reproduction repository as fast as possible.
This code suggests that at some point something called new PrismaClient either in the browser-side code or somehow it was bundled incorrectly. It would help us if you could share a small repro repository with what is happening.
I know it's not a Repo, but this error is happening to me when I use Prisma from a Server Function (server action) that's being called from a Client Component in NextJS using the App Router.
It would be really, really helpful if someone could take the time to put a case like this into a reproduction repository. We'll be able to take that and make it run any time, but setting it up from scratch will take us to properly reserve time for this - which we barely have any right now. Sorry - so please someone help us out. Thanks!
I had this same error and I was able to resolve it by updating the auth file.
Instead of using Prisma Adapter directly in the authOptions, use it seperately
Instead of this
export const authOptions: NextAuthOptions = {
secret: process.env.NEXTAUTH_SECRET,
adapter: PrismaAdapter(db),
session: {
strategy: 'jwt',
},
Do this
const adapter = PrismaAdapter(prisma);
export const authOptions: NextAuthOptions = {
secret: process.env.NEXTAUTH_SECRET,
adapter: adapter,
session: {
strategy: 'jwt',
},
Update me if this works out for you, it did for me anyway!
I got the same error because I was passing data as props that were called in the layout file I assigned as a client component, "use client".
Hi, everything good? I had the same problem few minutes ago, after a some tries, the thing was everything the frontend and how i treated the client side button.
I just made an AuthenticationButton recieving provider.id from getProviders method in Server Component:
import { AuthenticationButton } from '@/components/AuthenticationButton';
import { getProviders } from 'next-auth/react';
import Image from 'next/image';
import { FcGoogle } from 'react-icons/fc';
export default async function SignUp() {
const providers = await getProviders();
return (
<main>
<div className='flex flex-col min-h-[calc(100vh/2)] bg-[#5528ff] z-10'></div>
<div className='flex flex-col h-1/2'>
<div className='p-8 min-w-[400px] mx-auto bg-white -mt-12 z-20 border border-gray-300 rounded-lg'>
<div className='flex flex-col items-center pb-8'>
<Image src='/promogate-logo.svg' alt='Promogate Logo' width={120} height={60} />
</div>
{
Object.values(providers!).map((provider, i) => {
return (
<AuthenticationButton key={i} className='border border-gray-500 p-2 w-full text-sm text-center'
providerId={provider.id}>
<FcGoogle />
Cadastrar-se com {provider.name}
</AuthenticationButton>
)
})
}
</div>
</div>
</main>
)
'use client'
import { openSans } from '@/application/utils/fonts'
import { signIn } from 'next-auth/react'
import React, { ButtonHTMLAttributes } from 'react'
interface ButtonProps extends ButtonHTMLAttributes<HTMLButtonElement> {
children: React.ReactNode,
className?: string,
providerId: string
}
export function AuthenticationButton({ children, className, providerId, ...rest }: ButtonProps) {
return (
<button {...rest} className={'p-4 rounded-lg rounded-tl-none flex justify-center items-center gap-x-4' + ' ' + className}
style={openSans.style} onClick={() => signIn(providerId)}>
{children}
</button>
)
}
No prisma errors doing this.
I had a similar issue due to me importing a prisma instance in my nextauth-file. I then exported the authOptions to be used with getServerSession, which I then had imported in a file shared between both the backend and the frontend
The same error occurred in my monorepo repository as well.
The cause was that I consolidated both frontend and backend functions into index.ts. As you all have pointed out, it was directly because I imported authOptions, but be careful as there seems to be a pattern where it's indirectly imported.
index.ts (Bug occurs)
export { default } from "next-auth";
export { signIn, signOut, useSession } from "next-auth/react";
export { authOptions } from "./src/auth-options";
export { getServerSession } from "./src/get-session";
👇Fixed
client.ts
export { default } from "next-auth";
export { SessionProvider, signIn, signOut, useSession } from "next-auth/react";
server.ts
export { authOptions } from "./src/auth-options";
export { getServerSession } from "./src/get-session";
package.json
"files": [
"client.ts",
"server.ts"
],
Can someone please help us get a working reproduction of the problem? The code snippets above might be helpful if you know Next.js and next-auth, but it would take us quite some time to put that together. It would be amazing if one of you could create a new project and implement the minimal code to show the problem - and then put it on GitHub. Thank you!
While we still don't have a reproduction of this issue, I don't think this is a Prisma issue per se.
The problem to me is that Prisma Client should be used in a backend JS runtime (e.g., Node.js) only, but when developing isomorphic Next.js projects (in which the separation between frontend and backend isn't necessarily super clear at first glance), the Prisma Client instance can easily leak into the frontend.
And, since these leaks cause Prisma Client to run in a browser (which is not supported), the runtime error above is thrown.
In addition to what @jkomyno said, we recently fixed the fact that that we threw errors when doing new PrismaClient and now we moved these errors to only happen if you try emitting a query in your front-end (which obviously cannot work).
| gharchive/issue | 2023-07-10T09:39:32 | 2025-04-01T04:35:34.340500 | {
"authors": [
"AbelR007",
"Oliveira-86",
"TakashiAihara",
"alvesvaren",
"evanrosa",
"exosky12",
"gutodiasdev",
"janpio",
"jkomyno",
"millsp",
"nukkerone"
],
"repo": "prisma/prisma",
"url": "https://github.com/prisma/prisma/issues/20147",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
811067103 | POC using Prisma Migrate for CD
Acceptance criteria:
Generic steps to apply for CD
A working POC of applying migrations to a cloud deployment via prisma migrate deploy with Heroku Release
Rough documentation with the steps involved and any potential caveats
This is done:
Generic deployment guide
Vercel deployment guide
Heroku deployment guide
| gharchive/issue | 2021-02-18T12:45:11 | 2025-04-01T04:35:34.344789 | {
"authors": [
"albertoperdomo"
],
"repo": "prisma/prisma",
"url": "https://github.com/prisma/prisma/issues/5725",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1466112693 | ci(no-docker client tests): lower --max-old-space-size for linux and windows
https://prisma-company.slack.com/archives/CUXLS0Z6K/p1669489504119379
the conditional works!
| gharchive/pull-request | 2022-11-28T09:45:52 | 2025-04-01T04:35:34.346352 | {
"authors": [
"Jolg42"
],
"repo": "prisma/prisma",
"url": "https://github.com/prisma/prisma/pull/16485",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
539253246 | Prisma migration engine on s3 has been deleted, cannot install on CI
Problem
When installing Prisma2, certain files from an s3 bucket (prisma-native) are downloaded. Just starting this morning, these files can now not be found.
How To Reproduce
Run npm i -g --unsafe-perm prisma2@2.0.0-preview014.2 in CircleCI
Thanks for reporting, we will look into this.
Until then, installing a recent version of prisma2 will probably solve that problem. See https://github.com/prisma/prisma2/releases for the most recent release.
@janpio Thanks for the reply.
We have plans to upgrade but we found it will take significant effort due to our migrations not being compatible.
| gharchive/issue | 2019-12-17T19:02:43 | 2025-04-01T04:35:34.348811 | {
"authors": [
"aleccool213",
"janpio"
],
"repo": "prisma/prisma2",
"url": "https://github.com/prisma/prisma2/issues/1172",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2144982408 | Native node modules for downloading artifacts
Related Issue(s)
Closes #166
Checklist
[x] My code follows the style guidelines of this project
[x] I have performed a self-review of my code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[x] My changes generate no new warnings
[x] I have run yarn prettier and yarn lint without getting any errors
[x] I have added tests that prove my fix is effective or that my feature works
[x] New and existing unit tests pass locally with my changes
[x] Any dependent changes have been merged and published in downstream modules
LGTM. Just one thing, there is a duplication between eddsa-proof and poseidon-proof packages (the functions are the same). Wouldn't it make sense to have a shared utility?
Yes, I think some code could be moved to a shared function (maybe in the utils package). Could you open another issue for this?
LGTM. Just one thing, there is a duplication between eddsa-proof and poseidon-proof packages (the functions are the same). Wouldn't it make sense to have a shared utility?
Yes, I think some code could be moved to a shared function (maybe in the utils package). Could you open another issue for this?
Sure, tracked in https://github.com/privacy-scaling-explorations/zk-kit/issues/176
| gharchive/pull-request | 2024-02-20T18:10:13 | 2025-04-01T04:35:34.370009 | {
"authors": [
"0xjei",
"cedoor"
],
"repo": "privacy-scaling-explorations/zk-kit",
"url": "https://github.com/privacy-scaling-explorations/zk-kit/pull/173",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
932473105 | 🆕 Software Suggestion | Responding to reply on post being removed
Basic Information
Name: lazyweb.ai (redditor lazy-jem)
Category: Search engine
URL: lazyweb.ai
Description
Hi, I'm sorry. I posted earlier tonight seeking feedback about our alpha version of a new search engine app. It isn't a commercial app yet, so it wasn't intended as promotion, but the post was removed by /u/trai_dep with the following note:
If you have a project that you want to promote here, open an issue on our GitHub repo so our entire team can advise and evaluate it first.
I'm sorry I didn't appreciate that I should have cleared a post seeking feedback like this with the mods first. We aren't ready to promote it yet, as it isn't a commercial release. I thought the community would find it interesting and have useful feedback on how we could make what we're making more privacy-focused. We are mission driven and it wasn't our intention to be self promotional here as we are not at that stage.
With 129 upvotes and 28 comments in the first few hours, it did seem interesting to the community, and I'm sorry we misunderstood about doing this the right way, and would very much like to work with you to correct the mistake. A search engine is obviously not typical privacy software, as it isn't run directly by the user. I built this out of frustration with ad-tech and privacy invasion. So it is a personal project that I though other folks might find interesting.
Why I am making the suggestion
Here are some comments from the post:
Made a search web app that's anonymous, ad-free and non-tracking. It lets you read web content in a clean reader view anonymously through a proxy that strips ad-tech and tracking. Looking for feedback on our approach to privacy in the new alpha test version we just released.
It's called LazyWeb. It uses a chat interface and gives you control of how you view search results. The chat interface means that searches stay within the anonymous chat session rather than going through the browser history.
The alpha version is open to anyone to try here - https://lazyweb.ai
We're a small bootstrapped two-person team. I'm the technical co-founder. We'd be grateful for thoughts, suggestions, and feedback on how we're approaching privacy.
LazyWeb doesn't log searches, and it blocks tracking and ad-tech. We only collect and retain sufficient data to improve the service we provide and help our customers use the service effectively, or in future if customers want to create an account or be remembered between devices and sessions.
We use limited in-app analytics solely to help improve the application for people using it. The metrics are anonymous, reported in aggregate, and it do not contain any personal data or searches. They are not shared with anyone. You can disable all in-app analytics on the settings page.
One of the big challenge building a private, anonymous search engine is that nothing is logged, so we can't see what people search, or when things go wrong searching. So the only way to keep the results improving and fix problems is to get lots of feedback from people who are subject matter experts. Technical searches need a lot of work and we'd love any feedback on the search results in specialized areas (like privacy and security).
My connection with the software
Author
[YES ] I will keep the issue up-to-date if something I have said changes or I remember a connection with the software.
@jedwhite you mention that metrics are anonymous, so could you specify what exactly falls under 'metrics' for you and how do you ensure the anonymity hereof?
What is the GEO IP service for and are IP's shared with third parties for this? Doesn't sound very privacy friendly.
I also cannot find any real privacy policy, which you are required to have by the GDPR. I think this should be closed and maybe you could open a new issue when you have figured things out.
Thank you for taking the time to respond and for your feedback.
I think there may be some confusion about the request. I posted here at the suggestion of /u/trai_dep from the subreddit. We are not asking to be listed as suggested software or be recommended, and we are a long way from that point. The purpose of the subreddit post was explicitly to ask for help from the community to seek feedback on our approach at a very early stage, and to figure things out to do better.
As noted, the web app is an alpha test of a prototype and proof-of-concept only, and not a commercial service. We are not promoting it for commercial or production use, and searches are currently limited to en-US only.
I'm grateful for you taking to time to comment here, and this is exactly the type of feedback that we were seeking by posting on the subreddit. We definitely aren't asking for a recommendation or anything of the kind at this point, but we would like to work with you to be able to discuss with other people interested in privacy how we can build a better alternative to Google, DDG, Bing, Brave etc.
Thanks for the chance to answer some of your questions too. While it is an early prototype, we have a plain language privacy policy here (https://lazyweb.ai/privacy/). It is one of the things we are seeking feedback on at an early stage. Our informal legal advice is that it does meet the GDPR requirements broadly given we do not store any personal information but that we need to add more detail (especially how we collect the in-app analytics and storage).
With the GEOIP:
We don't log or store IP addresses. It's used to lookup your approximate location (nearest city) for location searches only, then discarded. It is never passed to third-parties. We only use a GPS or detailed location for searches with a user's express permission, and then only to approximate the area (nearest city). Your GPS location details are not stored or passed to any third-parties. We're using the maxmind database. We don't log or store IP addresses or any other geolocation data. For the analytics, we collect the city location name, and then summarise the number of users for each city by day. We don't log or store searches including searches that incorporate location elements. IP and actual geo coordinates are not passed with searches, and retrieved only by the browser client side. So the geo lookup API is isolated from searches and not sent to the search back end. That's why we lookup geo data on the client as a separate process, and pass location city name with search requests, rather than doing it server side from the search engine.
Originally, we completely disabled location until enabled. But people trying the app told us that was a terrible experience, because there is an expectation when someone asks the time, or weather, or the best coffee shop that the results will be at least nearby to them without them having to explictly enable location (we recognize that is a paradox). So we're trying to determine what the best trade-off is. Currently, we determine the broad city location from a separate independent client-side lookup, and a user can enable GPS (but it is still approximated to the closest city - just more accurately as IP/maxmind can be wildly inaccurate), and only the location city name is used for search localization. As far as we are aware, other major commercial search engines use IP geocoordinates with full fidelity server-side by default for location, so we think this is a better approach. But it is definitely one of the things we're seeking ideas and feedback on.
With the in-app analytics, there is no personal information passed (no IP, no geocoords, no search request content). We record signals on the broad search intent (for example, a programming search intent or food place search intent) grouped into about 60 intent categories. No search terms are passed. And we record the types of actions performed (again only by broad category - for example "external link clicked" but not what any links clicked were. We don't record referrer or any other information. We don't record what items people read or navigate to. This data is reduced to summary figures by hour and day, and the original events discarded. There is a session client identifier. We don't pass user agent strings, usernames, or any PII. We pass whether it is a first time session or returning session, mobile or desktop, and top-level browser name and OS type (again by category, not user agent string). It's worth noting that searches are not even passed to the web browser history, as they remain within the chat session, and that is destroyed with each page load. So they are even masked from Google's typical Chrome-level tracking. We have tried to find the best way to approach understanding how to improve the app while keeping use anonymous, and it is one of the areas we are hoping to get feedback on and improve.
Again, we aren't claiming to have solved this or that we have arrived at the right approach, but we are trying to talk with potential future users who care about privacy to work out how to build a good approach.
Thanks again for your feedback and questions, and I hope that's helpful.
@jedwhite thank you for clarification. Obviously you will need to get legal advice (which I cannot give you), but the thing you call a privacy policy does not seem to meet the requirements as defined in the GDPR.
The things you mention about metrics would need to be audited, more then often I have found systems like these not to be implemented accurately and find options remaining to recombine data and being rather pseudonym instead of anonymous. Sessions are connect to a user and therefor likely a pseudonym.
I don't think software suggestion as in your title is the right category. If you are looking for feedback this is likely not the right place to ask for help. Just to be clear, I like that people start up projects like this, but I see my role here to be hesitant and conservative against any new listings that yet not make the difference ;)
I'm going to close this because of the privacy policy issue - but I'll reopen it if/when that is fixed.
| gharchive/issue | 2021-06-29T10:32:16 | 2025-04-01T04:35:34.405428 | {
"authors": [
"freddy-m",
"jedwhite",
"ph00lt0"
],
"repo": "privacytools/privacytools.io",
"url": "https://github.com/privacytools/privacytools.io/issues/2360",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
916616092 | typescript definition
declare module 'html-to-docx'
{
interface Margins
{
/**
* top <Number> distance between the top of the text margins for the main document and the top of the page for all pages in this section in TWIP.
* Defaults to 1440. Supports equivalent measurement in pixel, cm or inch.
*/
top: number;
/**
* right <Number> distance between the right edge of the page and the right edge of the text extents for this document in TWIP.
* Defaults to 1800. Supports equivalent measurement in pixel, cm or inch.
*/
right: number;
/**
* bottom <Number> distance between the bottom of text margins for the document and the bottom of the page in TWIP.
* Defaults to 1440. Supports equivalent measurement in pixel, cm or inch.
*/
bottom: number;
/**
* left <Number> distance between the left edge of the page and the left edge of the text extents for this document in TWIP.
* Defaults to 1800. Supports equivalent measurement in pixel, cm or inch.
*/
left: number;
/**
* header <Number> distance from the top edge of the page to the top edge of the header in TWIP.
* Defaults to 720. Supports equivalent measurement in pixel, cm or inch.
*/
header: number;
/**
* footer <Number> distance from the bottom edge of the page to the bottom edge of the footer in TWIP.
* Defaults to 720. Supports equivalent measurement in pixel, cm or inch.
*/
footer: number;
/**
* gutter <Number> amount of extra space added to the specified margin, above any existing margin values. This setting is typically used when a document is being created for binding in TWIP.
* Defaults to 0. Supports equivalent measurement in pixel, cm or inch.
*/
gutter: number;
}
interface Row
{
/**
* cantSplit <?Boolean> flag to allow table row to split across pages. Defaults to false.
*/
cantSplit?: boolean;
}
interface Table
{
row?: Row;
}
export interface DocumentOptions
{
/**
* orientation <"portrait"|"landscape"> defines the general orientation of the document. Defaults to portrait.
*/
orientation?: "portrait" | "landscape";
margins?: Margins
/**
* title <?String> title of the document.
*/
title?: string;
/**
* subject <?String> subject of the document.
*/
subject?: string;
/**
* creator <?String> creator of the document. Defaults to html-to-docx
*/
creator?: string;
/**
* keywords <?Array<String>> keywords associated with the document. Defaults to ['html-to-docx'].
*/
keywords?: string[];
/**
* description <?String> description of the document.
*/
description?: string;
/**
* lastModifiedBy <?String> last modifier of the document. Defaults to html-to-docx.
*/
lastModifiedBy?: string;
/**
* revision <?Number> revision of the document. Defaults to 1.
*/
revision?: number;
/**
* createdAt <?Date> time of creation of the document. Defaults to current time.
*/
createdAt?: Date;
/**
* modifiedAt <?Date> time of last modification of the document. Defaults to current time.
*/
modifiedAt?: Date;
/**
* headerType <"default"|"first"|"even"> type of header. Defaults to default.
*/
headerType?: "default" | "first" | "even";
/**
* header <?Boolean> flag to enable header. Defaults to false.
*/
header?: boolean;
/**
* footerType <"default"|"first"|"even"> type of footer. Defaults to default.
*/
footerType?: "default" | "first" | "even";
/**
* footer <?Boolean> flag to enable footer. Defaults to false.
*/
footer?: boolean;
/**
* font <?String> font name to be used. Defaults to Times New Roman.
*/
font?: string;
/**
* fontSize <?Number> size of font in HIP(Half of point). Defaults to 22. Supports equivalent measure in pt.
*/
fontSize?: number;
/**
* complexScriptFontSize <?Number> size of complex script font in HIP(Half of point). Defaults to 22. Supports equivalent measure in pt.
*/
complexScriptFontSize?: number;
table?: Table;
/**
* pageNumber <?Boolean> flag to enable page number in footer. Defaults to false. Page number works only if footer flag is set as true.
*/
pageNumber?: boolean;
/**
* skipFirstHeaderFooter <?Boolean> flag to skip first page header and footer. Defaults to false.
*/
skipFirstHeaderFooter?: boolean;
}
/**
* @param htmlString <String> clean html string equivalent of document content.
* @param headerHTMLString <String> clean html string equivalent of header. Defaults to <p></p> if header flag is true.
* @param documentOptions <DocumentOptions>
* @param footerHTMLString <String> clean html string equivalent of footer. Defaults to <p></p> if footer flag is true.
* @constructor
* @private
*/
export default function HTMLtoDOCX(htmlString: string, headerHTMLString: string, documentOptions: DocumentOptions, footerHTMLString?: string): Promise<Buffer | Blob>;
}
@RoxnnyABarriosC Please raise a PR with this typescript definition, and also do we need any sort of typescript configuration to make use of this definition?
Any updates on this ?
@privateOmega RE: typescript. I've created a branch converted to typescript. It's working locally, if useful I can do up a PR.
https://github.com/jafin/html-to-docx/tree/feat/typescript
@jafin when will this PR be up ? it looks promising.
any update on this?
| gharchive/issue | 2021-06-09T19:42:36 | 2025-04-01T04:35:34.421162 | {
"authors": [
"RoxnnyABarriosC",
"jafin",
"lvndry",
"privateOmega",
"rossanodr",
"whis20"
],
"repo": "privateOmega/html-to-docx",
"url": "https://github.com/privateOmega/html-to-docx/issues/79",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1806392404 | feat: Add Theme Component for Toggling light/dark mode
What feature?
Want to add a Theme Component for Toggling light/dark mode in Next.js 13 ...Could you assign this issue to me as a GSSOC23 contributor? I want to make it this way when users click the moon icon UI will appear in dark mode & when it is the sun icon the UI will appear in a light mode
Add screenshots
Record
[X] I agree to follow this project's Code of Conduct
[X] I'm a GSSoC'23 contributor
[X] I want to work on this issue
@crocmons We decided to use only the dark mode, so we removed the option to switch between dark and light modes that we had before.
ok
| gharchive/issue | 2023-07-16T01:30:46 | 2025-04-01T04:35:34.428869 | {
"authors": [
"crocmons",
"priyankarpal"
],
"repo": "priyankarpal/ProjectsHut",
"url": "https://github.com/priyankarpal/ProjectsHut/issues/1719",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2139916238 | hoist change copyright date code from NeXus definitions
close #5
Needs some unit tests and documentation.
Pull Request Test Coverage Report for Build 7940777040
Details
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 63.636%
Totals
Change from base Build 7937257643:
0.0%
Covered Lines:
7
Relevant Lines:
11
💛 - Coveralls
| gharchive/pull-request | 2024-02-17T09:29:18 | 2025-04-01T04:35:34.433823 | {
"authors": [
"coveralls",
"prjemian"
],
"repo": "prjemian/murky",
"url": "https://github.com/prjemian/murky/pull/35",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
587316759 | Double attach is possibly bad?
The problem
In most cases we perform a double attach.
When the user creates a probe, we call attach on the probe: https://github.com/probe-rs/probe-rs/blob/master/probe-rs/src/probe/mod.rs#L123
And when the user attaches to a target, we call attach on the probe again: https://github.com/probe-rs/probe-rs/blob/master/probe-rs/src/probe/mod.rs#L210
This effect can be seen here: https://github.com/probe-rs/cargo-flash/blob/master/src/main.rs#L234-L268.
The effect is possibly bad because if a probe for example does not support SWD and the default is SWD and it tries to use that, hell breaks loose.
Solutions
If we remove the first one, it is not guaranteed that when the user operates on the probe, for example via DapAccess: https://github.com/probe-rs/probe-rs/blob/master/probe-rs/src/probe/mod.rs#L250, that the probe is actually attached to the chip.
If we remove the second one, all the settings like the protocol etc do not get applied.
I see 2 solutions: a) use a typestate to track whether it's attached or not or b) use an internal bool and attach if the probe is not attached.
ping: @Tiwalun @Disasm
Fixed by #200 .
| gharchive/issue | 2020-03-24T22:25:19 | 2025-04-01T04:35:34.514163 | {
"authors": [
"Yatekii"
],
"repo": "probe-rs/probe-rs",
"url": "https://github.com/probe-rs/probe-rs/issues/199",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1190496784 | Book 2 (version 2022-04-01), Section 8.8, Page 373: Code listing uses wrong string quote
The string quote before A needs to be corrected.
It should be jnp.einsum('A ...')
An alternative could be use LaTeX listings, that way the code would also be directly copy-pasteable (will not give quote problems). Example output
LaTeX code used to generate the above listing.
\documentclass{article}
\usepackage{listings}
\usepackage{xcolor}
%New colors defined below
\definecolor{codegreen}{rgb}{0,0.6,0}
\definecolor{codegray}{rgb}{0.5,0.5,0.5}
\definecolor{codepurple}{rgb}{0.58,0,0.82}
\definecolor{backcolour}{rgb}{0.95,0.95,0.92}
%Code listing style named "mystyle"
\lstdefinestyle{mystyle}{
backgroundcolor=\color{backcolour}, commentstyle=\color{codegreen},
keywordstyle=\color{magenta},
numberstyle=\tiny\color{codegray},
stringstyle=\color{codepurple},
basicstyle=\ttfamily\footnotesize,
breakatwhitespace=false,
breaklines=true,
captionpos=b,
keepspaces=true,
numbers=left,
numbersep=5pt,
showspaces=false,
showstringspaces=false,
showtabs=false,
tabsize=2
}
%"mystyle" code listing set
\lstset{style=mystyle}
\begin{document}
%Python code highlighting
\begin{lstlisting}[language=Python, caption=Python example]
import jax.numpy as jnp
from jax import grad
logZ_fun = lambda logpots: np.log(jnp.einsum("A,B,C,AB,BC",
*[jnp.exp(lp) for lp in logpots]))
probs = grad(logZ_fun)(logpots)
\end{lstlisting}
\end{document}
fixed, thanks
| gharchive/issue | 2022-04-02T04:45:29 | 2025-04-01T04:35:34.516652 | {
"authors": [
"murphyk",
"nipunbatra"
],
"repo": "probml/pml-book",
"url": "https://github.com/probml/pml-book/issues/299",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
559273871 | docs: update ZEIT Now deployment instructions
Fixes #985
View rendered docs/deployment.md
nice docs!
small comment (it's outside of your range of changes so couldn't comment on the code): you need to update the hyperlink to the Now docs, since you changed the section name (L17)
- 1. [Now](#now)
+ 1. [Now](#zeit-now)
@benjlevesque Nice catch! I updated that line 👍
Thanks you two :)
| gharchive/pull-request | 2020-02-03T19:06:00 | 2025-04-01T04:35:34.532403 | {
"authors": [
"benjlevesque",
"gr2m",
"styfle"
],
"repo": "probot/probot",
"url": "https://github.com/probot/probot/pull/1130",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
544084625 | Plugin not replacing img with Gatsby Img tag
I've installed everything locally and I'm able to display content from my local wordpress site. The issue I'm having is that the images are all urls from the wordpress site, when I thought that they would be downloaded as static images to the gatsby site and sourced from the static folder. Currently none of that is working. I'm wondering if you can advise.
my github of the demo install is here:
https://github.com/spencersmb/wordpress-graphql-gatsby-demo
I'm able to display my posts and show images, but it doesn't seem to make a difference when using dangerouslySetInnerHtml vs. ContentParser in my code.
In my demo - I added the code to layout.js
Thanks for any help or advice.
Same issue here, maybe these warnings at build time can be useful: warn "createResolvers" passed resolvers for type "wpgraphql_page" that doesn't exist in the schema. Use "createTypes" to add the type before adding resolvers.
Nevermind, with graphqlTypeName: "WPGraphQL" in this plugin configuration matching typeName: "WPGraphQL" in gatsby-source-graphql configuration, it works fine.
| gharchive/issue | 2019-12-31T04:27:48 | 2025-04-01T04:35:34.632746 | {
"authors": [
"jenkin",
"spencersmb"
],
"repo": "progital/gatsby-wpgraphql-blog-example",
"url": "https://github.com/progital/gatsby-wpgraphql-blog-example/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1996949892 | Pods with unready Containers exist on this node, we can't clean the slots yet
Describe the bug
Akri agent daemonset keeps reporting following error whenever any of pod running on a cluster is not ready.
2023-11-16T13:44:46Z TRACE agent::util::slot_reconciliation] reconcile - Pods with unready Containers exist on this node, we can't clean the slots yet
In my case failing POD doesn't use USB resources.
Output of kubectl get pods,akrii,akric -o wide
lpfe04@f1725b929a:~$ kubectl get pod,akrii,akric -n akri
NAME READY STATUS RESTARTS AGE
pod/akri-agent-daemonset-9gwl2 1/1 Running 0 10m
pod/akri-controller-deployment-7c6455f79-zt779 1/1 Running 0 11m
pod/akri-udev-discovery-daemonset-2d9hp 1/1 Running 0 10m
pod/akri-webhook-configuration-7bf6656b45-mclth 1/1 Running 0 11m
NAME CONFIG SHARED NODES AGE
instance.akri.sh/gsm-dongle-6e977d gsm-dongle false ["f1725b929a"] 10m
instance.akri.sh/wifi-dongle-254c38 wifi-dongle false ["f1725b929a"] 10m
instance.akri.sh/wifi-dongle-ac917e wifi-dongle false ["f1725b929a"] 10m
NAME CAPACITY AGE
configuration.akri.sh/gsm-dongle 1 11h
configuration.akri.sh/wifi-dongle 1 11h
Kubernetes Version: [e.g. Native Kubernetes 1.19, MicroK8s 1.19, Minikube 1.19, K3s]
kubernetes: v1.26.8+rke2r1"
Expected behavior
I would expect reconciliation process can be continue if failing pod is out of usb usage.management context.
@rpieczon just to clarify, are you saying if any pod (even if unassociated with Akri) is unready, it causes this slot reconciliation error? From what i remember slot reconciliation should only check pods with an expected annotation.
Exactly in my case I have failing Prometheus POD which has zero requirements related with USB allocation.
Any update on it?
| gharchive/issue | 2023-11-16T14:12:00 | 2025-04-01T04:35:34.649637 | {
"authors": [
"rpieczon"
],
"repo": "project-akri/akri",
"url": "https://github.com/project-akri/akri/issues/681",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2567649776 | Pap 221 frontend create UI components
Ticket
ID: PAP-221
Link: https://project-ascend-io.atlassian.net/browse/PAP-221?atlOrigin=eyJpIjoiNTNjY2Q4YTY4YTIzNDI3ZTgxNWJiYzE5YjQzMGU1ZTIiLCJwIjoiaiJ9
Design File: https://www.figma.com/design/siiCloyWGa7e19lYRr8RP4/Messanging?node-id=0-1&node-type=canvas&t=3M0OEULDR89OcSz6-0
Problem
The application needs a real-time direct messaging feature so users can converse with one another. The User Interface needed to be built to lay the foundation for the DM system.
Solution
Phase 1: I modified the existing codebase to enhance navigation, readability, and UI according to the design file. Full functional code and data persistence will be implemented in Phase 2.
Type of change
[X] Breaking change (fix or feature that would cause existing functionality not to work as expected)
How Has This Been Tested?
Visually tested to match the design file.
Proper testing will be conducted during phase 4.
Checklist:
[X] I have performed a self-review of my code
[X] I have commented my code, particularly in hard-to-understand areas
[X] My changes generate no new warnings
[X] I have added tests that prove my fix is effective or that my feature works
[X] New and existing unit tests pass locally with my changes
Great job, Robert! Everything looks good on I end. I approved.
https://github.com/user-attachments/assets/7d7083aa-16ff-448f-bcdc-5034403eb5f4
| gharchive/pull-request | 2024-10-05T04:23:20 | 2025-04-01T04:35:34.654563 | {
"authors": [
"bobbygrdn",
"oestrada1001"
],
"repo": "project-ascend-io/intracom-electron",
"url": "https://github.com/project-ascend-io/intracom-electron/pull/14",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2515266769 | service "backend" is not running
Describe the bug
This problem occurs when I execute the start script. How do I fix it?I tried upgrading and clearing, and then restarting the script, but it didn't work.
Steps to reproduce the behavior
1、Clone the TH repository
2、Goto to TH folder
3、Install/configure the TH dependencies
./scripts/pi-setup/auto-install.sh
Expected behavior
No response
Log files
No response
PICS file
No response
Screenshots
No response
Environment
No response
Additional Information
No response
The Raspberry PI system version is Ubuntu 24.04.1 LTS
The branch is commit f52d40a16def63edecdabbc8edcb14eb76b12ba6
It appears that the installation steps are incomplete:
Please try these:
cd ~/certification-tool
./scripts/stop.sh
cd ..
rm -rf certification-tool
git clone -b v2.11-beta3.1+fall2024 https://github.com/project-chip/certification-tool.git --recurse-submodules
cd ~/certification-tool
./scripts/pi-setup/auto-install.sh
Problem solved. Thank you for your help @fabiowmm @hiltonlima
| gharchive/issue | 2024-09-10T02:32:26 | 2025-04-01T04:35:34.659639 | {
"authors": [
"hiltonlima",
"song7788"
],
"repo": "project-chip/certification-tool",
"url": "https://github.com/project-chip/certification-tool/issues/412",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1960632125 | subprocess.CalledProcessError: Command '['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']' returned non-zero exit status 1.[BUG]
Reproduction steps
I use this link to setup Matter Device
https://community.arm.com/arm-community-blogs/b/internet-of-things-blog/posts/build-a-matter-home-automation-service-using-raspberry-pi-arm-virtual-hardware-and-python
but i create the step at 8 & 9 always show
error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: legacy-install-failure
× Encountered error while trying to install package.
╰─> grpcio-tools
note: This is an issue with the package mentioned above, not pip.
hint: See above for output from the failure.
ninja: build stopped: subcommand failed.
['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']
Traceback (most recent call last):
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/virtualenv_setup/install.py", line 327, in install_packages
subprocess.check_call(ninja_cmd, stdout=outs, stderr=outs)
File "/usr/lib/python3.11/subprocess.py", line 413, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']' returned non-zero exit status 1.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 795, in
sys.exit(main())
^^^^^^
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 787, in main
return EnvSetup(**vars(parse())).setup()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 457, in setup
result = step(spin)
^^^^^^^^^^
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/env_setup.py", line 606, in virtualenv
if not virtualenv_setup.install(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/virtualenv_setup/install.py", line 342, in install
install_packages(gn_target)
File "/home/patri/connectedhomeip/third_party/pigweed/repo/pw_env_setup/py/pw_env_setup/virtualenv_setup/install.py", line 330, in install_packages
raise subprocess.CalledProcessError(err.returncode, err.cmd,
subprocess.CalledProcessError: Command '['ninja', '-C', '/home/patri/connectedhomeip/.environment/gn_out', '-v', ':python_packages.install']' returned non-zero exit status 1.
So How can i fix this problem
Bug prevalence
everytimes
GitHub hash of the SDK that was being used
db49235c39635582ea522929f9905af03e3114c7
Platform
python, raspi
Platform Version(s)
python 3.11.2 raspi Raspberry Pi OS Lite(64bit) 2023-10-10
Anything else?
No response
The error seems to be not being able to install grpcio-tools.
Are you able to pip install grpcio-tools generally or in a virtualenv?
Platform also seems to state python 3.11.2 raspi Raspberry Pi OS Lite(64bit). For our RPI compiles, we always use an ubuntu version. I would try with Ubuntu 24.04 to make sure you have the highest changes of success. Using RPiOS you are likely using something that not many (guessing nobody else?) are using.
| gharchive/issue | 2023-10-25T06:22:06 | 2025-04-01T04:35:34.671067 | {
"authors": [
"andy31415",
"patrickwang0821"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/issues/29985",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1983978435 | [Build]
Build issue(s)
I have stated my initial question here: https://github.com/project-chip/connectedhomeip/issues/30237 but it was closed before I could test the proposed solution so I opened it as a new issue.
I have changed the lighting-app.zap as suggested to only have one slider in the HomeApp IOs for dimming rather than 6 for different colors by default in lighting-app):
(enable color control: 0)
that was my initial change which didn't help I then dowloaded ZAP GUI and opened lighting-app.zap using it and regenerated .zap then I run script to regenerate new .matter file. I put it in connectedhomeip/examples/lighting-app/light-common
Device says Dimmible Light in ZAP file:
After above changes I have deleted build files in the lighting-app and used idf.py build
I get the following errors when building the file:
I thought that should not be considered anymore as I disabled colourcontrol?
Could you please advise what is wrong ?
Platform
esp32
Anything else?
No response
examples/lighting-app/esp32/main/CMakeLists.txt has:
"${CMAKE_SOURCE_DIR}/third_party/connectedhomeip/src/app/clusters/color-control-server"
which will unconditionally try to compile the color control server. This needs to be changed if you change the ZAP config, no?
| gharchive/issue | 2023-11-08T16:27:07 | 2025-04-01T04:35:34.676154 | {
"authors": [
"BartJozwicki",
"bzbarsky-apple"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/issues/30324",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2047220145 | [Android] The QR code payload parser does not support optional fields
Reproduction steps
On Android call parseQrCode("MT:SA6K48.E15R5WY5.D3P0SOCE0LDQJ1DK5N1K8SQ1RYCU1O0") in SetupPayloadParser class
Examine returned payload
Expected:
An optional field for serialNumber should be present
Actual:
No optional fields are present
Platform
android
Platform Version(s)
No response
Type
Platform validated
(Optional) If manually tested please explain why this is only manually tested
No response
Anything else?
We were hoping to embed a serialNumber in the QR code but Android's payload parser does not support this yet. It is available on iOS as well as the chip tool
./chip-tool payload parse-setup-payload MT:SA6K48.E15R5WY5.D3P0SOCE0LDQJ1DK5N1K8SQ1RYCU1O0
[1702672725954] [5602:2751420] [SPL] Version: 0 [1702672725954] [5602:2751420] [SPL] VendorID: 1234 [1702672725954] [5602:2751420] [SPL] ProductID: 1234 [1702672725954] [5602:2751420] [SPL] Custom flow: 0 (STANDARD) [1702672725954] [5602:2751420] [SPL] Discovery Bitmask: 0x02 (BLE) [1702672725954] [5602:2751420] [SPL] Long discriminator: 717 (0x2cd) [1702672725954] [5602:2751420] [SPL] Passcode: 13360673 [1702672725954] [5602:2751420] [SPL] SerialNumber: CN0123456789
I noticed there is a TODO to add this to Android:
https://github.com/project-chip/connectedhomeip/blob/35455fa577cf114b68016660118e3d408fca253c/src/controller/java/src/matter/onboardingpayload/QRCodeOnboardingPayloadParser.kt#L63
@stevePalmerin does this removed Java setuppayload support this optional field? https://github.com/project-chip/connectedhomeip/pull/27630/files?
| gharchive/issue | 2023-12-18T18:35:39 | 2025-04-01T04:35:34.681322 | {
"authors": [
"stevePalmerin",
"yunhanw-google"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/issues/31079",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
837650639 | Compile error when CHIP_BLE_END_POINT_DEBUG_LOGGING_ENABLED is set
Problem
When compiling with the CHIP_BLE_END_POINT_DEBUG_LOGGING_ENABLED enabled, I get compile errors in src/ble/BLEEndPoint.cpp
Proposed Solution
I propose to use the following patch:
diff --git a/src/ble/BLEEndPoint.cpp b/src/ble/BLEEndPoint.cpp
index aaf0d52c..ec95fc2c 100644
--- a/src/ble/BLEEndPoint.cpp
+++ b/src/ble/BLEEndPoint.cpp
@@ -632,12 +632,12 @@ void BLEEndPoint::QueueTx(PacketBufferHandle && data, PacketType_t type)
if (mSendQueue.IsNull())
{
mSendQueue = std::move(data);
- ChipLogDebugBleEndPoint(Ble, "%s: Set data as new mSendQueue %p, type %d", __FUNCTION__, mSendQueue, type);
+ ChipLogDebugBleEndPoint(Ble, "%s: Set data as new mSendQueue %p, type %d", __FUNCTION__, &mSendQueue, type);
}
else
{
mSendQueue->AddToEnd(std::move(data));
- ChipLogDebugBleEndPoint(Ble, "%s: Append data to mSendQueue %p, type %d", __FUNCTION__, mSendQueue, type);
+ ChipLogDebugBleEndPoint(Ble, "%s: Append data to mSendQueue %p, type %d", __FUNCTION__, &mSendQueue, type);
}
QueueTxUnlock();
@@ -993,7 +993,7 @@ BLE_ERROR BLEEndPoint::DriveSending()
{
#ifdef CHIP_BLE_END_POINT_DEBUG_LOGGING_ENABLED
if (mRemoteReceiveWindowSize <= BTP_WINDOW_NO_ACK_SEND_THRESHOLD &&
- !mTimerStateFlags.Has(TimerStateFlag::kSendAckTimerRunning) && mAckToSend == NULL)
+ !mTimerStateFlags.Has(TimerStateFlag::kSendAckTimerRunning) && mAckToSend.IsNull())
{
ChipLogDebugBleEndPoint(Ble, "NO SEND: receive window almost closed, and no ack to send");
}
Given this is an old issue, is this still present? If so, please re-open, thanks!
| gharchive/issue | 2021-03-22T12:04:01 | 2025-04-01T04:35:34.683529 | {
"authors": [
"tlykkeberg-grundfos",
"woody-apple"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/issues/5534",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1066521117 | Improve doxygen Docker Image
Problem
The Doxygen GitHub action is not using the connectedhomeip/chip-build-doxygen image and also executes a script which only requires two external tools (doxygen and graphviz).
Change overview
This change fixes the source code to pass the doxygen execution and includes a lighter Docker image (from 2.83GB to 41.9MB) for Doxygen CI execution.
Testing
These changes was tested using the act -j doxygen command
Fast tracking, given this is updating the docker image.
| gharchive/pull-request | 2021-11-29T21:26:49 | 2025-04-01T04:35:34.685968 | {
"authors": [
"electrocucaracha",
"woody-apple"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/pull/12337",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
659411128 | error: invalid conversion from 'const char*' to 'char*' [-fpermissive…
…] in screen-framework
Problem
While building the ESP32 demo on my Mac.
Adding @mlepage-google since I believe it may knows this code.
The casts are necessary to compile because the TFT functions incorrectly take non-const arguments. The correct fix of course is to fix the upstream TFT repo so it doesn't take non-const arguments. I have already done the correct fix, so this workaround should not be necessary. If you still cannot build, it's likely your sub-repo needs to be updated?
https://github.com/jeremyjh/ESP32_TFT_library/commit/f4e07be511d1bf1c54ae657d3248c352de6b93b1
| gharchive/pull-request | 2020-07-17T16:38:52 | 2025-04-01T04:35:34.688186 | {
"authors": [
"mlepage-google",
"vivien-apple"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/pull/1639",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1288120762 | [BL602] Enable command line to flash chip
Problem
What is being fixed? Examples:
Enable command line to flash chip
Change overview
Enable command line to flash chip
Testing
How was this tested? (at least one bullet point required)
command line:
cd third_party/bouffalolab/bl602_sdk/repo/tools/flash_tool
ubuntu:
./bflb_iot_tool-ubuntu18 --chipname=BL602 --baudrate=115200 --port=/dev/ttyACM0 --pt=chips/bl602/partition/partition_cfg_4M.toml --dts=chips/bl602/device_tree/bl_factory_params_IoTKitA_40M.dts --firmware=../../../../../../out/bl602-light/chip-bl602-lighting-example.bin
macos:
./bflb_iot_tool-macos --chipname=BL602 --baudrate=115200 --port=/dev/ttyACM0 --pt=chips/bl602/partition/partition_cfg_4M.toml --dts=chips/bl602/device_tree/bl_factory_params_IoTKitA_40M.dts --firmware=../../../../../../out/bl602-light/chip-bl602-lighting-example.bin
Fast tracking platform changes.
| gharchive/pull-request | 2022-06-29T03:20:09 | 2025-04-01T04:35:34.691835 | {
"authors": [
"jczhang777",
"woody-apple"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/pull/20094",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
721638397 | Drive the vscode esp32 wifi example through a separate script
Problem
The existing ESP32 script is the CI script, which has sideffects:
clobbers the sdkconfig defaults on every run (would lose things like wifi configuration)
builds two boards at once every time (extra work)
every bloard clean is after a sdk config reset, so build is not incremental
Summary of Changes
Created a python script that can drive esp32 builds (used python to have prettier argument parsing, although in hindsight it required quite a few workarounds regarding environment)
Changed tasks.json to reference this script instead.
@mspang ? I believe there may be better ways of doing this (calling shell from python, knowing that the shell will call python in turn is odd), however maybe we could incrementally improve. For now this works better than before (is incremental, builds only what is needed).
We could also convert it to just shell, for a cleaner feeling (but without as clean argument parsing).
Either way, I would propose to do improvements in future PRs.
@mspang ? I believe there may be better ways of doing this (calling shell from python, knowing that the shell will call python in turn is odd), however maybe we could incrementally improve. For now this works better than before (is incremental, builds only what is needed).
We could also convert it to just shell, for a cleaner feeling (but without as clean argument parsing).
Either way, I would propose to do improvements in future PRs.
This looks very awkward to me, couldn't you split this two and have a activate_and_run_command.sh wrapper?
| gharchive/pull-request | 2020-10-14T17:14:50 | 2025-04-01T04:35:34.695776 | {
"authors": [
"andy31415",
"mspang"
],
"repo": "project-chip/connectedhomeip",
"url": "https://github.com/project-chip/connectedhomeip/pull/3258",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1097645380 | Dependency package not found
I am trying to get the tool running by following the instructions the the main README.md and docs/instructions.md. In the root directory of the project I have tried executing npm ci and npm install, following from the documentation files listed above. Both of these fail to install the dependency https://nexus.tecnotree.com/repository/npm-public/listr2/-/listr2-3.13.5.tgz. Failing with the error
npm ERR! code ENOTFOUND
npm ERR! syscall getaddrinfo
npm ERR! errno ENOTFOUND
npm ERR! network request to https://nexus.tecnotree.com/repository/npm-public/listr2/-/listr2-3.13.5.tgz failed, reason: getaddrinfo ENOTFOUND nexus.tecnotree.com
npm ERR! network This is a problem related to network connectivity.
npm ERR! network In most cases you are behind a proxy or have bad network settings.
npm ERR! network
npm ERR! network If you are behind a proxy, please make sure that the
npm ERR! network 'proxy' config is set properly. See: 'npm help config'
I am running node version v16.13.1 and npm version 8.3.0.
I have made sure that I am not behind a proxy and ran the following npm configuration commands
npm config rm proxy
npm config rm https-proxy
npm config --global rm proxy
npm config --global rm https-proxy
This was fixed. There was some "tecnotree" nonsense that weasled it's way into the package.json at some point.
| gharchive/issue | 2022-01-10T09:23:25 | 2025-04-01T04:35:34.698307 | {
"authors": [
"hicklin",
"tecimovic"
],
"repo": "project-chip/zap",
"url": "https://github.com/project-chip/zap/issues/361",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2352519037 | feat(erofs): initial commit for erofs support
Fixes https://github.com/opencontainers/image-spec/issues/1190
What type of PR is this?
Which issue does this PR fix:
What does this PR do / Why do we need it:
If an issue # is not available please add repro steps and logs showing the issue:
Testing done on this change:
Automation added to e2e:
Will this break upgrades or downgrades?
Does this PR introduce any user-facing change?:
By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Hi @rchincha, I'm currently working on incremental builds and I plan to release erofs-utils v1.8 this month (with multi-threaded mkfs, incremental builds, Intel QPL, etc.), but anyway, I'd suggest use cgo for initial EROFS support (with formal liberofs APIs exported) for now (or even binary integration) since this go implemention seems somewhat incomplete (and maybe even broken.. I don't have enough time to look into that since other prioritied stuffs are on hands..)
@hsiangkao I suspect that cgo path may be what ends up happening. Currently, just prototyping quickly to understand the interfaces and scope of changes/work.
One thing I might need to mention here is that erofs supports external blobs or chunks since Linux 5.16 compared to Squashfs because Dragonfly Nydus once asked this feature to deduplicate data in chunks among different container images.
I'm not sure if it's worthwhile to highlight this, you could just make a tiny metadata with external blobs (which can be used for multiple images) for reference.
Currently only Nydus has userspace tools to generate chunk blobs, if that is interested in other use cases, I could seek more time to implement this in mkfs.erofs too.
$ cat stacker.yaml
build:
from:
type: docker
url: docker://public.ecr.aws/docker/library/busybox:1.37-glibc
$ /tmp/stacker build -layer-type erofs
Creating new OCI Layout at "/data/hdd/rchincha/tmp/lxc/oci"
preparing image build...
imported file hashes (after substitutions):
loading docker://public.ecr.aws/docker/library/busybox:1.37-glibc
Copying blob a46fbb00284b done
Copying config 27a71e19c9 done
Writing manifest to image destination
Storing signatures
mkfs.erofs 1.4 c_version: [ 1.4]
c_dbg_lvl: [ 2]
c_dry_run: [ 0]
filesystem build built successfully
@hsiangkao
Is there an equivalent of "unsquashfs" in erofs-utils?
https://manpages.debian.org/testing/squashfs-tools/unsquashfs.1.en.html
SYNOPSIS
unsquashfs [OPTIONS] FILESYSTEM [files to extract or exclude (with -excludes) or cat (with -cat )]
Hi!
@hsiangkao
Is there an equivalent of "unsquashfs" in erofs-utils?
https://manpages.debian.org/testing/squashfs-tools/unsquashfs.1.en.html
SYNOPSIS
unsquashfs [OPTIONS] FILESYSTEM [files to extract or exclude (with -excludes) or cat (with -cat )]
fsck.erofs --extract= FILESYSTEM to extract the whole filesystem...
Currently, there is no exclude way but it can be added later.
| gharchive/pull-request | 2024-06-14T05:08:51 | 2025-04-01T04:35:34.750706 | {
"authors": [
"hsiangkao",
"rchincha"
],
"repo": "project-stacker/stacker",
"url": "https://github.com/project-stacker/stacker/pull/626",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
95047829 | atomic-reactor fails on RHEL7 because of python-requests version
The version of python-requests that is in RHEL7 (core OS package set), and derivative distros (CentOS, et al), is too old.
I built an rpm from my git checkout with tito build --test --builder mock --arg mock=epel-7-x86_64 --rpm and then installed the resulting rpm. I get the following error:
[root@losbs ~]# atomic-reactor create-build-image --reactor-tarball-path /usr/share/atomic-reactor/atomic-reactor.tar.gz /usr/share/atomic-reactor/images/dockerhost-builder/ buildroot
Traceback (most recent call last):
File "/usr/bin/atomic-reactor", line 5, in <module>
from pkg_resources import load_entry_point
File "/usr/lib/python2.7/site-packages/pkg_resources.py", line 3011, in <module>
parse_requirements(__requires__), Environment()
File "/usr/lib/python2.7/site-packages/pkg_resources.py", line 626, in resolve
raise DistributionNotFound(req)
pkg_resources.DistributionNotFound: requests>=2.2.1,<2.5.0
[root@losbs ~]# rpm -q python-requests
python-requests-1.1.0-8.el7.noarch
this is actually issue of docker-py; therefore we build python-requests and supply it from our own repo
Yeah, I'm doing this currently with my COPR space but this isn't something we can to in EPEL because it will violate the policy on not overriding core RHEL packages. Is the plan to maintain this separately? I fear for what might potentially be broken by replacing a core RHEL package.
I've discussed this with docker-py RHEL maintainer and am not sure what was the final statement. CC @lsm5 @arrfab
CentOS SIGs can and do override base packages, I can add a new python-requests along with docker-py and atomic-reactor to the Virt SIG repositories. @maxamillion and @TomasTomecek would you like to be a part of CentOS virt SIG?
RE: RHEL, @TomasTomecek @maxamillion I did patch docker-py earlier so that atomic could run well with an older python-requests. Looks like this would need another patch :\ . I vaguely remember a bug filed to update python-requests on RHEL 7, but can't find it atm :(
@lsm5 I'm actually planning to patch python-docker-py in Fedora to don't depend pretty much on any version of python-requests: https://github.com/projectatomic/atomic-reactor/issues/226
@TomasTomecek please use @rhatdan's repo instead of original upstream moving forward.
@lsm5 because of https://github.com/rhatdan/docker-py/commit/0fc135c6dbe994aae1530e96071ea9aa33d94c88?
python-requests got an update in RHEL7 official to 2.6.0 -> python-requests-2.6.0-1.el7_1.noarch so now atomic-reactor is erroring the same way Fedora is in https://github.com/projectatomic/atomic-reactor/issues/226
@maxamillion I already reported this at https://bugzilla.redhat.com/show_bug.cgi?id=1249651 and am waiting for a fix there. Feel free to add yourself to CC of that bug to know when this gets fixed.
BTW I believe the correct fix is pinning the dependencies in the specfile precisely the same way they are pinned in upstream. Then the package will fail to install if dependencies change (and even some compose tools may throw errors about unsatisfied dependencies). When the versions are pinned, then it's only a matter of patching upstream requirements (and the code) for dependencies that have different version in RHEL7/EPEL7.
@bkabrda if i'm reading above comment right, if we have python-requests >= 2.5.3 in the python-docker-py specfile, it will make python-docker-py uninstallable on rhel7 which is just not an option, as there's atomic depending on python-docker-py
can't we just upgrade python-requests on rhel7 and move on?
ugh ..re-reading ...so we gotta have python-requests == 2.5.3, hmm :( ...will check on that :( :(
@lsm5 the point is, that to do packaging correctly, you should do it this way:
pin dependencies in the specfile to be precisely the same as dependencies in requirements.txt
try to install the package
see what breaks
patch both requirements and the actual code to work with the version that is in RHEL 7
profit :)
this appears to have been resolved, can others verify? if so, I'm good with closing
Seems to be fixed. I can install python-docker-py from rhel7-server-extras-rpms and atomic-reactor from EPEL without problems.
| gharchive/issue | 2015-07-14T21:37:26 | 2025-04-01T04:35:34.824743 | {
"authors": [
"TomasTomecek",
"bkabrda",
"lsm5",
"maxamillion",
"mmilata"
],
"repo": "projectatomic/atomic-reactor",
"url": "https://github.com/projectatomic/atomic-reactor/issues/225",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
255640385 | Avoid traceback when pulp_pull fails
Signed-off-by: Tim Waugh twaugh@redhat.com
Release note:
Bugfix:
** failure of the pulp_pull plugin no longer causes the store_metadata_in_osv3 plugin to fail
| gharchive/pull-request | 2017-09-06T15:12:39 | 2025-04-01T04:35:34.826524 | {
"authors": [
"twaugh"
],
"repo": "projectatomic/atomic-reactor",
"url": "https://github.com/projectatomic/atomic-reactor/pull/817",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
190162644 | We want to execute docker-storage-setup on the host system
If docker-storage-setup is run from a Super Privileged Container (SPC)
we want to execute the script on the host machine.
$HOST indicates the mountpoint of the host root fs of the container.
This patch will chroot to this rootfs and execute the script.
@rhvgoyal PTAL
This assumes that all the /usr/lib/docker-storage-setup files are installed on host. That means docker-storage-setup and all dependencies are installed on host. But AFAIK, we are not doing any of that?
Yes you are right. I thought the script was all encompassing.
Another version, this one just adds the $HOST/usr/bin and $HOST/usr/sbin to $PATH.
Also changes a couple of other small parts of script. Probably need to look at other helper apps.
We have decided not to do this for now, closing.
| gharchive/pull-request | 2016-11-17T21:00:10 | 2025-04-01T04:35:34.832233 | {
"authors": [
"rhatdan",
"rhvgoyal"
],
"repo": "projectatomic/docker-storage-setup",
"url": "https://github.com/projectatomic/docker-storage-setup/pull/173",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
86848529 | docker-storage-setup.1: fix typos
Signed-off-by: Alex Jia ajia@redhat.com
:+1:
| gharchive/pull-request | 2015-06-10T05:40:34 | 2025-04-01T04:35:34.833264 | {
"authors": [
"cgwalters",
"chuanchang"
],
"repo": "projectatomic/docker-storage-setup",
"url": "https://github.com/projectatomic/docker-storage-setup/pull/41",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2726912147 | V0.0.18
added ERC721 deployable and smart character token address to the artifacts as part of the deployer image output
agreed! from v2 it can be written directly to json file from the scripts instead of console.logs
| gharchive/pull-request | 2024-12-09T12:31:38 | 2025-04-01T04:35:34.835132 | {
"authors": [
"0xxlegolas"
],
"repo": "projectawakening/world-chain-contracts",
"url": "https://github.com/projectawakening/world-chain-contracts/pull/340",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1211855459 | Is the -root-tld flag still working ?
./interactsh-server -domain example.com -hostmaster admin@example.com -ip X.X.X.X -listen-ip X.X.X.X -root-tld
flag provided but not defined: -root-tld
https://github.com/projectdiscovery/interactsh/issues/17
Is it now enabled by default ?
@greckko flag name is updated with below:
-wc, -wildcard enable wildcard interaction for interactsh domain (authenticated)
| gharchive/issue | 2022-04-22T05:54:51 | 2025-04-01T04:35:34.842649 | {
"authors": [
"ehsandeep",
"greckko"
],
"repo": "projectdiscovery/interactsh",
"url": "https://github.com/projectdiscovery/interactsh/issues/268",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
788500898 | resolves infinite loop when received value was zero
this commit solves the problem in projectdiscovery/naabu#123
hi, is there any plan to release a new version to fix this problems?
@LazyMaple the latest release is after this commit - are you still having this issue?
| gharchive/pull-request | 2021-01-18T19:27:58 | 2025-04-01T04:35:34.843965 | {
"authors": [
"LazyMaple",
"nicolascb",
"olearycrew"
],
"repo": "projectdiscovery/ipranger",
"url": "https://github.com/projectdiscovery/ipranger/pull/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.