id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
222426441
systemd error: Default-Start contains no runlevels, aborting. I've a Ubuntu 16.04 with systemd. The package has this in the recipe: poise_service "node_exporter" do command "/some/command" user user end But I get this error when deploying: update-rc.d: error: node_exporter Default-Start contains no runlevels, aborting. Did I miss anything in the recipe? The same recipe works fine on a Ubuntu 14.04 system. Thanks! poise-service 1.4.2 according to my Berksfile.lock Solved, ugh. There were old files in /etc/sv/, /etc/service/ and /etc/init.d/ from a previous version. Removing those solved the problem. Thanks :)
gharchive/issue
2017-04-18T13:59:29
2025-04-01T06:45:24.964954
{ "authors": [ "alicebob" ], "repo": "poise/poise-service", "url": "https://github.com/poise/poise-service/issues/43", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2404303651
🛑 Pokko - au-syd1 is down In ba8c1b9, Pokko - au-syd1 (https://au-syd1.pokko.io) was down: HTTP code: 0 Response time: 0 ms Resolved: Pokko - au-syd1 is back up in f08167f after 5 minutes.
gharchive/issue
2024-07-11T22:51:07
2025-04-01T06:45:24.992241
{ "authors": [ "brendanmckenzie" ], "repo": "pokkocms/pokko-status", "url": "https://github.com/pokkocms/pokko-status/issues/122", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1332229550
Update README.md Updates to deps REPLACE_ME: Please make sure to update the Assigness, Labels, Projects, Milestone before submitting the PR for review Description REPLACE_ME: Summary of the change. Include motivation, reasons, context, dependencies, etc... Fixes issue/<ISSUE_NUMBER> Type of change Please mark the options that are relevant. [ ] New feature (non-breaking change which adds functionality) [ ] Bug fix (non-breaking change which fixes an issue) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Documentation [ ] Other (<REPLACE_ME_WITH_DETAILS>) How Has This Been Tested? [Not needed for integration PR] _REPLACE_ME: Describe the tests and that you ran to verify your changes. If applicable, provide steps to reproduce. Bonus points for images and videos or gifs. [ ] make test_all [ ] LocalNet Checklist [ ] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have tested my changes using the available tooling [ ] If applicable, I have made corresponding changes to related or global README [ ] If applicable, I have added added new diagrams using mermaid.js [ ] If applicable, I have added tests that prove my fix is effective or that my feature works @phthan0 Please see & update all the placeholder text in the github PR template. Also, can you go through all the steps in the dev guide and submit one PR with your improvements to run this on linux?
gharchive/pull-request
2022-08-08T18:21:32
2025-04-01T06:45:24.999968
{ "authors": [ "Olshansk", "phthan0" ], "repo": "pokt-network/pocket", "url": "https://github.com/pokt-network/pocket/pull/155", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
400029750
Spike detection part of newest version? Hello! Someone in my lab came across this link https://github.com/poldracklab/fmriprep/issues/699 Which appeared to indicate that spike detection was being applied at the subject level. But when we went through the documentation we didn't see anything indicating this had been implemented. I wanted to check in to see if spike detection is part of the current fmriprep pipeline and if so, is it being applied to raw images or later in the pipeline? Thanks! Please follow up in the link above.
gharchive/issue
2019-01-16T22:38:37
2025-04-01T06:45:25.061034
{ "authors": [ "oesteban", "zanderso13" ], "repo": "poldracklab/fmriprep", "url": "https://github.com/poldracklab/fmriprep/issues/1471", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
609411740
ENH: Use new DerivativesDataSink from NiWorkflows 1.2.0 This PR integrates the new API that will be released with 1.2.0. References: nipreps/niworkflows#507 Codecov Report Merging #183 into master will decrease coverage by 46.20%. The diff coverage is 6.00%. @@ Coverage Diff @@ ## master #183 +/- ## =========================================== - Coverage 71.72% 25.52% -46.21% =========================================== Files 15 15 Lines 955 995 +40 Branches 134 135 +1 =========================================== - Hits 685 254 -431 - Misses 224 731 +507 + Partials 46 10 -36 Flag Coverage Δ #ds005 ? #ds054 ? Impacted Files Coverage Δ smriprep/workflows/base.py 29.50% <ø> (-50.82%) :arrow_down: smriprep/workflows/anatomical.py 16.32% <4.54%> (-51.27%) :arrow_down: smriprep/workflows/outputs.py 15.23% <16.66%> (-55.92%) :arrow_down: smriprep/cli/run.py 2.85% <0.00%> (-74.77%) :arrow_down: smriprep/workflows/norm.py 30.76% <0.00%> (-69.24%) :arrow_down: smriprep/__main__.py 16.66% <0.00%> (-66.67%) :arrow_down: smriprep/workflows/surfaces.py 14.44% <0.00%> (-64.45%) :arrow_down: smriprep/interfaces/freesurfer.py 19.69% <0.00%> (-40.91%) :arrow_down: ... and 3 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 07e19f1...d70010f. Read the comment docs. Locally checking the correct indexing of probseg and dseg results, w/wo FreeSurfer. Once that is done, this is all yours, @mgxd.
gharchive/pull-request
2020-04-29T22:46:30
2025-04-01T06:45:25.075739
{ "authors": [ "codecov-io", "oesteban" ], "repo": "poldracklab/smriprep", "url": "https://github.com/poldracklab/smriprep/pull/183", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
892699793
Add type annotations? Comes from #1223: Although I have to admit I am a big fan of type annotations to empower mypy to detect as much problems as possible. Yes! I'm a fan as well, but I have concerns about the verbosity of the syntax, the fact that MyPy has some limitations when working with stubs that are either very difficult to fix or not a priority for the developers (https://github.com/python/mypy/issues/5028), the changes to come in Python 3.10, and so forth. Python type annotations and gradual typing have evolved in a very organic way: first annotations were introduced without any particular intent, then they were made for gradual typing, then they just simplified the syntax... Even though I am normally happy with living in the bleeding edge in terms of libraries and such, for type annotations I have some reservations. That's why I didn't jump immediately and said "let's merge this". Besides, there's the question of annotating all the library to be consistent, and finally we have another problem which is the interaction with astropy.units. As a result, before starting the path of annotating poliastro, I would like to see some discussion, and a way forward. For example: should we start by slowly making MyPy more strict, and adding types more gradually? This is the "full MyPy strict mode" I've been using in other projects: [mypy] files = src/**/*.py, tests/**/*.py warn_redundant_casts = True warn_unused_configs = True pretty = True show_error_codes = True disallow_any_generics = True disallow_subclassing_any = True disallow_untyped_calls = True disallow_incomplete_defs = True check_untyped_defs = True disallow_untyped_decorators = True no_implicit_optional = True warn_unused_ignores = True warn_return_any = True no_implicit_reexport = True # More strict checks for library code [mypy-poliastro] disallow_untyped_defs = True which is very far from what we have now: https://github.com/poliastro/poliastro/blob/5ee92de8b620efbf215d3fca1ba248beda41428f/mypy.ini#L1-L4 I think we could start by enabling some of these configs (the easiest ones), doing small tweaks, and then seeing where is the "inflection point" where we need to annotate large chunks of the code. And one last note: did the recent changes in Astropy made the interaction between MyPy and quantity_input easier, or more robust, or more appealing? Do we need to wait until Astropy 4.3 for any particular functionality? Thanks @astrojuanlu for opening this issue. The verbosity of type annotations, as you said, is definitely an issue although overall I have found adding type annotations quite useful. In case you would like to go for adding type annotations, one proposal would be to follow a guideline similar to Pandas (https://pandas.pydata.org/pandas-docs/stable/development/contributing.html) were the new code is required to have type annotations, so they are adding type annotations incrementally instead of going for a big refactoring. With respect to the quantity_input from astropy, I would recommend to use it whenever the method inputs are unit types. For those methods, if there are parameters that are not unit types we could also try to annotate them although I have found that sometimes quantity_input has conflicts with type annotations. My proposal is that in case of conflict the use of quantity_input trumps general type annotations. I also think that it would be a good idea to go step by step towards a mypy configuration like the one you proposed. What do you think? A little side-note: If you add type annotations, you may also consider using typeguard for run-time type checking. It's slow like hell, but it is absolutely useful for testing. For production, you can always deactivate it by running Python in optimized mode. Run-time type checking actually goes a few steps further than what static code analysis can provide and show you even more bugs (than you ever wanted to find). I have yet one more gripe with annotations, and is that they are still in flux. In fact, they had to postpone a change from Python 3.10 to Python 3.11 because otherwise lots of web stuff would break https://mail.python.org/archives/list/python-dev@python.org/message/CLVXXPQ2T2LQ5MP2Y53VVQFCXYWQJHKZ/ . In fact, our own czml3 will be broken by this. Therefore, with all the caveats above, and taking a (perhaps overly conservative) stance, my vote for type annotations is "not yet". Speaking as someone who even runs Linux LTS distributions beyond their due date, I am always in favor of a very stable, conservative approach ;) Or I am simply lacking the time to upgrade ... Besides, I agree: This stuff being in flux is annoying and makes it unreliable. Just the fact that you had to import a special alias for lists, i.e. List, so you could write List[int] and now you're supposed to write list[int] all the sudden (in 3.9 or 3.10?), which is actually confusing in multiple ways, drives me nuts. Having said that, the question is what you want to use type annotations for and how much you want to rely on them. Here is where the pros come in: Amazing support in IDEs, i.e. way faster development and fewer bugs. Besides, both static type checking and run time checks actually provide an amazing amount of insight into your code while testing and debugging. But it's like with coverage: Is it worth it to have 100% coverage at all cost? Or is it merely a tool to be on the safe side for most of your critical code? Besides, how often do you run coverage analysis in production setups? I'd say do not rely on every last detail of the spec and use it with care and in the right quantity. Then it actually makes sense (although it is still in flux). I have gone through this process and ended up trying it in a project of mine. It is worth it. Have I mentioned that Sphinx automatically picks up the notations and uses them when generating API docs? Less complex/confusing doc strings, actually more readable code, ... TIL: https://pypi.org/project/typing-copilot/
gharchive/issue
2021-05-16T15:54:33
2025-04-01T06:45:25.087876
{ "authors": [ "astrojuanlu", "ismaelJimenez", "s-m-e" ], "repo": "poliastro/poliastro", "url": "https://github.com/poliastro/poliastro/issues/1230", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
301946129
Plotting segments of Orbits in 2D as well Refactored the OrbitPlotter . Now it can plot segments of Orbits in 2D as well due to the plot_trajectory method. Addresses Issue #274 Awesome! Can you add some screenshot of how the plot_trajectory method looks like? For instance, plot some Hohmann transfer with only the traversed parts of the orbit. Sure! I will do that in night @Juanlu001 , Do you have some data about the hohmann transfer? Do you have any idea on how to plot this? I'm being able to plot in your branch without problems: Using this notebook: https://gist.github.com/Juanlu001/e574886ccf8bff10c8c08459981a674f Please review the reasons why it does not work for you. Leaving some extra comments. Rebased! :) This is looking better :) Now we can improve notebooks like this one without cluttering the final plot: http://docs.poliastro.space/en/latest/examples/Going to Jupiter with Python using Jupyter and poliastro.html Merging! True:)
gharchive/pull-request
2018-03-03T00:18:57
2025-04-01T06:45:25.093452
{ "authors": [ "Juanlu001", "shreyasbapat" ], "repo": "poliastro/poliastro", "url": "https://github.com/poliastro/poliastro/pull/332", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
974658935
Error on API initialization An attempt to initialize the API produces the attached errors. Initialization code: val wsProvider = WsProvider("wss://rpc.polkadot.io/") val ready: Promise<ApiPromise> = ApiPromise.create(wsProvider) Is there any way to get it working? 13:42:56.234 [WebSocketConnectReadThread-139] INFO o.p.rpc.provider.ws.WsProvider - WebSocket onOpen: wss://rpc.polkadot.io/ java.lang.reflect.InvocationTargetException at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.polkadot.types.TypesUtils$1.newInstance(TypesUtils.java:103) at org.polkadot.types.codec.CodecUtils.decodeU8a(CodecUtils.java:38) at org.polkadot.types.codec.CodecUtils.decodeU8a(CodecUtils.java:44) at org.polkadot.types.codec.CodecUtils.decodeU8a(CodecUtils.java:25) at org.polkadot.types.codec.Struct.decodeStruct(Struct.java:69) at org.polkadot.types.codec.Struct.<init>(Struct.java:37) at org.polkadot.types.codec.Struct.<init>(Struct.java:46) at org.polkadot.types.metadata.MetadataVersioned.<init>(MetadataVersioned.java:79) at org.polkadot.types.metadata.Metadata.decodeMetadata(Metadata.java:29) at org.polkadot.types.metadata.Metadata.<init>(Metadata.java:17) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.polkadot.types.TypesUtils$1.newInstance(TypesUtils.java:103) at org.polkadot.types.codec.CreateType.createType(CreateType.java:237) at org.polkadot.rpc.core.RpcCore.formatOutput(RpcCore.java:176) at org.polkadot.rpc.core.RpcCore.access$200(RpcCore.java:37) at org.polkadot.rpc.core.RpcCore$1.lambda$invoke$0(RpcCore.java:122) at com.onehilltech.promises.OnResolvedExecutor.execute(OnResolvedExecutor.java:45) at com.onehilltech.promises.OnResolvedExecutor.lambda$execute$0(OnResolvedExecutor.java:32) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.IndexOutOfBoundsException: Index 13 out of bounds for length 4 at java.base/jdk.internal.util.Preconditions.outOfBounds(Preconditions.java:64) at java.base/jdk.internal.util.Preconditions.outOfBoundsCheckIndex(Preconditions.java:70) at java.base/jdk.internal.util.Preconditions.checkIndex(Preconditions.java:248) at java.base/java.util.Objects.checkIndex(Objects.java:372) at java.base/java.util.ArrayList.get(ArrayList.java:459) at org.polkadot.types.codec.EnumType.createValue(EnumType.java:130) at org.polkadot.types.codec.EnumType.decodeViaValue(EnumType.java:88) at org.polkadot.types.codec.EnumType.decodeEnumType(EnumType.java:80) at org.polkadot.types.codec.EnumType.<init>(EnumType.java:37) at org.polkadot.types.metadata.MetadataVersioned$MetadataEnum.<init>(MetadataVersioned.java:25) ... 28 more java.lang.NullPointerException at org.polkadot.types.codec.CodecUtils.decodeU8a(CodecUtils.java:42) at org.polkadot.types.codec.CodecUtils.decodeU8a(CodecUtils.java:44) at org.polkadot.types.codec.CodecUtils.decodeU8a(CodecUtils.java:25) at org.polkadot.types.codec.Struct.decodeStruct(Struct.java:69) at org.polkadot.types.codec.Struct.<init>(Struct.java:37) at org.polkadot.types.codec.Struct.<init>(Struct.java:46) at org.polkadot.types.metadata.MetadataVersioned.<init>(MetadataVersioned.java:79) at org.polkadot.types.metadata.Metadata.decodeMetadata(Metadata.java:29) at org.polkadot.types.metadata.Metadata.<init>(Metadata.java:17) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.polkadot.types.TypesUtils$1.newInstance(TypesUtils.java:103) at org.polkadot.types.codec.CreateType.createType(CreateType.java:237) at org.polkadot.rpc.core.RpcCore.formatOutput(RpcCore.java:176) at org.polkadot.rpc.core.RpcCore.access$200(RpcCore.java:37) at org.polkadot.rpc.core.RpcCore$1.lambda$invoke$0(RpcCore.java:122) at com.onehilltech.promises.OnResolvedExecutor.execute(OnResolvedExecutor.java:45) at com.onehilltech.promises.OnResolvedExecutor.lambda$execute$0(OnResolvedExecutor.java:32) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) java.lang.reflect.InvocationTargetException at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.polkadot.types.TypesUtils$1.newInstance(TypesUtils.java:103) at org.polkadot.types.codec.CreateType.createType(CreateType.java:237) at org.polkadot.rpc.core.RpcCore.formatOutput(RpcCore.java:176) at org.polkadot.rpc.core.RpcCore.access$200(RpcCore.java:37) at org.polkadot.rpc.core.RpcCore$1.lambda$invoke$0(RpcCore.java:122) at com.onehilltech.promises.OnResolvedExecutor.execute(OnResolvedExecutor.java:45) at com.onehilltech.promises.OnResolvedExecutor.lambda$execute$0(OnResolvedExecutor.java:32) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.NullPointerException at org.polkadot.types.metadata.Metadata.decodeMetadata(Metadata.java:32) at org.polkadot.types.metadata.Metadata.<init>(Metadata.java:17) ... 14 more error.log Please include log between 2 lines for code formating "```" like this is log example Please see above I see, Is there any way to get it working? I wonder the same. As in mai example https://github.com/polkadot-java/api/blob/master/examples/src/main/java/org/polkadot/example/Staking.java#L68-L70 WsProvider wsProvider = new WsProvider(endPoint); Promise<ApiPromise> ready = ApiPromise.create(wsProvider); Well, I bet it will not run on newest v3.0 Substrate, see https://github.com/polkadot-java/api/issues/34 @mnowrot I wonder if you had to dance with gradle build, as I had to , see #35 Would you care to give a try? See also https://github.com/polkadot-java/api/issues/34#issuecomment-914217953 @paulvi Thank you for your message. I've tried both fixing the build and polkaj, and finally settled for the TypeScript Polkadot API as polkaj implements only a small part of that API OK, there are 3 APIs actually from ParityTech see https://wiki.polkadot.network/docs/build-node-interaction Substrate API Sidecar is actually using Polkadot-JS RPC API, so is actually usage example. BTW, wss:// and not plain ws:// may be the issue, I have not noticed yet TLS related stuff, and could make TCP connect to ws://
gharchive/issue
2021-08-19T13:21:55
2025-04-01T06:45:25.117042
{ "authors": [ "mnowrot", "paulvi" ], "repo": "polkadot-java/api", "url": "https://github.com/polkadot-java/api/issues/33", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
305346472
Support audio sources to control microphone or line-in I searched for a while about microphone support and can't found anything, exist any way to connect the volume or pulseaudio internal modules to sources instead sinks? Or maybe in future the project support that? Hi, Are there any hacks to enable microphone in polybar ? An hacky option is to parse the pacmd output pacmd list-sources | grep -e '* index:' -A 11 | grep -Po '(?(?<=muted: )(.*))' # no if not muted Otherwise you have to use the PulseAudio API. I created a binary to get the muted state of the microphone joshuachp/pulseaudio-mic-state This can be used to create a script module [module/microphone] type = custom/script ; Available tokens: ; %counter% ; Command to be executed (using "/bin/sh -c [command]") exec = [[ "$(pulseaudio-mic-state)" == 'true' ]] && echo '<icon-muted>' || echo '<icon-unmuted>' This might help for the time being: https://github.com/MarcDonald/polybar-mic-volume This might help for the time being: https://github.com/MarcDonald/polybar-mic-volume Related PR: #2238 Does anyone known why this #2238 PR is not yet merged? That would be a great addition as many of us want to show the status of the microphone or any input device currently in use. I have opened this #3062 discussion to start a conversation about it. If helps, there's this module. I made a hybrid between the default configuration and this one for output, but I don't know if it's possible for input (basically because I'm newbie so still don't know the basics, I'm still learning how this works): [module/pulseaudio] type = internal/pulseaudio ; format-volume-prefix = "VOL " format-volume-prefix = " " format-volume-prefix-foreground = ${colors.primary} format-volume = <label-volume> label-volume = %percentage%% label-muted = 󰖁 label-muted-foreground = ${colors.disabled}
gharchive/issue
2018-03-14T22:07:39
2025-04-01T06:45:25.124154
{ "authors": [ "Jaffex", "j0n4t", "joshuachp", "marcovergueira", "moshpirit", "rjshrjndrn", "tzeikob" ], "repo": "polybar/polybar", "url": "https://github.com/polybar/polybar/issues/1096", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
432766630
Alvr crash I just attempted to use the newest version of alvr, first time using this program so I might have missed something but I get thus error when I press connect to headset: Unhandled Exception.ExceptionCode=C0000005Address=000007FED88AA036(000007FED8890000 + 000000000001A036) I updated drivers, set the video res and encoder to AVC and tried disabling streaming sound. I also noticed steamVR gets a 301 error when alvr gets its' error and weirdly enough it doesnt crash I just hit okay and the server still runs but I get the error again when I hit connect. I also tried reinstalling steamvr but no luck. Where would I find the error log if I need to attach that here? I'm getting the same error, ran fine the first time but came back a week later and I'm now seeing this error Can you provide crash log file? The file is located on C:\Program Files\ALVR\logs. I updated my graphics driver via Device Manager and the issue is now resolved
gharchive/issue
2019-04-12T22:23:14
2025-04-01T06:45:25.151564
{ "authors": [ "Nataliecatchat", "korncob", "polygraphene" ], "repo": "polygraphene/ALVR", "url": "https://github.com/polygraphene/ALVR/issues/282", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1154090453
Adding description to one manifest file Change Adding description fields to json manifest schema files. Testing yarn build & yarn test is happy. yarn lint is happy. Future changes Update description for metadata manifest schema. Hi @sarath-hotspot, can you please add descriptions for all manifest types & versions? Hi @sarath-hotspot, can you please add descriptions for all manifest types & versions? Updated.
gharchive/pull-request
2022-02-28T13:24:55
2025-04-01T06:45:25.163353
{ "authors": [ "dOrgJelli", "sarath-hotspot" ], "repo": "polywrap/monorepo", "url": "https://github.com/polywrap/monorepo/pull/698", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
249602620
population phenotypes showing in cell phenotype section The population phenotypes are also showing in the cell phenotype section It looks like the problem is that I recently added FYPO:0000300 (biological process phenotype) to the list of terms to include in the "Cell phenotype" section. But those "sensitive to ..." terms are descendants of FYPO:0000300 and are also descendants of 0000003 (cell population phenotype). We added FYPO:0000300 because there were some terms that weren't in either list on the gene pages: https://github.com/pombase/website/issues/471#issuecomment-320228727 I'll take FYPO:0000300 out of the config. The only term that will go missing from the gene page when I do that is FYPO:0000680 ("abnormal ascus development"). Maybe an ontology change can fix that? After that change the populate phenotype section will include all descendants of: FYPO:0000003 and FYPO:0001155 The cell phenotype section will include descendants of: FYPO:0000002, FYPO:0000652, FYPO:0000300, FYPO:0000301, FYPO:0001312, FYPO:0002928, FYPO:0002931, FYPO:0003093, FYPO:0003324, FYPO:0003427, FYPO:0003480, FYPO:0004914, FYPO:0004915, FYPO:0004942, FYPO:0005157 Sorry about the long list. It was needed to cover everything. It's looking much better now. You'll need to shift-reload to see the change. The only term that will go missing from the gene page when I do that is FYPO:0000680 ("abnormal ascus development"). Maybe an ontology change can fix that? Sorry, FYPO:0000680 wouldn't fit biologically under any of the other high-level terms configured to show up in the cell phenotype section. Better to add it itself. Also, is there a regular(ish) check for terms that are used in annotations but wouldn't show up on pages? I noticed that FYPO:0002930 isn't in the configuration list - it doesn't have annotations at the moment but we may use it somewhere along the line, and when we do we would want it displayed. Sorry, FYPO:0000680 wouldn't fit biologically under any of the other high-level terms configured to show up in the cell phenotype section. Better to add it itself. OK, I've added that and I'll rerun things soon. Also, is there a regular(ish) check for terms that are used in annotations but wouldn't show up on pages? It's a manual check at the moment. If you think it's worthwhile checking on every load, could you add an issue about it? I noticed that FYPO:0002930 isn't in the configuration list FYPO:0002929 is slightly more general. Would it be better to add that to the config? Also, is there a regular(ish) check for terms that are used in annotations but wouldn't show up on pages? It's a manual check at the moment. If you think it's worthwhile checking on every load, could you add an issue about it? Every load feel like overkill, but I'll open a ticket anyway. Chado tracker (for logs)? I noticed that FYPO:0002930 isn't in the configuration list FYPO:0002929 is slightly more general. Would it be better to add that to the config? ooh, yes, well spotted Chado tracker (for logs)? Yes please. ooh, yes, well spotted OK, I'll add FYPO:0002929 and re-run. https://github.com/pombase/pombase-chado/issues/641 now open for business I'll add FYPO:0002929 and re-run. I've done that. The one annotation to "abnormal ascus development" will pop back in the preview next time I re-release.
gharchive/issue
2017-08-11T11:02:11
2025-04-01T06:45:25.172706
{ "authors": [ "Antonialock", "kimrutherford", "mah11" ], "repo": "pombase/website", "url": "https://github.com/pombase/website/issues/483", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1559045150
Avoid negative width/height in Buffer::set_size Negative width/height may cause unexpected effects on the layout, which we can avoid by clamping. That would probably make sense. I wanted to keep the upper bound at i32::MAX instead of u32::MAX for now because we use these values in other places as i32. We could probably switch to u32 if we either clamp the u32 to i32::MAX or audit all the places these are used to make sure they could work with u32 instead.
gharchive/pull-request
2023-01-27T01:48:54
2025-04-01T06:45:25.202895
{ "authors": [ "grovesNL" ], "repo": "pop-os/cosmic-text", "url": "https://github.com/pop-os/cosmic-text/pull/66", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2388110279
feat: APIドキュメントを自動生成できるようにする 概要 Honoには Zod OpenAPI Hono というミドルウェアが存在していて、Zodのスキーマ定義からルーティングを書くことができる ルーティングからは自動的にOpenAPIのスキーマファイル(JSON形式)が出力される また、同じくHonoのミドルウェアに Scalar Hono API Reference Plugin というものがあり、これを設定すると自動的にスキーマ定義から見やすいドキュメントを自動生成してくれる これを活用してAPIのドキュメントを自動で生成できるようにしたい closed #463
gharchive/issue
2024-07-02T14:37:26
2025-04-01T06:45:25.217588
{ "authors": [ "laminne", "tufusa" ], "repo": "poporonnet/kcmsx", "url": "https://github.com/poporonnet/kcmsx/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
752621448
initial_size<1 or final_size<1 should be an error We currently enforce that Ne>0, but it should be Ne>=1. Unless there is some circumstance under which 0 < Ne < 1 makes sense? Oh, and our current upper limit is "finite", which in practise is 1e308. Maybe we could limit this to something less than the number of atoms in the universe? I.e. below 1e80?
gharchive/issue
2020-11-28T07:55:21
2025-04-01T06:45:25.226524
{ "authors": [ "grahamgower" ], "repo": "popsim-consortium/demes-python", "url": "https://github.com/popsim-consortium/demes-python/issues/149", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
2467356921
Add sequencing library and cohort back as covariates Was removed before because wasn't planning to use them as covariates. needs blacking but otherwise looks good to me! Good to include cohort as well for possible joint analysis :)
gharchive/pull-request
2024-08-15T04:06:42
2025-04-01T06:45:25.227519
{ "authors": [ "annacuomo", "hopedisastro" ], "repo": "populationgenomics/saige-tenk10k", "url": "https://github.com/populationgenomics/saige-tenk10k/pull/106", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1804135931
React & Redux group project - Space Travelers' Hub Hi, this week *no coding partner* has been assigned to me in my dashboard. Therefore, I am the only one have been working on this project. Space Travelers' Hub: Live Demo Highlights Upon the first render fetch data from the SpaceX API endpoints. Rendered UI lists accordingly Written actions for booking rockets and joining missions Written actions for canceling rockets and leaving missions Rendered My Profile Section with booked rockets and missions list. @Olamarx Hi, thanks for the review, however, as per the instruction All tasks should be based on the cards from your Kanban board. Writing tests are not mentioned as mandatory, So, it would be great if you could shed some light on that if it could be a blocker on PR approval, please? Thanks
gharchive/pull-request
2023-07-14T04:19:08
2025-04-01T06:45:25.230854
{ "authors": [ "porag-m06" ], "repo": "porag-m06/The-Space-Travelers-Hub", "url": "https://github.com/porag-m06/The-Space-Travelers-Hub/pull/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1465312798
🛑 PrivilegedBA is down In c19ec84, PrivilegedBA (https://privilegedba.tur.ar) was down: HTTP code: 0 Response time: 0 ms Resolved: PrivilegedBA is back up in 042fea3.
gharchive/issue
2022-11-27T02:46:33
2025-04-01T06:45:25.242235
{ "authors": [ "porrale" ], "repo": "porrale/status", "url": "https://github.com/porrale/status/issues/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
179066857
Add an action to clean orphaned volumes In the volumes view, add an action that removes all dangling volumes. Equivalent of the command docker volume rm $(docker volume ls -qf dangling=true). The #347 was rejected because it was considered too dangerous for the user, which I reckon too. Rejected. I am not comfortable with this feature. It exposes users to unexpected data loss should a volume be unattached but retained. Rather, I would prefer a filter that showed all dangling volumes, allowing the user to then select which volumes to delete. Even better, once we add "multi-select" with SHIFT keys (hold shift key whilst selecting the first and last sequential item) then that makes this filter scenario much safer. Actually, this is not blocked by #344 To implement the functionality, #344 must be implemented and dangling/orphaned volumes must be displayed in the volumes view (adding a extra column for example). By having both of these functionalities implemented, any user will be able to select multiple dangling volumes and trigger the remove action.
gharchive/issue
2016-09-25T04:19:47
2025-04-01T06:45:25.244623
{ "authors": [ "deviantony" ], "repo": "portainer/portainer", "url": "https://github.com/portainer/portainer/issues/231", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
837086729
I can't add Endpoint I can access API (http://x.x.x.x.155:2375/v1.40/containers/json) but when i add it (endpoint( on portainer didn't add it !!!! Raspberry 4 Portainer --> 2.1.1 Docker Version --> 20.10.5 *** I have 8 more on portainer and worked perfect Ticket update by Michael Cooper (Michael.cooper@portainer.io) Hello, I apologize that it took so long to get back to you. Could you show me exactly what it is you are trying to do to get this endpoint working? A few screenshots should do it. Could you show me one you did that works properly? I am just trying to get a feel of what is going on I your environment... Thank you for your time, Ticket update by Michael Cooper (Michael.cooper@portainer.io) No Response for over a month, Closing this ticket.
gharchive/issue
2021-03-21T14:04:28
2025-04-01T06:45:25.249060
{ "authors": [ "PortainerSupport", "cod3r-dev" ], "repo": "portainer/portainer", "url": "https://github.com/portainer/portainer/issues/4934", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
1213622582
Unable to retrieve templates Bug description can not load app templates Expected behavior load the templates Portainer Logs level=info msg="2022/04/22 08:48:19 [INFO] [http,server] [message: starting HTTPS server on port :9443]" level=info msg="2022/04/22 08:48:19 [DEBUG] [internal,init] [message: start initialization monitor ]" level=info msg="2022/04/22 08:48:19 [INFO] [http,server] [message: starting HTTP server on port :9000]" level=info msg="2022/04/22 08:48:45 http: TLS handshake error from 110.184.8.181:1576: remote error: tls: unknown certificate" level=info msg="2022/04/22 08:48:48 http error: A valid authorisation token is missing (err=Unauthorized) (code=401)" level=info msg="2022/04/22 08:48:53 http error: Invalid credentials (err=Unauthorized) (code=422)" level=info msg="2022/04/22 09:08:14 http error: Unable to write templates from templates URL (err=write tcp 172.19.0.2:9443->172.19.0.1:53154: write: broken pipe) (code=500)" level=info msg="2022/04/22 09:08:14 http: superfluous response.WriteHeader call from github.com/portainer/libhttp/error.writeErrorResponse (error.go:37)" level=info msg="2022/04/23 09:13:02 http error: Unable to write templates from templates URL (err=write tcp 172.19.0.2:9443->172.19.0.1:60030: write: broken pipe) (code=500)" level=info msg="2022/04/23 09:13:02 http: superfluous response.WriteHeader call from github.com/portainer/libhttp/error.writeErrorResponse (error.go:37)" level=info msg="2022/04/24 10:59:13 http error: A valid authorisation token is missing (err=Unauthorized) (code=401)" level=info msg="2022/04/24 11:04:09 http error: Unable to write templates from templates URL (err=write tcp 172.19.0.2:9443->172.19.0.1:40390: write: broken pipe) (code=500)" level=info msg="2022/04/24 11:04:09 http: superfluous response.WriteHeader call from github.com/portainer/libhttp/error.writeErrorResponse (error.go:37)" level=info msg="2022/04/24 11:04:09 http error: Unable to write templates from templates URL (err=write tcp 172.19.0.2:9443->172.19.0.1:40306: write: broken pipe) (code=500)" level=info msg="2022/04/24 11:04:09 http: superfluous response.WriteHeader call from github.com/portainer/libhttp/error.writeErrorResponse (error.go:37)" level=info msg="2022/04/24 11:04:11 http error: Unable to write templates from templates URL (err=write tcp 172.19.0.2:9443->172.19.0.1:40450: write: broken pipe) (code=500)" level=info msg="2022/04/24 11:04:11 http: superfluous response.WriteHeader call from github.com/portainer/libhttp/error.writeErrorResponse (error.go:37)" Steps to reproduce the issue: Enter the App Templates menu of the management platform ,waiting the page load Technical details: Portainer version:2.11.1 Docker version (managed by Portainer): 20.10.14 Kubernetes version (managed by Portainer): Platform (windows/linux): centos7 Command used to start Portainer (docker run -p 9443:9443 portainer/portainer): Browser: Use Case (delete as appropriate): Using Portainer at Home, Using Portainer in a Commercial setup. Have you reviewed our technical documentation and knowledge base? Yes/No Additional context Add any other context about the problem here. @flyingCrp Do you have outbound internet access from the Portainer host? Can you check the url from Settings--> App Templates? sorry bro ,maybe i know whats going on . im in china ,Sometimes access is unstable due to GFW problems
gharchive/issue
2022-04-24T11:07:43
2025-04-01T06:45:25.255583
{ "authors": [ "flyingCrp", "samdulam" ], "repo": "portainer/portainer", "url": "https://github.com/portainer/portainer/issues/6810", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
1587307233
Portainer 2.17.0 Business Edition no longer able to recreate containers - Error 500 Previously version 2.16 I could click on a container in Portainer, click on recreate, and then select to pull a new image - everything worked just fine. Recently when I try this, I'll get a spinning circle in the top right of the screen, and after a minute or two I'll get an error message saying; Failure Request failed with status code 500 The container will then be restarted with the original non-updated image, although the image will show up as having been renamed from its original name to a hex string such as fdf04daf6952. Can confirm this. You can workaround by pulling the latest image first, then recreate the container without the "Re-Pull image" switch. @socza Thank you for the information. I am currently investigating. I will update you as I learn more. Thanks! @socza Can you please open the Web Developer Tools >> open up the Network tab >> and screenshot the Request and Response for review? Thanks! I was the original reporter on Reddit. This happened on every container I’ve tried to update so far. This happens on specific containers, e.g. "minio" (I have it deployed too), "gitlab-ce", gitlab-runner", "jenkins". They are all tagged with "latest". As you wrote eg "registry:2" action "recreate" works. I was the original reporter on Reddit. This happened on every container I’ve tried to update so far. Yes, I copied your post and pasted it on github. Sorry, I wanted to report a case quickly. I was the original reporter on Reddit. This happened on every container I’ve tried to update so far. @Andrew-H2O Yes, I copied your post and pasted it on GitHub. Sorry, I wanted to report a case quickly (I didn't want to describe the problem again on GitHub). You're not angry, I owe you a beer... ;) Haha, it’s all good - I was just establishing my credentials :) @socza The containers you mentioned: "minio", "gitlab-ce", gitlab-runner", "jenkins" Were these created via docker run or via the Portainer UI? Thanks! The containers were created via docker run. Update: Upon further testing, this only occurs on containers that were created via docker run and not Portainer. I have forwarded this information to the team for review. Thanks! Happens to me on containers built in Portainer (stacks/web ui). Whether containers/stacks are on the host (win 10 / wsl2) or agents (synology). @theRealKLH Please provide container example for further review. Thanks! Hi I have the same problem. For containers I have not yet tried to update via the Pull new Image button as usual, if I first pulle the image and then Recreate it works. Unfortunately, this does not work for those containers that I have already tried before. Can someone help me? Thanks! We've just released 2.17.1 which should fix this issue - are you able to update and retest? We've just released 2.17.1 which should fix this issue - are you able to update and retest? Just updated and tested - it worked as expected. Thanks for the fix! Yes, I confirm it works correct. For me it's the same error again... Von: James Carppe @.> Gesendet: Mittwoch, 22. Februar 2023 03:58 An: portainer/portainer @.> Cc: Kassel Marco Daniel Andrin BZBS INS20a @.>; Comment @.> Betreff: Re: [portainer/portainer] Portainer 2.17.0 Business Edition no longer able to recreate containers - Error 500 (Issue #8507) Sie erhalten nicht oft eine E-Mail von @.@.>. Erfahren Sie, warum dies wichtig isthttps://aka.ms/LearnAboutSenderIdentification We've just released 2.17.1 which should fix this issue - are you able to update and retest? Reply to this email directly, view it on GitHubhttps://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fportainer%2Fportainer%2Fissues%2F8507%23issuecomment-1439369789&data=05|01|marco.kassel%40stud.bzbs.ch|1aa1a18025a24f8637c408db148094d9|5daf41bd338c4311b1b0e1299889c34b|0|0|638126314708616069|Unknown|TWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D|3000|||&sdata=OcBUNpvgE5uGMVN14O9%2BJBsvqm2L5go1QjU2btWTFtA%3D&reserved=0, or unsubscribehttps://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Fnotifications%2Funsubscribe-auth%2FA5YDIFA2A5ITK4X6PET4ONDWYV6CXANCNFSM6AAAAAAU54MC2U&data=05|01|marco.kassel%40stud.bzbs.ch|1aa1a18025a24f8637c408db148094d9|5daf41bd338c4311b1b0e1299889c34b|0|0|638126314708616069|Unknown|TWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D|3000|||&sdata=Ss9I1q5FjKRkp8kZD%2B92ofZu5Kl%2FktyIaEeE2g%2BCXs4%3D&reserved=0. You are receiving this because you commented.Message ID: @.@.>> @marco-kasl Can you provide the following: [1] What are the container images [2] How did you create the containers [3] Web Developer Tools >> open up the Network tab >> and screenshot the Request and Response for review I have 2 containers minio ldap that I had initially created via docker run that were not able to be re-created in 2.17.0. Upon upgrade, I am no longer seeing the 500 error. https://user-images.githubusercontent.com/20426210/220763497-ae452c83-aa69-465e-ad70-d288e7f2194c.mov Thanks! Problem still present in 2.17.1 with containers deployed trough the stack feature in portainer (not via docker run). Recreate container with re-pull fails with code 500, while updating the stack (containing the target container) with re-pull works as intended, which is what I'm using as a workaround. @Jikeh I performed the following: [1] Create Wordpress Stack [2] Re-create wordpress-wordpress-1 without re-pull [3] No Error [4] Re-create wordpress-wordpress-1 with re-pull [5] No Error https://user-images.githubusercontent.com/20426210/220787941-2cdcf083-27cd-4fe7-9df0-bc3d166483cc.mov https://user-images.githubusercontent.com/20426210/220787975-f6c6aa19-10bc-4c83-9cc8-dc226309adde.mov Can you provide the following: [1] What are the container images [2] How did you create the containers [3] Web Developer Tools >> open up the Network tab >> and screenshot the Request and Response for review Thanks! Sure thing. I also created a wordpress stack and I got no errors when recreating it, so I thought it was a problem with my config and I played around. I found out that containers with 0-1 networks can be recreated, while containers with multiple networks attached cannot (error 500). All containers in my setup (except for Portainer itself) are deployed through stacks. And in fact, testing recreation on a simple cointainer (mpepping/cyberchef) with 2 networks got the following response message: {"message":"Error recreating container","details":"create container error: Error response from daemon: Container cannot be connected to network endpoints: cyberchef-net, nginx-proxy-manager-net"} Should I open another issue or do you thing it is related? @theRealKLH Please provide container example for further review. Thanks! ooof.. sorry for the late reply. AFK since posting. Before coming back here, saw there was an upgrade to 2.17.1. Applied it and same outcome. I am seeing the same as @Jikeh both now and prior to upgrading. Stack update w/pull works, but container recreate w/pull does not. However, my response is different... {"message":"Error recreating container","details":"pull image error docker.io/willfarrell/autoheal:latest: Cannot connect to the Docker daemon at tcp://192.168.1.225:9001. Is the docker daemon running?"} I have also received... {"message":"Error recreating container","details":"pull image error docker.io/willfarrell/autoheal:latest: context deadline exceeded (Client.Timeout or context cancellation while reading body)"} It does appear that it is only happening on my 'agent' based systems (2 Synology NAS). I am not able to reproduce it on the stand-alone host (Win 10) even though I'm pretty sure it was acting the same prior to upgrading Portainer. I've tested all containers on that system and they are not showing the issue. I was just about to write that "it's only affecting containers pre-2.17" but I just tried a container that I created via stack post 2.17.1 and it failed. AND tried again just now and it was successful. So my issue seems to be one of timeout with my agents? Now, is that a "coincidental environmental" issue (net/hw/os-at times my Synos can be bogged down) or a "portainer" issue (app version)? GRIN I will say that I reverted everything down to 2.16.2 and didn't experience a single failure with the recreate process no matter how many times I tried (still failure until host Portainer system downgraded). Going to try upgrading again. @Jikeh Can you share one of your stack's code that is not re-creating the containers and I can test further on my end? Thanks! @Jikeh Can you share one of your stack's code that is not re-creating the containers and I can test further on my end? Thanks! There it is: version: '3.8' services: cyberchef: container_name: cyberchef image: mpepping/cyberchef restart: unless-stopped environment: - "TZ=Europe/Rome" user: 1000:1000 ports: - '10230:8000' networks: - cyberchef-net - nginx-proxy-manager-net networks: cyberchef-net: name: cyberchef-net driver: bridge nginx-proxy-manager-net: name: nginx-proxy-manager-net driver: bridge @Jikeh Thank you for the additional information. I am able to reproduce. I have logged an internal request for the unable to re-create containers with multiple networks. I do not yet have a timeline or release. I will update you as I learn more. Thanks! @Jikeh Thank you for the additional information. I am able to reproduce. I have logged an internal request for the unable to re-create containers with multiple networks. I do not yet have a timeline or release. I will update you as I learn more. Thanks! That's great to hear, thank you! @kroese Can you provide the following: [1] How did you create the container? docker run or Portainer UI? [2] Is this container part of a Stack? [3] Stack code or image tag Thanks! I have a similar problem. I added a stack of grafana, prometheus, endlessh and node-exporter and called it "monitoring". I upgraded my portainer to Business Edition and lost the possibility to edit or see the docker compose code inside the stack I tried to update grafana to the latest version, but failed with Request failed with status code 500. The latest Image gets downlaoded, but it doesn't get used. The iamge gets renamed to a random hex string. What I tried and didn't help: moving the container to another network and retry recreate without pulling the new container Duplicate/Edit and then Deploy the container throws Duplicate mount point I cannot delete the stack without the docker compose code I once put into stack. I cannot reach the docker-compose code because of the upghrade to BE. First of all, if you have an idea, how to get the docker compose code back, would be awesome. Secondly please explain how to update my grafana container Also noticed the issue in the latest version along with the errors in the logs which are highlighted in another issue. The main reason I use Portainer is for the ease of recreating containers and pulling a new image with little effort but the current version just seems to fail at every turn. I have also experienced this issue. An ETA on the fix would be appreciated. I solved this problem today with a short run of watchtower. This solved it flawlessly for me. docker-compose.yml: version: '3.7' services: watchtower: image: containrrr/watchtower container_name: watchtower restart: unless-stopped volumes: - /var/run/docker.sock:/var/run/docker.sock:ro environment: - TZ=Europe/Berlin - WATCHTOWER_CLEANUP=true - WATCHTOWER_INCLUDE_STOPPED=true - WATCHTOWER_POLL_INTERVAL=1800 - WATCHTOWER_DEBUG=true @kroese I created a homeassistance container via docker run with :latest and :stable images sudo docker run -d --name homeassistant-2 --privileged --restart=unless-stopped -p 8123:8123 -e TZ="Europe/London" -v homeassistant_config-2:/config --network=host ghcr.io/home-assistant/home-assistant:latest I am able to Recreate without re-pull image and with re-pull image: https://user-images.githubusercontent.com/20426210/228091449-b71f287d-cebc-4872-9273-f310bdb25ac9.mov Can you provide the following: [1] Launch Web Dev Tools >> Network tab [2] Try recreate [3] Post screenshots of the Request and Response tabs [4] What type of network do you create in your docker run? Thanks! @GFNiko Can you post your Stack code and I can test on my end? @zinkscott Please provide the following information so I can test your issue: [1] How did you create the container? docker run or Portainer UI? [2] Is this container part of a Stack? [3] Stack code or image tag Thanks! @GFNiko Can you post your Stack code and I can test on my end? No either there is an export function I couldn't find or, like I already wrote, there is no access to the stack code since I upgraded to portainer-be It just shows 'limited' in the stack overview I mostly got those error 500 errors when using re-create without stopping the container first. since I changed my habit to first stopping the containers and then use re-recreate with pull new image the errors are gone. I have just about give up on this as a solution, that a Business Edition application can have such huge problems for so many weeks without it being fixed or rolled back is ridiculous. For me it stopped working one i added my account to dockerhub This happened with 2.18 for me and 2.19.0, but only for containers created via stack. For non-stack containers, I can do recreate + re-pull, but for containers created via stack, I get a 500 and it fails. can confirm 2.19.0 CE has this issue again. I'm having to change ports every time to get containers back up and running. Just had my portainer freak out and kick 4 apps and have this 500 Status Failure when trying to start them back up. I feel like this issue may have deviated into a few different scenarios that may be unrelated to each other - unfortunately 500 is a very generic error message. Originally this thread was referring to stacks with containers that were on multiple networks, but I'm not sure if this is the case for all of you experiencing issues now. Those that have posted recently here, can you confirm whether your containers are on multiple networks? Try this stack. Start it up, wait for an update to either container. Try doing recreate + re-pull on the outdated container and you’ll get a 500. version: '3.3' services: tailscale: hostname: fastenhealth image: tailscale/tailscale:latest volumes: - /volume1/docker/fastenhealth/tailscale:/var/lib/tailscale:rw environment: - TS_USERSPACE=1 - TS_STATE_DIR=/var/lib/tailscale restart: unless-stopped fastenhealth: image: ghcr.io/fastenhealth/fasten-onprem:main network_mode: service:tailscale depends_on: - tailscale volumes: - /volume1/docker/fastenhealth/db:/opt/fasten/db:rw restart: unless-stopped @tamarahenson Just returning to this since Feb. I no longer have the issue "across the board". Now it has become isolated to a single container. That container is homeassistant. Like @kroese I am starting to believe it's some kind of timeout happening. When recreate from containers fail, recreate from stacks usually works, but I have also seen it pull the image but fail to stop/recreate the actual container. The container is left running referencing the "old" image as if I only performed a pull. One of the things I believe helped me was the system I was running back in Feb was kinda on the "slow" side. now on a more powerful NAS (but not TOO powerful I guess). Not working for me again after updating to 2.19.0 Try this stack. Start it up, wait for an update to either container. Try doing recreate + re-pull on the outdated container and you’ll get a 500. version: '3.3' services: tailscale: hostname: fastenhealth image: tailscale/tailscale:latest volumes: - /volume1/docker/fastenhealth/tailscale:/var/lib/tailscale:rw environment: - TS_USERSPACE=1 - TS_STATE_DIR=/var/lib/tailscale restart: unless-stopped fastenhealth: image: ghcr.io/fastenhealth/fasten-onprem:main network_mode: service:tailscale depends_on: - tailscale volumes: - /volume1/docker/fastenhealth/db:/opt/fasten/db:rw restart: unless-stopped I also get error 500's when doing this with a stack that has a depends on and network mode. Other stacks work fine. Stopping, removing images, starting seems to work fine, and it's likely that recreating twice would work as it does download the new images. There is another even bigger issues with occurring as well. The old container is being deleted after failing to create the new one. This leaves no means of recovery. Having the same issue since update to 2.19.4 Happening to all my containers regardless of network config. And also happening regardless of how the container was created. Getting a 500 error on recreate and repull, 2nd attempt works but wipes env Having the same issue since update to 2.19.4 BE Happening to all my containers regardless of network config. And also happening regardless of how the container was created. Getting a 500 error on recreate and repull, 2nd attempt works but wipes env Same issue here with Portainer BE Same here. Please fix this :( Since my last post back in 09/2023.. (we are now 1.5 years since my first post 02/2023. that was on version 2.16) Not sure when the issue went back to "across the board" from "only two remaining containers". I've updated versions multiple times and the issue remains. My current situation.... ALL containers (local & via agent) whether the image is already downloaded (doesn't matter if the image is the same or newer) require updating from "stacks editor" as opposed to "containers recreate". The "two remaining containers" from my last post (calibre & home assistant) require that I first manually download the images otherwise no matter how I attempt (via stacks editor or container recreate) the update fails. the exacty issue is alive on my sys aswell as of today Same issue: what should i do guys? Same here, running appreciate 40 containers. A lot containers can't even ping or connect to the docker host using omda stuff here I gave up on this and used https://github.com/mag37/dockcheck it was more reliable.
gharchive/issue
2023-02-16T09:10:12
2025-04-01T06:45:25.307953
{ "authors": [ "6XGate", "Aevrin", "Andrew-H2O", "GFNiko", "Hadatko", "IIIdefconIII", "Jikeh", "Milor123", "appiekap653", "avidrissman", "cros13", "danferd13", "hulleyrob", "jamescarppe", "marco-kasl", "mastermueller", "socza", "tamarahenson", "tarunx", "theRealKLH", "tytanick", "zinkscott" ], "repo": "portainer/portainer", "url": "https://github.com/portainer/portainer/issues/8507", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
1375650829
js: i18n: fix translation fallback to English Checklist [ ] All new and existing tests are passing [ ] (If adding features:) I have added tests to cover my changes [ ] (If docs changes needed:) I have updated the documentation accordingly. [ ] I have added an entry to CHANGES.rst because this is a user-facing change or an important bugfix [x] I have written proper commit message(s) What changes does this Pull Request introduce? Fix error when certain translation not present. Why is this necessary? When borwser's language is not English, and certain msgid dose not exist in target translation, we want provide the string in English, or [?${msgid}] format. But current implementation would just throw an error and interrupt the whole script, because en is not defined. Tested an LGTM
gharchive/pull-request
2022-09-16T09:01:01
2025-04-01T06:45:25.329118
{ "authors": [ "ix5", "rocka" ], "repo": "posativ/isso", "url": "https://github.com/posativ/isso/pull/927", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
450778698
Add support for Packet We adapted Typhoon to run on Packet in Lokomotive. In order to do this, we had to make a number of changes and include additional add-ons (e.g. MetalLB). Would Packet support be something that would be accepted in Typhoon if we did the work to prepare that and get it in? I've looked into Packet.net before I'm not interested in it yet. There are a number of factors: Packet: Use of Packet still has some warts / missing bits. Like, in your fork you've had to choose AWS Route53 for records since Packet has none. The multi-cluster balancing setup doesn't have a route toward stable, it relies on RR DNS of an external provider (contrast with bare-metal where the sky is the limit for how advanced users wanna get or DO (beta) where RR is used, but mainly due to cost cutting, LBs are present). Or firewalling just being done via the hosts. One could work around these issues, but there are also a variety of hosting platforms to choose from. I'm optimistic for Packet in the long run and really like their service, but I'm content to wait for the right time. Supportability: I use and run Typhoon clusters on each supported platform. But I don't run clusters on Packet. They're too costly without the direct support of the provider. I also worry that same cost calculus skews the potential user base of a Packet variant, favoring enterprise interests. Basically, until I can justify a personal need for Typhoon on Packet, it shouldn't be added to the distro bc that directly affects quality and decision making. Impl: I'd prefer to just implement new platforms afresh to know it inside and out, when the time is right. I think the minute decisions play a big role and initial working variants are straight forward enough. For example, I'd probably prefer direct integration via Matchbox and some other choices if I explored it. So I'm currently not aiming to add support for Packet. I'm glad you guys are using it though. Packet guys are great!
gharchive/issue
2019-05-31T12:38:16
2025-04-01T06:45:25.332709
{ "authors": [ "blixtra", "dghubble" ], "repo": "poseidon/typhoon", "url": "https://github.com/poseidon/typhoon/issues/489", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
460880521
README.md: fix link to azure module Signed-off-by: Mateusz Gozdek mgozdekof@gmail.com Thanks!
gharchive/pull-request
2019-06-26T10:09:16
2025-04-01T06:45:25.334755
{ "authors": [ "dghubble", "invidian" ], "repo": "poseidon/typhoon", "url": "https://github.com/poseidon/typhoon/pull/502", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1914115267
Maintain calculation results in session state @vnijs had a request to have the results of a long-running reactive calculation "persist" even through navigating away and back. In talking this through in person, maybe the equivalent of Shiny for R's bindCache (combined with bookmarkable state) would do it. @GShotwell mentioned the idea of saving models, which I really liked. Being able to load them back in + show output and create additional plots, diagnostics, etc. would also be really interesting. You could do that with a separate app designed just to evaluate a particular type of saved models. However, to be able to do it in the same estimation app would be really nice (i.e., restore input values, reactive.Calc state, and output). I encountered the same issue. I would like to ask if there have been any recent developments about the bindCache function. Thank you very much. Just curious if there are any updates on this, or people have any workarounds. Is there a Python equivalent of R's bindCache? My use case is that I have to hit an external web API, and that API request will be the same for each user of the app. So I would prefer to cache it, and include a time component in the key to invalidate the cached result after X minutes. I've read the docs but based on what I can tell there isn't currently this type of functionality supported in Shiny for Python? @sambrilleman I bet you could get this behaviour by using a reactive outside of the server function. This will let all of the users of the app share the same API call, which you could invalidate on a schedule with invalidate_later. Alternatively you could use reactive.poll to call a cheap API call regularly and only do the full API refresh when something changed. Here's an example I think https://github.com/posit-dev/py-shiny-templates/blob/main/monitor-database/shared.py#L26 https://github.com/posit-dev/py-shiny-templates/blob/main/monitor-database/app-core.py#L4 Legend! Thanks @gshotwell , we'll give that approach a go 😀 I guess I didn't realise we can use a reactive outside of the context of a server. I guess if it's out there on its own, not part of the server, it just forms its own reactive dependency graph or something like that 🤔 (i.e. only linked to the invalidation/polling like you mention). Interesting that it still fires on start up as part of the App though. I guess because its still called from the server (even if not declared there) so must be called/loaded by the app atleast once. Anyhow ignore me if I'm talking some gibberish, pretty new to Shiny. I'll take a closer look at your example, and we can just copy that behaviour, thanks! The reactive graph actually works the same way, it's just that when you put it outside of the server function it's shared across all users. So it's instantiated once when the app starts up, and all users share that reactive. When it invalidates it also invalidates all of the session objects that depend on it. The one thing that I'd be careful of with this pattern is altering the global reactive from within a session. I think it's better to just have a rule that global reactives are a kind of one-way trip and send data to sessions, but aren't modified by them. Yeah epic, that sounds reasonable. Thanks again! On Thu, 19 Sept 2024, 10:33 pm Gordon Shotwell, @.***> wrote: The reactive graph actually works the same way, it's just that when you put it outside of the server function it's shared across all users. So it's instantiated once when the app starts up, and all users share that reactive. When it invalidates it also invalidates all of the session objects that depend on it. The one thing that I'd be careful of with this pattern is altering the global reactive from within a session. I think it's better to just have a rule that global reactives are a kind of one-way trip and send data to sessions, but aren't modified by them. — Reply to this email directly, view it on GitHub https://github.com/posit-dev/py-shiny/issues/740#issuecomment-2360864588, or unsubscribe https://github.com/notifications/unsubscribe-auth/AEVECTCD53HGSSXMIHMNCSDZXLABDAVCNFSM6AAAAAA5IF4TWGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNRQHA3DINJYHA . You are receiving this because you were mentioned.Message ID: @.***>
gharchive/issue
2023-09-26T18:54:10
2025-04-01T06:45:25.347569
{ "authors": [ "gshotwell", "jcheng5", "liuzj039", "sambrilleman", "vnijs" ], "repo": "posit-dev/py-shiny", "url": "https://github.com/posit-dev/py-shiny/issues/740", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1876198395
Stop using deprecated pandas function This switches away from pandas.api.types.is_categorical_dtype, which started raising deprecation warnings in Pandas 2.1.0. Closing in favor of #726
gharchive/pull-request
2023-08-31T19:45:55
2025-04-01T06:45:25.349096
{ "authors": [ "schloerke", "wch" ], "repo": "posit-dev/py-shiny", "url": "https://github.com/posit-dev/py-shiny/pull/725", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2467862961
Update KATS for FIPS 203 Final FIPS 203 Final was issued on 13th August 2024. As such, KATs will need to be updated for all ML-KEM parameter sets. I have an updated kat_MLKEM_768.rsp file locally. Happy to contribute if anyone else can verify the results independently. See https://github.com/post-quantum-cryptography/KAT/pull/15
gharchive/issue
2024-08-15T11:13:46
2025-04-01T06:45:25.350936
{ "authors": [ "rod-chapman" ], "repo": "post-quantum-cryptography/KAT", "url": "https://github.com/post-quantum-cryptography/KAT/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
411242743
Add Sanity.io [x] verified that the CMS I'm adding is still maintained. [x] read CONTRIBUTING.md. [x] did not generate README.md. Thank you, @good-idea!
gharchive/pull-request
2019-02-17T22:20:43
2025-04-01T06:45:25.387285
{ "authors": [ "good-idea", "mutewinter" ], "repo": "postlight/awesome-cms", "url": "https://github.com/postlight/awesome-cms/pull/82", "license": "cc0-1.0", "license_type": "permissive", "license_source": "bigquery" }
65787631
Pep8 compliance These changes are for files which are in cocotb/cocotb. Changes include mostly: Deleting/putting whitespaces. Deleting/puttig blank lines. Wrapping lines which are >79 chars. Silencing some warnings when checking with flake. Removing unnecessary 'pass' statements. This PR is about issue #6 There are still some unused imports/variables and some other pep8 (and also logical) errors on the code. I left them untouched for now. Good work! Note the @landscape-bot "new" errors reported above are simply places where you've touched the code and are actually false positives. Just a note on workflow. It's a good idea to reference the issue in your commit messages. For PEP8 we already have an issue so you could write something like this: #6 Numerous style improvements for PEP8 Also it would be preferable to rebase so that this all squashes to a single commit. Also it would be preferable to rebase so that this all squashes to a single commit. Done!
gharchive/pull-request
2015-04-01T21:55:48
2025-04-01T06:45:25.477026
{ "authors": [ "chiggs", "mfkaptan" ], "repo": "potentialventures/cocotb", "url": "https://github.com/potentialventures/cocotb/pull/225", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2060244850
Stream works but no HSKV recording Hello team,I have Ubuntu 22.4 LTS - homebridge installation- latest version of nest-sdm v1.1.20. Nest outdoor camera stream works fine and events are captured in homebridge debug mode logs. But the hskv recordings are not happening. I have homebridge ffmpeg installed. My earlier installation on rpi ended up with this issue https://github.com/potmat/homebridge-google-nest-sdm/issues/101 and no success .So decided to give a try on ubuntu Please see complete log below Main errors I see are [h264 @ 0x560df6cd9940] [error] non-existing PPS 0 referenced Driveway Camera decode_slice_header error Driveway Camera [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:18:53 PM] [homebridge-google-nest-sdm] Video stream requested: 1280 x 720, 30 fps, 299 kbps Driveway Camera [12/29/2023, 3:18:53 PM] [homebridge-google-nest-sdm] Executing command sdm.devices.commands.CameraLiveStream.GenerateWebRtcStream with parameters {"offerSdp":"v=0\r\no=- 82755371 0 IN IP4 0.0.0.0\r\ns=-\r\nt=0 0\r\na=group:BUNDLE 0 1 2\r\na=extmap-allow-mixed\r\na=msid-semantic:WMS *\r\nm=audio 9 UDP/TLS/RTP/SAVPF 96\r\nc=IN IP4 0.0.0.0\r\na=ice-ufrag:629e\r\na=ice-pwd:9ab5288d5b990dec7bfd24\r\na=ice-options:trickle\r\na=fingerprint:sha-256 6D:2D:98:37:2A:E9:1E:1A:D0:24:9E:2E:98:E5:26:4E:F5:E1:42:40:A8:BD:CD:00:16:C0:4F:44:A3:A9:AE:CA\r\na=setup:actpass\r\na=recvonly\r\na=mid:0\r\na=msid:4f5d7569-5d2d-4167-81c6-6ef1d6c78258 7ff12b87-9d99-487a-bb4f-9809c0183fc3\r\na=rtcp:9 IN IP4 0.0.0.0\r\na=rtcp-mux\r\na=ssrc:145330168 cname:9e92855c-b13f-4d67-971c-17e21020c6a3\r\na=rtpmap:96 opus/48000/2\r\nm=video 9 UDP/TLS/RTP/SAVPF 97\r\nc=IN IP4 0.0.0.0\r\na=ice-ufrag:629e\r\na=ice-pwd:9ab5288d5b990dec7bfd24\r\na=ice-options:trickle\r\na=fingerprint:sha-256 6D:2D:98:37:2A:E9:1E:1A:D0:24:9E:2E:98:E5:26:4E:F5:E1:42:40:A8:BD:CD:00:16:C0:4F:44:A3:A9:AE:CA\r\na=setup:actpass\r\na=recvonly\r\na=mid:1\r\na=msid:b68a27ae-28e5-40c8-b3c0-87e3851568fe c8940906-7f60-41ad-be9a-ddc6728fbe9b\r\na=rtcp:9 IN IP4 0.0.0.0\r\na=rtcp-mux\r\na=ssrc:1819314906 cname:9e92855c-b13f-4d67-971c-17e21020c6a3\r\na=rtpmap:97 H264/90000\r\na=rtcp-fb:97 transport-cc\r\na=rtcp-fb:97 ccm fir\r\na=rtcp-fb:97 nack\r\na=rtcp-fb:97 nack pli\r\na=rtcp-fb:97 goog-remb\r\na=fmtp:97 level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=42e01f\r\nm=application 9 UDP/DTLS/SCTP webrtc-datachannel\r\nc=IN IP4 0.0.0.0\r\na=ice-ufrag:629e\r\na=ice-pwd:9ab5288d5b990dec7bfd24\r\na=ice-options:trickle\r\na=fingerprint:sha-256 6D:2D:98:37:2A:E9:1E:1A:D0:24:9E:2E:98:E5:26:4E:F5:E1:42:40:A8:BD:CD:00:16:C0:4F:44:A3:A9:AE:CA\r\na=setup:actpass\r\na=mid:2\r\na=sctp-port:5000\r\na=max-message-size:65536\r\n"} Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] Execution of command sdm.devices.commands.CameraLiveStream.GenerateWebRtcStream returned {"answerSdp":"v=0\r\no=- 0 2 IN IP4 127.0.0.1\r\ns=-\r\nt=0 0\r\na=group:BUNDLE 0 2 1\r\na=msid-semantic: WMS 2595933019133906505/416178219 virtual-6666\r\na=ice-lite\r\nm=audio 19305 UDP/TLS/RTP/SAVPF 96\r\nc=IN IP4 142.250.111.127\r\na=rtcp:9 IN IP4 0.0.0.0\r\na=candidate: 1 udp 2113939711 2607:f8b0:4023:1401::7f 19305 typ host generation 0\r\na=candidate: 1 tcp 2113939710 2607:f8b0:4023:1401::7f 19305 typ host tcptype passive generation 0\r\na=candidate: 1 ssltcp 2113939709 2607:f8b0:4023:1401::7f 443 typ host generation 0\r\na=candidate: 1 udp 2113932031 142.250.111.127 19305 typ host generation 0\r\na=candidate: 1 tcp 2113932030 142.250.111.127 19305 typ host tcptype passive generation 0\r\na=candidate: 1 ssltcp 2113932029 142.250.111.127 443 typ host generation 0\r\na=ice-ufrag:QSekoobsOlBM3goKWigaIAEQABgHCAMGBQQBAggSDAiS\r\na=ice-pwd:DP25PYLF/CZHEYPD8BGUQPXZ\r\na=fingerprint:sha-256 75:CF:58:21:2A:11:A6:C1:E1:39:E6:00:71:01:CD:C6:D3:4D:F5:19:D2:C9:4E:D9:A7:EE:02:E4:52:FD:83:F9\r\na=setup:passive\r\na=mid:0\r\na=sendrecv\r\na=msid:virtual-6666 virtual-6666\r\na=rtcp-mux\r\na=rtpmap:96 opus/48000/2\r\na=fmtp:96 minptime=10;useinbandfec=1\r\na=ssrc:6666 cname:6666\r\nm=video 9 UDP/TLS/RTP/SAVPF 97\r\nc=IN IP4 0.0.0.0\r\na=rtcp:9 IN IP4 0.0.0.0\r\na=ice-ufrag:QSekoobsOlBM3goKWigaIAEQABgHCAMGBQQBAggSDAiS\r\na=ice-pwd:DP25PYLF/CZHEYPD8BGUQPXZ\r\na=fingerprint:sha-256 75:CF:58:21:2A:11:A6:C1:E1:39:E6:00:71:01:CD:C6:D3:4D:F5:19:D2:C9:4E:D9:A7:EE:02:E4:52:FD:83:F9\r\na=setup:passive\r\na=mid:1\r\na=sendrecv\r\na=msid:2595933019133906505/416178219 2595933019133906505/416178219\r\na=rtcp-mux\r\na=rtpmap:97 H264/90000\r\na=rtcp-fb:97 ccm fir\r\na=rtcp-fb:97 nack\r\na=rtcp-fb:97 nack pli\r\na=rtcp-fb:97 goog-remb\r\na=fmtp:97 level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=42e01f\r\na=ssrc:416178219 cname:416178219\r\nm=application 9 DTLS/SCTP 5000\r\nc=IN IP4 0.0.0.0\r\na=ice-ufrag:QSekoobsOlBM3goKWigaIAEQABgHCAMGBQQBAggSDAiS\r\na=ice-pwd:DP25PYLF/CZHEYPD8BGUQPXZ\r\na=fingerprint:sha-256 75:CF:58:21:2A:11:A6:C1:E1:39:E6:00:71:01:CD:C6:D3:4D:F5:19:D2:C9:4E:D9:A7:EE:02:E4:52:FD:83:F9\r\na=setup:passive\r\na=mid:2\r\na=sctpmap:5000 webrtc-datachannel 1024\r\n","expiresAt":"2023-12-29T15:23:57.007541Z","mediaSessionId":"QSekoobsOlBM3goKWigaIAEQABgHCAMGBQQBAggSDAiS"} Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] Stream command: ffmpeg -protocol_whitelist pipe,crypto,udp,rtp,fd -analyzeduration 15000000 -probesize 100000000 -i - -an -sn -dn -codec:v libx264 -preset ultrafast -tune zerolatency -pix_fmt yuv420p -color_range mpeg -bf 0 -r 30 -b:v 1196k -bufsize 1196k -maxrate 2392k -filter:v scale='min(1280,iw)':'min(720,ih)':force_original_aspect_ratio=decrease,scale=trunc(iw/2)*2:trunc(ih/2)*2 -payload_type 99 -ssrc 3128821 -f rtp -srtp_out_suite AES_CM_128_HMAC_SHA1_80 -srtp_out_params JnEj5Ei2taIpcHBbV/06MUJkwEag+nA3tKSP5jci srtp://192.168.5.60:50715?rtcpport=50715&pkt_size=1378 -vn -sn -dn -codec:a libfdk_aac -profile:a aac_eld -flags +global_header -ar 16k -b:a 24k -ac 1 -payload_type 110 -ssrc 2825176 -f rtp -srtp_out_suite AES_CM_128_HMAC_SHA1_80 -srtp_out_params fi5TDk4gdLj0k04E3jmiG79fi68RE5VqjGiZA03I srtp://192.168.5.60:63149?rtcpport=63149&pkt_size=188 -loglevel level+verbose v=0 o=- 0 0 IN IP4 127.0.0.1 s=- c=IN IP4 127.0.0.1 t=0 0 m=audio 15127 UDP 96 a=rtpmap:96 opus/48000/2 a=fmtp:96 minptime=10;useinbandfec=1 a=rtcp-fb:96 transport-cc a=sendrecv m=video 14646 UDP 97 a=rtpmap:97 H264/90000 a=rtcp-fb:97 ccm fir a=rtcp-fb:97 nack a=rtcp-fb:97 nack pli a=rtcp-fb:97 goog-remb a=fmtp:97 level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=42e01f a=sendrecv Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] ffmpeg version 5.0 Copyright (c) 2000-2022 the FFmpeg developers Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] built with gcc 6.3.0 (Debian 6.3.0-18+deb9u1) 20170516 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] configuration: --pkgconfigdir=/build/workspace/lib/pkgconfig --prefix=/build/workspace --pkg-config-flags=--static --extra-cflags=-I/build/workspace/include --extra-ldflags=-L/build/workspace/lib --extra-libs='-lpthread -lm' --enable-static --disable-debug --disable-shared --disable-ffplay --disable-doc --enable-openssl --enable-gpl --enable-version3 --enable-nonfree --enable-pthreads --enable-libvpx --enable-libmp3lame --enable-libopus --enable-libtheora --enable-libvorbis --enable-libx264 --enable-runtime-cpudetect --enable-libfdk-aac --enable-avfilter --enable-libopencore_amrwb --enable-libopencore_amrnb --enable-filters --enable-decoder=h264 --enable-network --enable-protocol=tcp --enable-libspeex --enable-demuxer=rtsp --enable-indev=alsa --enable-outdev=alsa Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libavutil 57. 17.100 / 57. 17.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libavcodec 59. 18.100 / 59. 18.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libavformat 59. 16.100 / 59. 16.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libavdevice 59. 4.100 / 59. 4.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libavfilter 8. 24.100 / 8. 24.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libswscale 6. 4.100 / 6. 4.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libswresample 4. 3.100 / 4. 3.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [info] libpostproc 56. 3.100 / 56. 3.100 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [sdp @ 0x560b9cf0d400] [verbose] setting jitter buffer size to 500 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [sdp @ 0x560b9cf0d400] Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] [verbose] setting jitter buffer size to 500 Driveway Camera [12/29/2023, 3:18:57 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:18:59 PM] [homebridge-google-nest-sdm] [h264 @ 0x560b9cf13940] [verbose] Reinit context to 1920x1088, pix_fmt: yuv420p Driveway Camera [12/29/2023, 3:18:59 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Input #0, sdp, from 'pipe:': Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] title : - Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Duration: N/A, start: 0.000000, bitrate: N/A Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream #0:0: Audio: opus, 48000 Hz, stereo, fltp Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream #0:1: Video: h264 (High), 1 reference frame, yuv420p(progressive, left), 1920x1080 (1920x1088), 30 fps, 30 tbr, 90k tbn Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream mapping: Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream #0:1 -> #0:0 (h264 (native) -> h264 (libx264)) Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream #0:0 -> #1:0 (opus (native) -> aac (libfdk_aac)) Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [graph_1_in_0_0 @ 0x560b9d0a2c00] [verbose] tb:1/48000 samplefmt:fltp samplerate:48000 chlayout:0x3 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [format_out_1_0 @ 0x560b9d0a3e00] [verbose] auto-inserting filter 'auto_aresample_0' between the filter 'Parsed_anull_0' and the filter 'format_out_1_0' Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [auto_aresample_0 @ 0x560b9d065600] [verbose] ch:2 chl:stereo fmt:fltp r:48000Hz -> ch:1 chl:mono fmt:s16 r:16000Hz Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Output #1, rtp, to 'srtp://192.168.5.60:63149?rtcpport=63149&pkt_size=188': Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] title : - Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] encoder : Lavf59.16.100 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream #1:0: Audio: aac (ELD), 16000 Hz, mono, s16, delay 288, 24 kb/s Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] encoder : Lavc59.18.100 libfdk_aac Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [h264 @ 0x560b9cf19400] [verbose] Reinit context to 1920x1088, pix_fmt: yuv420p Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [Parsed_scale_0 @ 0x560b9d099840] [verbose] w:min(1280,iw) h:min(720,ih) flags:'' interl:0 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [Parsed_scale_1 @ 0x560b9cfd4500] [verbose] w:trunc(iw/2)*2 h:trunc(ih/2)*2 flags:'' interl:0 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [graph 0 input from stream 0:1 @ 0x560b9cfd4a00] [verbose] w:1920 h:1080 pixfmt:yuv420p tb:1/90000 fr:30/1 sar:0/1 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [Parsed_scale_0 @ 0x560b9d099840] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] w:1920 h:1080 fmt:yuv420p sar:0/1 -> w:1280 h:720 fmt:yuv420p sar:0/1 flags:0x0 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [Parsed_scale_1 @ 0x560b9cfd4500] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] w:1280 h:720 fmt:yuv420p sar:0/1 -> w:1280 h:720 fmt:yuv420p sar:0/1 flags:0x0 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [libx264 @ 0x560b9cfd1e80] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] using cpu capabilities: MMX2 SSE2Fast SSSE3 SSE4.1 Cache64 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [libx264 @ 0x560b9cfd1e80] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] profile Constrained Baseline, level 3.1, 4:2:0, 8-bit Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Output #0, rtp, to 'srtp://192.168.5.60:50715?rtcpport=50715&pkt_size=1378': Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Stream #0:0: Video: h264, 1 reference frame, yuv420p(tv, progressive, left), 1280x720 (0x0), q=2-31, 1196 kb/s, 30 fps, 90k tbn Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] encoder : Lavc59.18.100 libx264 Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] Side data: Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] cpb: bitrate max/min/avg: 2392000/0/1196000 buffer size: 1196000 vbv_delay: N/A Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:19:00 PM] [homebridge-google-nest-sdm] [info] frame= 1 fps=0.0 q=38.0 size= 37kB time=00:00:01.77 bitrate= 172.5kbits/s speed=38.4x Driveway Camera [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] non-existing PPS 0 referenced Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] non-existing PPS 0 referenced Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] decode_slice_header error Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] no frame! Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] non-existing PPS 0 referenced Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] non-existing PPS 0 referenced Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] decode_slice_header error Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [error] no frame! Driveway Camera [12/29/2023, 3:16:13 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:14 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cd9940] [verbose] Reinit context to 1920x1088, pix_fmt: yuv420p Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Input #0, sdp, from 'pipe:': Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] title : - Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Duration: N/A, start: 0.000000, bitrate: N/A Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream #0:0: Audio: opus, 48000 Hz, stereo, fltp Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream #0:1: Video: h264 (High), 1 reference frame, yuv420p(progressive, left), 1920x1080 (1920x1088), 30 fps, 30 tbr, 90k tbn Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream mapping: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream #0:1 -> #0:0 (h264 (native) -> h264 (libx264)) Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream #0:0 -> #1:0 (opus (native) -> aac (libfdk_aac)) Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [graph_1_in_0_0 @ 0x560df6e22740] [verbose] tb:1/48000 samplefmt:fltp samplerate:48000 chlayout:0x3 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [format_out_1_0 @ 0x560df6e23ac0] [verbose] auto-inserting filter 'auto_aresample_0' between the filter 'Parsed_anull_0' and the filter 'format_out_1_0' Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [auto_aresample_0 @ 0x560df6e25180] [verbose] ch:2 chl:stereo fmt:fltp r:48000Hz -> ch:1 chl:mono fmt:s16 r:16000Hz Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Output #1, rtp, to 'srtp://192.168.5.60:62836?rtcpport=62836&pkt_size=188': Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] title : - Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] encoder : Lavf59.16.100 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream #1:0: Audio: aac (ELD), 16000 Hz, mono, s16, delay 288, 24 kb/s Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] encoder : Lavc59.18.100 libfdk_aac Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [h264 @ 0x560df6cddf40] [verbose] Reinit context to 1920x1088, pix_fmt: yuv420p Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [Parsed_scale_0 @ 0x560df6d9e040] [verbose] w:min(1280,iw) h:min(720,ih) flags:'' interl:0 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [Parsed_scale_1 @ 0x560df6e3a440] [verbose] w:trunc(iw/2)*2 h:trunc(ih/2)*2 flags:'' interl:0 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [graph 0 input from stream 0:1 @ 0x560df6d9bf40] [verbose] w:1920 h:1080 pixfmt:yuv420p tb:1/90000 fr:30/1 sar:0/1 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [Parsed_scale_0 @ 0x560df6d9e040] [verbose] w:1920 h:1080 fmt:yuv420p sar:0/1 -> w:1280 h:720 fmt:yuv420p sar:0/1 flags:0x0 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [Parsed_scale_1 @ 0x560df6e3a440] [verbose] w:1280 h:720 fmt:yuv420p sar:0/1 -> w:1280 h:720 fmt:yuv420p sar:0/1 flags:0x0 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [libx264 @ 0x560df6daadc0] [info] using cpu capabilities: MMX2 SSE2Fast SSSE3 SSE4.1 Cache64 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [libx264 @ 0x560df6daadc0] [info] profile Constrained Baseline, level 3.1, 4:2:0, 8-bit Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Output #0, rtp, to 'srtp://192.168.5.60:62187?rtcpport=62187&pkt_size=1378': Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] title : - Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] encoder : Lavf59.16.100 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Stream #0:0: Video: h264, 1 reference frame, yuv420p(tv, progressive, left), 1280x720 (0x0), q=2-31, 1196 kb/s, 30 fps, 90k tbn Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Metadata: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] encoder : Lavc59.18.100 libx264 Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] Side data: Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [info] cpb: bitrate max/min/avg: 2392000/0/1196000 buffer size: 1196000 vbv_delay: N/A Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] [verbose] *** 78 dup! Driveway Camera [12/29/2023, 3:16:16 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [info] frame= 79 fps=0.0 q=20.0 size= 272kB time=00:00:02.60 bitrate= 857.6kbits/s dup=78 drop=0 speed= 44x Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [info] frame= 79 fps=0.0 q=20.0 size= 272kB time=00:00:02.60 bitrate= 857.6kbits/s dup=78 drop=0 speed=3.56x Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] *** dropping frame 79 from stream 0 at ts 1 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] *** dropping frame 79 from stream 0 at ts 2 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] *** dropping frame 79 from stream 0 at ts 3 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] *** dropping frame 79 from stream 0 at ts 4 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] *** dropping frame 79 from stream 0 at ts 5 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] *** dropping frame 79 from stream 0 at ts 6 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] *** dropping frame 79 from stream 0 at ts 7 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] *** dropping frame 79 from stream 0 at ts 8 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] *** dropping frame 79 from stream 0 at ts 9 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] *** dropping frame 79 from stream 0 at ts 10 Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera 12/29/2023, 3:16:17 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:18 PM] [homebridge-google-nest-sdm] [info] frame= 99 fps= 57 q=23.0 size= 454kB time=00:00:04.91 bitrate= 758.3kbits/s dup=78 drop=75 speed=2.83x Driveway Camera [12/29/2023, 3:16:18 PM] [homebridge-google-nest-sdm] [info] frame= 132 fps= 59 q=23.0 size= 632kB time=00:00:06.22 bitrate= 831.6kbits/s dup=78 drop=75 speed=2.77x Driveway Camera [12/29/2023, 3:16:18 PM] [homebridge-google-nest-sdm] [verbose] Driveway Camera [12/29/2023, 3:16:18 PM] [homebridge-google-nest-sdm] *** dropping frame 133 from stream 0 at ts 131 Driveway Camera [12/29/2023, 3:16:18 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:19 PM] [homebridge-google-nest-sdm] [info] frame= 166 fps= 60 q=21.0 size= 707kB time=00:00:06.73 bitrate= 860.0kbits/s dup=78 drop=76 speed=2.44x Driveway Camera [12/29/2023, 3:16:19 PM] [homebridge-google-nest-sdm] [info] frame= 190 fps= 58 q=20.0 size= 785kB time=00:00:07.27 bitrate= 883.8kbits/s dup=78 drop=76 speed=2.23x Driveway Camera [12/29/2023, 3:16:20 PM] [homebridge-google-nest-sdm] [info] frame= 190 fps= 50 q=20.0 size= 785kB time=00:00:07.79 bitrate= 825.7kbits/s dup=78 drop=76 speed=2.05x Driveway Camera [12/29/2023, 3:16:20 PM] [homebridge-google-nest-sdm] Received request to reconfigure: 640 x 360, 30 fps, 132 kbps (Ignored) Driveway Camera [12/29/2023, 3:16:20 PM] [homebridge-google-nest-sdm] [info] frame= 193 fps= 44 q=24.0 size= 841kB time=00:00:08.30 bitrate= 830.3kbits/s dup=78 drop=76 speed=1.91x Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [info] frame= 219 fps= 45 q=21.0 size= 986kB time=00:00:08.87 bitrate= 910.3kbits/s dup=78 drop=76 speed=1.83x Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] max delay reached. need to consume packet Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] RTP: missed 3 packets Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] max delay reached. need to consume packet Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] RTP: missed 1 packets Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [info] frame= 219 fps= 40 q=21.0 size= 986kB time=00:00:09.13 bitrate= 885.1kbits/s dup=78 drop=76 speed=1.68x Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] max delay reached. need to consume packet Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] RTP: missed 1 packets Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] max delay reached. need to consume packet Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] RTP: missed 2 packets Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [warning] RTP: dropping old packet received too late Driveway Camera [12/29/2023, 3:16:21 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [sdp @ 0x560df6cd3400] Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] [error] Empty H.264 RTP packet Driveway Camera [12/29/2023, 3:16:30 PM] [homebridge-google-nest-sdm] Driveway Camera @vgnnair: Those errors are perfectly normal and should not affect anything. According to the logs everything looks fine, did you read the HKSV section in the readme? Thanks @potmat. Make sense .Therefore moved to a different kind of implementation for hskv's
gharchive/issue
2023-12-29T16:36:37
2025-04-01T06:45:25.577322
{ "authors": [ "potmat", "vgnnair" ], "repo": "potmat/homebridge-google-nest-sdm", "url": "https://github.com/potmat/homebridge-google-nest-sdm/issues/130", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1475296250
Added an optional config to use a switch for Eco Mode on thermostats. Default to the current behavior of using a custom characteristic. Custom characteristics are not visible through the default Home application. Using a switch within the accessory provides an option for anyone who wants to keep all automations compatible with the Home app. @hn-git: It seems possible to create automations with custom characteristics using an app like Eve: What happens if you do this? Does the Home app throw up errors or cause problems? Yes, it's possible to create using other apps. When the scene with this action is viewed in the Home app it displays the accessory but not the characteristic being updated (eg. it just shows the thermostat being on). I prefer to keep everything managed through the Home app without using scenes so it was useful to me. Feel free to close if you don't want to support maintaining/customizing this. @hn-git: My personal preference has been to avoid putting in "fake" accessories like switches/lights just to get around the limitations of the Home app.
gharchive/pull-request
2022-12-04T23:27:14
2025-04-01T06:45:25.581544
{ "authors": [ "hn-git", "potmat" ], "repo": "potmat/homebridge-google-nest-sdm", "url": "https://github.com/potmat/homebridge-google-nest-sdm/pull/46", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1100413221
[Bug] right clicking on profile while in a voice chat crashed the client (since 5d13037) Describe the bug A clear and concise description of what the bug is. To Reproduce Steps to reproduce the behavior: Expected behavior A clear and concise description of what you expected to happen. this is the error in the console: if u give me 30 minutes i'll open a PR with a bug fix ah, I didn't expect that there might be empty, I will push it with the next commit with other small fixes
gharchive/issue
2022-01-12T14:33:09
2025-04-01T06:45:25.602049
{ "authors": [ "Giotje", "userXinos" ], "repo": "powerfart-plugins/image-tools", "url": "https://github.com/powerfart-plugins/image-tools/issues/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2143847153
bug: sanitize username with special characters Description When username contains special chars like ., -, '/, PPResume would fail to generate and set a valid logto username as logto do not allow special chars (with _` as the only exception) as part of the username. To Reproduce Steps to reproduce the behavior: Register a new account with email like a.b@xxx.com Check whether the new account has a valid username Actual Behavior The new registered account would have null username Expected Behavior The new registered account should have a sanitized username as well. [Optional] Screenshots NA Additional Context This is a similar bug to https://github.com/ppresume/ppresume/issues/166 Done. We also revised the style for username in /settings page:
gharchive/issue
2024-02-20T08:51:28
2025-04-01T06:45:25.659720
{ "authors": [ "xiaohanyu" ], "repo": "ppresume/community", "url": "https://github.com/ppresume/community/issues/30", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1537437194
chore: updates Description chore: upgrade packages Additional context What is the purpose of this pull request? [ ] Bug fix [ ] New Feature [ ] Documentation update [ ] Other Before submitting the PR, please make sure you do the following [ ] Read the Contributing Guidelines. [ ] Check that there isn't already a PR that solves the problem the same way to avoid creating a duplicate. [ ] Provide a description in this PR that addresses what the PR is solving, or reference the issue that it solves (e.g. fixes #123). [ ] Ideally, include relevant tests that fail without this PR but pass with it. Codecov Report Merging #231 (b2bc759) into master (60e1357) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #231 +/- ## ======================================= Coverage 77.49% 77.49% ======================================= Files 130 130 Lines 12264 12264 Branches 1305 1305 ======================================= Hits 9504 9504 Misses 2760 2760 Flag Coverage Δ unittests 77.49% <ø> (ø) Flags with carried forward coverage won't be shown. Click here to find out more. Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
gharchive/pull-request
2023-01-18T05:49:02
2025-04-01T06:45:25.721090
{ "authors": [ "codecov-commenter", "prabhuignoto" ], "repo": "prabhuignoto/react-creme", "url": "https://github.com/prabhuignoto/react-creme/pull/231", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
270020767
License? Please remember to add a LICENSE(.md) file to this repository. That way, people will know how they may use/modify the code. Hi, Thank you for looking at my code. Its a good idea to add a license file. I will look into that.
gharchive/issue
2017-10-31T16:12:37
2025-04-01T06:45:25.722916
{ "authors": [ "brylie", "kimfalk" ], "repo": "practical-recommender-systems/moviegeek", "url": "https://github.com/practical-recommender-systems/moviegeek/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
892759446
Reset Banco Por conta da alteração da estura do banco pela issue #281 tanto os banco de teste e produção devem receber o migrate com as alterações. Mas aproveitar para resetar ambos com php artisan migrate:refresh php artisan db:seed isso já resolveria a issue #286 também! Com o Mural V2 existira novas funcionalidades e acredito que será realizado esse processo. No banco do dev foi realizado o processo.
gharchive/issue
2021-05-16T21:04:49
2025-04-01T06:45:25.727978
{ "authors": [ "CleissonVieira", "oraphaBorges" ], "repo": "practice-uffs/mural", "url": "https://github.com/practice-uffs/mural/issues/300", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
716084698
Edição de vídeo Letras CH (PTA E LPT): 11º vídeo - Resenha parte 1. Profª Morgana Essa issue em questão refere-se ao vídeo 11 - Resenha parte 1. Profª Morgana. O vídeo está na pasta deste link Essa série de issues são relacionadas a issue #151, cada uma delas representa um vídeo a ser editado. Todas as edições seguem o mesmo padrão (que pode ser visto aqui) e são feitas no software ShotCut. Os materiais do programa podem ser encontrados na pasta de materiais e as ferramentas no guia do programa As etapas de edição desses vídeos são: Primeira etapa: [ ] Realizar edição inicial do vídeo seguindo orientações do respectivo professor e agregar slide enviado (arquivos na pasta). Os slides devem ser baixados em formato de imagem, para serem mais facilmente utilizados na edição. [x] Caso tenha imagens que precisem de áudio descrição deve-se gravar e agregar ao vídeo. [x] Criar uma pasta no link dessa issue chamada Em progresso e armazenar todo conteúdo e o andamento da edição nela. [ ] Criação da pagina dos créditos seguindo o template feito pela equipe de conteúdo. [ ] Inserção das vinhetas e da pagina de créditos com a seguinte ordem: Vinheta de Letras - Vinheta PRACTICE - Vídeo - Vinheta Letras - Créditos - Vinheta PRACTICE Segunda etapa: Com a edição inicial concluída, deve ser feita a legenda do vídeo. [x] Upar vídeo editado no YouTube para criação da legenda automática (Para saber como fazer isso, veja o vídeo deste link) [x] Corrigir as palavras erradas da legenda automática gerada e baixar o arquivo .srt (seguindo o vídeo do tópico anterior) [x] Com a legenda finalizada, disponibilizar o arquivo e marcar a equipe de conteúdo para revisão textual. [ ] Ao final da revisão as legendas devem ser adicionadas ao vídeo através do software Format Factory (Veja como fazer aqui) [ ] Com o vídeo finalizado e com as legendas, enviar ao professor para revisão e apontar possíveis ajustes. [ ] Caso tenha, realizar os ajustes solicitados. Responsável: @brendaellen É necessário fazer a transposição dos slides disponibilizados pelo(a) professor(a) referente a esse vídeo. Para fazer a transposição você deve: [ ] Localizar o template elaborado para o curso nesta pasta; Os templates utilizados serão os de capa, créditos 02 e 05; Sendo o 02 para trechos com título e texto e o 05 para trechos com imagem; [ ] Copiar o template e transferi-lo para a pasta dessa tarefa (edição) em questão (que está posta na descrição dessa tarefa); [ ] Adicionar títulos, textos, imagens e referências em sua totalidade; [ ] Conferir se margens, fontes, tamanhos das fontes e parágrafos estão padronizados em toda a apresentação; [ ] Conferir erros de digitação. É bom lembrar que os slides devem estar em 1920x1080 (16x9). Pois os slides que a professora disponibilizou estão em 4x4 A transcrição do vídeo e o link para a primeira versão dele está nesta pasta, @practice-uffs/con-conteudo. Irei fazer a segunda versão do vídeo quando os slides normatizados e a revisão textual estiverem prontos. Responsável: @thaliafriedrich Para a revisão textual das legendas: [ ] Assistir ao vídeo concomitantemente a leitura da transcrição com a minutagem; [ ] Fazer alterações de pontuação e ortografia; [ ] Mudar palavras caso elas não estiverem na fala do prof (as vezes a legenda automática coloca umas palavras nada a ver com o que está sendo dito); [ ] Rever o vídeo e reler a transcrição da legenda, para ter certeza que nenhum trecho ficou errado. Observações: Não há necessidade de deixar no texto da legenda vícios de linguagem, como "tipo", "né" ou similares. A transcrição do vídeo e o link para a primeira versão dele está nesta pasta, @practice-uffs/con-conteudo. Irei fazer a segunda versão do vídeo quando os slides normatizados e a revisão textual estiverem prontos. @robisonjose, estou sem acesso ao vídeo contido na pasta que você mencionou. O link direciona para um vídeo indisponível (ou privado). Peço que permita o acesso para eu revisar a legenda. O mesmo ocorre na issue #196. Obrigada. @thaliafriedrich Realmente, havia colocado privado, mas agora alterei a privacidade nos dois. Creio que agora está dando para acessar. Revisão da legenda finalizada. Acho que quando finaliza tem que colocar "texto:ok" nas labels. Só faltam os slides para que eu possa iniciar a versão final do vídeo @brendaellen @practice-uffs/con-conteudo. Isso vale para a issue #196 também. @brendaellen @practice-uffs/con-conteudo Será que seria possível, onde as professoras utilizam imagens, pegar imagens da "fonte" com uma qualidade melhor para refazer os slides? Tem algumas imagens com uma qualidade bem ruim que as professoras utilizaram para fazer os slides.. Isso também acontece na issue #196. @brendaellen @practice-uffs/con-conteudo Será que seria possível, onde as professoras utilizam imagens, pegar imagens da "fonte" com uma qualidade melhor para refazer os slides? Tem algumas imagens com uma qualidade bem ruim que as professoras utilizaram para fazer os slides.. Isso também acontece na issue #196. Alterei todas as imagens em todas a issues, para aquelas que possuíam qualidade melhor em sua origem. Os slides que continuam com as imagens de qualidade ruim, é porque não encontrei uma segunda opção melhor em minha pesquisa (principalmente as charges). @MaNegrao o link para a versão do vídeo a ser corrigido está aqui. Obs: Em torno do minuto 27:20 o vídeo começa a ficar com um pequeno problema de atualização que foi causado pela má renderização do Format Factory, isso será corrigido nas próximas versões. @robisonjose A professora mandou as alterações necessárias para esse vídeo e o 12 na issue #151.
gharchive/issue
2020-10-06T23:21:42
2025-04-01T06:45:25.747581
{ "authors": [ "MaNegrao", "alessandrapedrotti", "brendaellen", "robisonjose", "thaliafriedrich" ], "repo": "practice-uffs/programa", "url": "https://github.com/practice-uffs/programa/issues/195", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
716307781
Access violation in finalizer of PreparedSqlLiteInsertCommand I have some code that generates a complex object, then serialises it to JSON for storage in a sqlite database. Takes just over a minute to generate, then saves. I ran it and after about 80 inserts, it failed in the debugger saying that memory was corrupt. Odd, but I ran it again, and after about 110 more, same thing. System.AccessViolationException HResult=0x80004003 Message=Attempted to read or write protected memory. This is often an indication that other memory is corrupt. Source=<Cannot evaluate the exception source> StackTrace: <Cannot evaluate the exception stack trace> Fatal error. System.AccessViolationException: Attempted to read or write protected memory. This is often an indication that other memory is corrupt. at SQLitePCL.sqlite3_stmt.manual_close() at SQLite.PreparedSqlLiteInsertCommand.Finalize() I have seen issue 667 and it looks similar, but obviously not quite. On reviewing my code, I create an object for the generation, which creates a member property which is the database connection. This is created in the constructor, and used after the generation. There is no finaliser in the object that is dealing with the connection, so I guess it is garbage collected at some later time. The loop creates one of these objects, which does the work, inserts, and then it goes around again, so the original object is left to the garbage collector. I shall add database closing to the object, but I am posting this for reference as it obviously will need fixing if it can be pinned down. Using version 1.7.355 - latest. Running under Windows 10 in Visual Studio for the fail. Okay, I added IDisposable to my class, Added .Dispose to the database connection, and it has run 150 loops with no issue. So that's a workaround, even if there is an underlying issue. @mj2015 Thank you for your issue report. We were having the exact same issue, but ensuring that Dispose is called (in stead of the finalizer) does seem to be a functioning workaround :-)
gharchive/issue
2020-10-07T08:18:10
2025-04-01T06:45:25.759083
{ "authors": [ "JakobAnarkyLabs", "mj2015" ], "repo": "praeclarum/sqlite-net", "url": "https://github.com/praeclarum/sqlite-net/issues/1003", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
62306754
Feature/addCORS Added CORS To API :+1:
gharchive/pull-request
2015-03-17T06:38:04
2025-04-01T06:45:25.760028
{ "authors": [ "DanielKeun", "dbreedt" ], "repo": "praekelt/txtalert-webapi", "url": "https://github.com/praekelt/txtalert-webapi/pull/12", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2820702
Support tox Vumi should have a tox.ini, to ease testing. setup.py (and related) also needs some cleaning up, to support Vumi being installed as a normal Python package. Merged, along with general cleanups.
gharchive/issue
2012-01-12T17:15:18
2025-04-01T06:45:25.761445
{ "authors": [ "piet-praekelt" ], "repo": "praekelt/vumi", "url": "https://github.com/praekelt/vumi/issues/123", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1401781670
Create cargame.py A simple car game Please update contributing,md Update the CONTRIBUTING.md file
gharchive/pull-request
2022-10-08T01:38:38
2025-04-01T06:45:25.790786
{ "authors": [ "Rohit-cyber696", "npranavb", "pranav-nb" ], "repo": "pranav-nb/Hacktoberfest22", "url": "https://github.com/pranav-nb/Hacktoberfest22/pull/34", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
1718392035
Project : Student Enrollment form Describe the Feature you want to end I would like 1)add components on the front page in enrol section 2) Mark the sections as mandatory in order to enrol so that we can filter duplicate entries Expected behavior This way we can increase user experience and also makes the site more functional Screenshots (optional) Additional context (optional ) I would like to contribute to this site under GSSOC'23 Contributor. @pranjay-poddar Please assign this to me. Hey @SnigdhaP2022 , I would like to work on this project by using Html, Css. I'll add all the components on the front page in enrol section and will mark the sections as mandatory in order to enrol so that we can filter duplicate entries and it will help increase user experience. I would like to contribute to this site under GSSOC'23 Contributor. Thank you! I would like to work on this issue. Assign it to me
gharchive/issue
2023-05-21T06:48:21
2025-04-01T06:45:25.795304
{ "authors": [ "SnigdhaP2022", "bhoomigupta1406", "dk172923" ], "repo": "pranjay-poddar/Dev-Geeks", "url": "https://github.com/pranjay-poddar/Dev-Geeks/issues/1138", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1822322064
Changes in Advice_Generator Title and Issue number Advice_Generator earlier Now Close #4064 Checklist: [x] I have mentioned the issue number in my Pull Request. [] I have commented my code, particularly in hard-to-understand areas [] I have created a helpful and easy to understand README.md Resolve conflicts Resolved
gharchive/pull-request
2023-07-26T12:34:03
2025-04-01T06:45:25.798755
{ "authors": [ "TusharKesarwani", "akankshachanana1" ], "repo": "pranjay-poddar/Dev-Geeks", "url": "https://github.com/pranjay-poddar/Dev-Geeks/pull/4237", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2560787638
🛑 Natural Language <2> SQL Query is down In 16610d3, Natural Language <2> SQL Query (https://natural-language-2-sql-queries.onrender.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Natural Language <2> SQL Query is back up in 5bb4ec6 after 9 minutes.
gharchive/issue
2024-10-02T06:47:47
2025-04-01T06:45:27.236621
{ "authors": [ "prateekralhan" ], "repo": "prateekralhan/status", "url": "https://github.com/prateekralhan/status/issues/2090", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2114825990
🛑 Prateek Ralhan-Resume is down In 33eecd7, Prateek Ralhan-Resume (https://prateek-online-resume-render.onrender.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Prateek Ralhan-Resume is back up in 3511105 after 19 minutes.
gharchive/issue
2024-02-02T12:31:01
2025-04-01T06:45:27.239281
{ "authors": [ "prateekralhan" ], "repo": "prateekralhan/status", "url": "https://github.com/prateekralhan/status/issues/351", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2119808632
🛑 Bangalore - Realtime Weather Stats is down In 4002eb7, Bangalore - Realtime Weather Stats (https://bangalore-weather-stats.onrender.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Bangalore - Realtime Weather Stats is back up in d1b4ab2 after 5 minutes.
gharchive/issue
2024-02-06T02:24:58
2025-04-01T06:45:27.241876
{ "authors": [ "prateekralhan" ], "repo": "prateekralhan/status", "url": "https://github.com/prateekralhan/status/issues/402", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2165332532
🛑 ChatGPT Beyond 2021 is down In 5d2a480, ChatGPT Beyond 2021 (https://chatgpt-beyond-2021.onrender.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: ChatGPT Beyond 2021 is back up in 8e2f4b7 after 17 minutes.
gharchive/issue
2024-03-03T12:25:25
2025-04-01T06:45:27.244678
{ "authors": [ "prateekralhan" ], "repo": "prateekralhan/status", "url": "https://github.com/prateekralhan/status/issues/520", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2451683332
cutouts and masks are not the same in segment_weeds.py cutout masks and cutouts are not the same in terms of what pixels are being masked. This is because you're removing gray from the cutout before saving it but not the mask. make the mask and cutout the same size (the bounding box) DONE.
gharchive/issue
2024-08-06T20:58:07
2025-04-01T06:45:27.358888
{ "authors": [ "mkutu", "navjot-nangia" ], "repo": "precision-sustainable-ag/Field-AnnotationPipeline", "url": "https://github.com/precision-sustainable-ag/Field-AnnotationPipeline/issues/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
935053228
Farm YRV in KS This farm is showing 999 for the deepest sensor in treatment number one. Sensor will be replaced. At the same time the middle sensor for bare 2 has to be readdressed, looks like the depth is not input correctly. Recurring issue with depth not reading or input correctly
gharchive/issue
2021-07-01T17:13:05
2025-04-01T06:45:27.359784
{ "authors": [ "elhenriq", "saseehav" ], "repo": "precision-sustainable-ag/On-farm-Protocols", "url": "https://github.com/precision-sustainable-ag/On-farm-Protocols/issues/138", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2420884108
if LoRAX is based on punica kernels will it be able to support LoRA Adapters for Mistral NeMO 12B? Feature request if LoRAX is based on punica kernels will it be able to support LoRA Adapters for Mistral NeMO 12B? which has a vocab size > 130k. Currently Vllm for example doesn't support vocab_size > 128512 when enable_lora=True I think if huggingface and LoRAX are based on punica kernels they will also have this limitation or this limitation does not exist for TGI and LoRAX? Thank you! Motivation be able to run inference with Mistral NeMO + LoRA Adapter (in a multi-lora world) Your contribution Checked various deployment providers and found the limitation Feature request if LoRAX is based on punica kernels will it be able to support LoRA Adapters for Mistral NeMO 12B? which has a vocab size > 130k. Currently Vllm for example doesn't support vocab_size > 128512 when enable_lora=True I think if huggingface and LoRAX are based on punica kernels they will also have this limitation or this limitation does not exist for TGI and LoRAX? Thank you! Motivation be able to run inference with Mistral NeMO + LoRA Adapter (in a multi-lora world) Your contribution Checked various deployment providers and found the limitation did you figure out if Mistral Nemo 12B works with lora adapters with lorax? It does not work with VLLM or Aphrodite still and I am looking for alternatives.
gharchive/issue
2024-07-20T12:43:54
2025-04-01T06:45:27.363197
{ "authors": [ "Nero10578", "tensimixt" ], "repo": "predibase/lorax", "url": "https://github.com/predibase/lorax/issues/549", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2229354773
pip compile equivalent / trigger solve Problem description I'd like to use pixi to give me just the lock file, not install packages, similar to how I can use pip-compile or uv pip compile. However, pixi seems to generate the lock file only after a pixi add command. However, if I change the manifest file (e.g. generate), I cannot do a pixi compile/solve/lock or pixi install --dry-run to obtain a lockfile. Maybe I misunderstood how pixi works, but I expect 3 phases: add solve install pixi seems to run solve and install by default, which is great, and I can skip the install using --no-install, which is also great. However, I cannot trigger the solve phase independently. Yeah this is a more often requested feature. I though I would be able to link you to that issue but I guess it is lost in conversations. Having a pixi lock or something similar would be good indeed For documentation purposes, there is an requirement for the pypi solve to have a python interpreter available which we install in the conda ([dependencies] table) environment. Thus installing atleast part of the environment is required for the full locking of a pixi project. For example this would still need to install python to only lock: [project] name = "test" channels = ["conda-forge"] platforms = ["linux-64"] [dependencies] python = ">=3.10" [pypi-dependencies] solara = "==1.30.1" Until there is a solution to this issue https://github.com/prefix-dev/pixi/issues/1130, specifically the wheel only install. I have a similar use case, and my workaround is running pixi list or pixi tree. These are otherwise diagnostic commands that function like no-ops but still trigger a lockfile check. The main thing that can be confusing from a UX perspective is that if you haven't installed anything and no lockfile update is needed, then both of those commands succeed without printing anything to console.
gharchive/issue
2024-04-06T18:21:16
2025-04-01T06:45:27.372462
{ "authors": [ "jayqi", "maartenbreddels", "olivier-lacroix", "ruben-arts" ], "repo": "prefix-dev/pixi", "url": "https://github.com/prefix-dev/pixi/issues/1131", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
574689475
Make space delimiter an option on a per language basis Right now we have a global preference for using spaces after delimiters: let g:NERDSpaceDelims = 1 I would like to overhaul this such that each language in our array of supported file types specifies a default integer number of spaces. Example: let s:delimiterMap = { \ 'aap': { 'left': '#', 'spaces': 1 }, \} This would make it easier to manage the cases where some language have stringent style guides, others even mandate spaces for proper lexing, while many people prefer no spaces at all. Additionally it keeps our current "uncomment" code resilient to variations in these preferences, something hard coding the space into the delimiter ruins. See also #328, #278, #202, #183, #420, #96. This idea can probably be expanded to other settings such as default alignment (see e.g. this comment for a use case).
gharchive/issue
2020-03-03T14:01:48
2025-04-01T06:45:27.380773
{ "authors": [ "alerque" ], "repo": "preservim/nerdcommenter", "url": "https://github.com/preservim/nerdcommenter/issues/421", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
907337994
fix(bugs): fixed a lot of bugs Signed-off-by: cndoit18 cndoit18@outlook.com Too much, I need to find Issues for it. Can I help me together? please fixes #643 fixes #605 fixes #692 [x] I've made sure the Changelog.md will remain up-to-date after this PR is merged. Hi @cndoit18, Thank you very much for your contribution! It looks good. I have only one question and a suggestion. Sorry for the delay, I was caught in other projects. 😞 Thank you very much, I hope to bring more contributions to this project and be able to speed up the progress of the roadmap. I would like to ask what are the requirements to be the core maintainer of this project? Beside some minor comments all looks good. After you update the code I will merge this. Thanks again! I would like to ask what are the requirements to be the core maintainer of this project? We don't have yet requirements for maintainers. Please continue the good work, we would love to have more people working along us on this project. 😄 Thanks!!!! Thanks!!!! It is recommended that you turn on graceful scheduling for more stability Thanks!!!! It is recommended that you turn on graceful scheduling for more stability update mysqlcluster fail 2021-06-09T08:46:31.261593Z 0 [System] [MY-010116] [Server] /usr/sbin/mysqld (mysqld 8.0.20-11) starting as process 1 2021-06-09T08:46:31.273844Z 1 [System] [MY-013576] [InnoDB] InnoDB initialization has started. 2021-06-09T08:46:31.769019Z 1 [System] [MY-013577] [InnoDB] InnoDB initialization has ended. 2021-06-09T08:46:32.079049Z 0 [System] [MY-011323] [Server] X Plugin ready for connections. Socket: '/var/lib/mysql/mysqlx.sock' bind-address: '::' port: 33060 2021-06-09T08:46:32.200204Z 0 [Warning] [MY-010068] [Server] CA certificate ca.pem is self signed. 2021-06-09T08:46:32.263329Z 0 [Warning] [MY-010604] [Repl] Neither --relay-log nor --relay-log-index were used; so replication may break when this MySQL server acts as a slave and has his hostname changed!! Please use '--relay-log=testww-mysql-0-relay-bin' to avoid this problem. 2021-06-09T08:46:32.266324Z 0 [Warning] [MY-010533] [Repl] Error during --relay-log-recovery: Could not locate rotate event from the master. 2021-06-09T08:46:32.266607Z 0 [Warning] [MY-013504] [Repl] Server was not able to find a rotate event from master server to initialize relay log recovery for channel ''. Skipping relay log recovery for the channel. 2021-06-09T08:46:32.302295Z 0 [System] [MY-010931] [Server] /usr/sbin/mysqld: ready for connections. Version: '8.0.20-11' socket: '/var/lib/mysql/mysql.sock' port: 3306 Percona Server (GPL), Release 11, Revision 5b5a5d2. 2021-06-09T08:46:54.305007Z 30 [System] [MY-010597] [Repl] 'CHANGE MASTER TO FOR CHANNEL '' executed'. Previous state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. 2021-06-09T08:46:54.311761Z 32 [Warning] [MY-010897] [Repl] Storing MySQL user name or password information in the master info repository is not secure and is therefore not recommended. Please consider using the USER and PASSWORD connection options for START SLAVE; see the 'START SLAVE Syntax' in the MySQL Manual for more information. 2021-06-09T08:46:54.315089Z 32 [System] [MY-010562] [Repl] Slave I/O thread for channel '': connected to master 'sys_replication@testww-mysql-1.mysql.kme-mysql-10272:3306',replication started in log 'FIRST' at position 4 2021-06-09T08:46:54.319253Z 32 [ERROR] [MY-010557] [Repl] Error reading packet from server for channel '': Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been (server_errno=1236) 2021-06-09T08:46:54.319292Z 32 [ERROR] [MY-013114] [Repl] Slave I/O for channel '': Got fatal error 1236 from master when reading data from binary log: 'Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been', Error_code: MY-013114 When update MySQL, I want to set master to read-only. How about this? Thanks!!!! It is recommended that you turn on graceful scheduling for more stability update mysqlcluster fail 2021-06-09T08:46:31.261593Z 0 [System] [MY-010116] [Server] /usr/sbin/mysqld (mysqld 8.0.20-11) starting as process 1 2021-06-09T08:46:31.273844Z 1 [System] [MY-013576] [InnoDB] InnoDB initialization has started. 2021-06-09T08:46:31.769019Z 1 [System] [MY-013577] [InnoDB] InnoDB initialization has ended. 2021-06-09T08:46:32.079049Z 0 [System] [MY-011323] [Server] X Plugin ready for connections. Socket: '/var/lib/mysql/mysqlx.sock' bind-address: '::' port: 33060 2021-06-09T08:46:32.200204Z 0 [Warning] [MY-010068] [Server] CA certificate ca.pem is self signed. 2021-06-09T08:46:32.263329Z 0 [Warning] [MY-010604] [Repl] Neither --relay-log nor --relay-log-index were used; so replication may break when this MySQL server acts as a slave and has his hostname changed!! Please use '--relay-log=testww-mysql-0-relay-bin' to avoid this problem. 2021-06-09T08:46:32.266324Z 0 [Warning] [MY-010533] [Repl] Error during --relay-log-recovery: Could not locate rotate event from the master. 2021-06-09T08:46:32.266607Z 0 [Warning] [MY-013504] [Repl] Server was not able to find a rotate event from master server to initialize relay log recovery for channel ''. Skipping relay log recovery for the channel. 2021-06-09T08:46:32.302295Z 0 [System] [MY-010931] [Server] /usr/sbin/mysqld: ready for connections. Version: '8.0.20-11' socket: '/var/lib/mysql/mysql.sock' port: 3306 Percona Server (GPL), Release 11, Revision 5b5a5d2. 2021-06-09T08:46:54.305007Z 30 [System] [MY-010597] [Repl] 'CHANGE MASTER TO FOR CHANNEL '' executed'. Previous state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. 2021-06-09T08:46:54.311761Z 32 [Warning] [MY-010897] [Repl] Storing MySQL user name or password information in the master info repository is not secure and is therefore not recommended. Please consider using the USER and PASSWORD connection options for START SLAVE; see the 'START SLAVE Syntax' in the MySQL Manual for more information. 2021-06-09T08:46:54.315089Z 32 [System] [MY-010562] [Repl] Slave I/O thread for channel '': connected to master 'sys_replication@testww-mysql-1.mysql.kme-mysql-10272:3306',replication started in log 'FIRST' at position 4 2021-06-09T08:46:54.319253Z 32 [ERROR] [MY-010557] [Repl] Error reading packet from server for channel '': Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been (server_errno=1236) 2021-06-09T08:46:54.319292Z 32 [ERROR] [MY-013114] [Repl] Slave I/O for channel '': Got fatal error 1236 from master when reading data from binary log: 'Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been', Error_code: MY-013114 Did you create a new mysqlcluster? If it's old, you should have problems with dirty data, do you have slack? We can use it to chat Thanks!!!! It is recommended that you turn on graceful scheduling for more stability update mysqlcluster fail 2021-06-09T08:46:31.261593Z 0 [System] [MY-010116] [Server] /usr/sbin/mysqld (mysqld 8.0.20-11) starting as process 1 2021-06-09T08:46:31.273844Z 1 [System] [MY-013576] [InnoDB] InnoDB initialization has started. 2021-06-09T08:46:31.769019Z 1 [System] [MY-013577] [InnoDB] InnoDB initialization has ended. 2021-06-09T08:46:32.079049Z 0 [System] [MY-011323] [Server] X Plugin ready for connections. Socket: '/var/lib/mysql/mysqlx.sock' bind-address: '::' port: 33060 2021-06-09T08:46:32.200204Z 0 [Warning] [MY-010068] [Server] CA certificate ca.pem is self signed. 2021-06-09T08:46:32.263329Z 0 [Warning] [MY-010604] [Repl] Neither --relay-log nor --relay-log-index were used; so replication may break when this MySQL server acts as a slave and has his hostname changed!! Please use '--relay-log=testww-mysql-0-relay-bin' to avoid this problem. 2021-06-09T08:46:32.266324Z 0 [Warning] [MY-010533] [Repl] Error during --relay-log-recovery: Could not locate rotate event from the master. 2021-06-09T08:46:32.266607Z 0 [Warning] [MY-013504] [Repl] Server was not able to find a rotate event from master server to initialize relay log recovery for channel ''. Skipping relay log recovery for the channel. 2021-06-09T08:46:32.302295Z 0 [System] [MY-010931] [Server] /usr/sbin/mysqld: ready for connections. Version: '8.0.20-11' socket: '/var/lib/mysql/mysql.sock' port: 3306 Percona Server (GPL), Release 11, Revision 5b5a5d2. 2021-06-09T08:46:54.305007Z 30 [System] [MY-010597] [Repl] 'CHANGE MASTER TO FOR CHANNEL '' executed'. Previous state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. 2021-06-09T08:46:54.311761Z 32 [Warning] [MY-010897] [Repl] Storing MySQL user name or password information in the master info repository is not secure and is therefore not recommended. Please consider using the USER and PASSWORD connection options for START SLAVE; see the 'START SLAVE Syntax' in the MySQL Manual for more information. 2021-06-09T08:46:54.315089Z 32 [System] [MY-010562] [Repl] Slave I/O thread for channel '': connected to master 'sys_replication@testww-mysql-1.mysql.kme-mysql-10272:3306',replication started in log 'FIRST' at position 4 2021-06-09T08:46:54.319253Z 32 [ERROR] [MY-010557] [Repl] Error reading packet from server for channel '': Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been (server_errno=1236) 2021-06-09T08:46:54.319292Z 32 [ERROR] [MY-013114] [Repl] Slave I/O for channel '': Got fatal error 1236 from master when reading data from binary log: 'Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been', Error_code: MY-013114 Did you create a new mysqlcluster? If it's old, you should have problems with dirty data, do you have slack? We can use it to chat new mysqlcluster. wechart ? Thanks!!!! It is recommended that you turn on graceful scheduling for more stability update mysqlcluster fail 2021-06-09T08:46:31.261593Z 0 [System] [MY-010116] [Server] /usr/sbin/mysqld (mysqld 8.0.20-11) starting as process 1 2021-06-09T08:46:31.273844Z 1 [System] [MY-013576] [InnoDB] InnoDB initialization has started. 2021-06-09T08:46:31.769019Z 1 [System] [MY-013577] [InnoDB] InnoDB initialization has ended. 2021-06-09T08:46:32.079049Z 0 [System] [MY-011323] [Server] X Plugin ready for connections. Socket: '/var/lib/mysql/mysqlx.sock' bind-address: '::' port: 33060 2021-06-09T08:46:32.200204Z 0 [Warning] [MY-010068] [Server] CA certificate ca.pem is self signed. 2021-06-09T08:46:32.263329Z 0 [Warning] [MY-010604] [Repl] Neither --relay-log nor --relay-log-index were used; so replication may break when this MySQL server acts as a slave and has his hostname changed!! Please use '--relay-log=testww-mysql-0-relay-bin' to avoid this problem. 2021-06-09T08:46:32.266324Z 0 [Warning] [MY-010533] [Repl] Error during --relay-log-recovery: Could not locate rotate event from the master. 2021-06-09T08:46:32.266607Z 0 [Warning] [MY-013504] [Repl] Server was not able to find a rotate event from master server to initialize relay log recovery for channel ''. Skipping relay log recovery for the channel. 2021-06-09T08:46:32.302295Z 0 [System] [MY-010931] [Server] /usr/sbin/mysqld: ready for connections. Version: '8.0.20-11' socket: '/var/lib/mysql/mysql.sock' port: 3306 Percona Server (GPL), Release 11, Revision 5b5a5d2. 2021-06-09T08:46:54.305007Z 30 [System] [MY-010597] [Repl] 'CHANGE MASTER TO FOR CHANNEL '' executed'. Previous state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. 2021-06-09T08:46:54.311761Z 32 [Warning] [MY-010897] [Repl] Storing MySQL user name or password information in the master info repository is not secure and is therefore not recommended. Please consider using the USER and PASSWORD connection options for START SLAVE; see the 'START SLAVE Syntax' in the MySQL Manual for more information. 2021-06-09T08:46:54.315089Z 32 [System] [MY-010562] [Repl] Slave I/O thread for channel '': connected to master 'sys_replication@testww-mysql-1.mysql.kme-mysql-10272:3306',replication started in log 'FIRST' at position 4 2021-06-09T08:46:54.319253Z 32 [ERROR] [MY-010557] [Repl] Error reading packet from server for channel '': Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been (server_errno=1236) 2021-06-09T08:46:54.319292Z 32 [ERROR] [MY-013114] [Repl] Slave I/O for channel '': Got fatal error 1236 from master when reading data from binary log: 'Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been', Error_code: MY-013114 Did you create a new mysqlcluster? If it's old, you should have problems with dirty data, do you have slack? We can use it to chat new mysqlcluster. wechart ? Thanks!!!! It is recommended that you turn on graceful scheduling for more stability update mysqlcluster fail 2021-06-09T08:46:31.261593Z 0 [System] [MY-010116] [Server] /usr/sbin/mysqld (mysqld 8.0.20-11) starting as process 1 2021-06-09T08:46:31.273844Z 1 [System] [MY-013576] [InnoDB] InnoDB initialization has started. 2021-06-09T08:46:31.769019Z 1 [System] [MY-013577] [InnoDB] InnoDB initialization has ended. 2021-06-09T08:46:32.079049Z 0 [System] [MY-011323] [Server] X Plugin ready for connections. Socket: '/var/lib/mysql/mysqlx.sock' bind-address: '::' port: 33060 2021-06-09T08:46:32.200204Z 0 [Warning] [MY-010068] [Server] CA certificate ca.pem is self signed. 2021-06-09T08:46:32.263329Z 0 [Warning] [MY-010604] [Repl] Neither --relay-log nor --relay-log-index were used; so replication may break when this MySQL server acts as a slave and has his hostname changed!! Please use '--relay-log=testww-mysql-0-relay-bin' to avoid this problem. 2021-06-09T08:46:32.266324Z 0 [Warning] [MY-010533] [Repl] Error during --relay-log-recovery: Could not locate rotate event from the master. 2021-06-09T08:46:32.266607Z 0 [Warning] [MY-013504] [Repl] Server was not able to find a rotate event from master server to initialize relay log recovery for channel ''. Skipping relay log recovery for the channel. 2021-06-09T08:46:32.302295Z 0 [System] [MY-010931] [Server] /usr/sbin/mysqld: ready for connections. Version: '8.0.20-11' socket: '/var/lib/mysql/mysql.sock' port: 3306 Percona Server (GPL), Release 11, Revision 5b5a5d2. 2021-06-09T08:46:54.305007Z 30 [System] [MY-010597] [Repl] 'CHANGE MASTER TO FOR CHANNEL '' executed'. Previous state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. New state master_host='testww-mysql-1.mysql.kme-mysql-10272', master_port= 3306, master_log_file='', master_log_pos= 4, master_bind=''. 2021-06-09T08:46:54.311761Z 32 [Warning] [MY-010897] [Repl] Storing MySQL user name or password information in the master info repository is not secure and is therefore not recommended. Please consider using the USER and PASSWORD connection options for START SLAVE; see the 'START SLAVE Syntax' in the MySQL Manual for more information. 2021-06-09T08:46:54.315089Z 32 [System] [MY-010562] [Repl] Slave I/O thread for channel '': connected to master 'sys_replication@testww-mysql-1.mysql.kme-mysql-10272:3306',replication started in log 'FIRST' at position 4 2021-06-09T08:46:54.319253Z 32 [ERROR] [MY-010557] [Repl] Error reading packet from server for channel '': Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been (server_errno=1236) 2021-06-09T08:46:54.319292Z 32 [ERROR] [MY-013114] [Repl] Slave I/O for channel '': Got fatal error 1236 from master when reading data from binary log: 'Slave has more GTIDs than the master has, using the master's SERVER_UUID. This may indicate that the end of the binary log was truncated or that the last binary log file was lost, e.g., after a power or disk failure when sync_binlog != 1. The master may or may not have rolled back transactions that were already replicated to the slave. Suggest to replicate any transactions that master has rolled back from slave to master, and/or commit empty transactions on master to account for transactions that have been', Error_code: MY-013114 Did you create a new mysqlcluster? If it's old, you should have problems with dirty data, do you have slack? We can use it to chat new mysqlcluster. wechart ? Ok, what is yours? @cndoit18 wx234258202 Thanks @cndoit18 , I forgot to update the image of mysql-operator-sidecar , MySQLCluster is working properly.
gharchive/pull-request
2021-05-31T10:25:43
2025-04-01T06:45:27.411198
{ "authors": [ "AMecea", "cndoit18", "lizhongxuan" ], "repo": "presslabs/mysql-operator", "url": "https://github.com/presslabs/mysql-operator/pull/690", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
254789559
Turn off function-paren-newline, introduced in ESLint v4.6.0 Turn off function-paren-newline and upgrade to ESLint 4.6.0. Thanks! Released in 2.4.0.
gharchive/pull-request
2017-09-02T04:56:36
2025-04-01T06:45:27.447029
{ "authors": [ "lydell", "vanduynslagerp" ], "repo": "prettier/eslint-config-prettier", "url": "https://github.com/prettier/eslint-config-prettier/pull/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2754705906
🛑 prevhelp.com is down In 804d9ba, prevhelp.com (https://prevhelp.com) was down: HTTP code: 0 Response time: 0 ms Resolved: prevhelp.com is back up in 40c3e18 after 9 minutes.
gharchive/issue
2024-12-22T15:17:08
2025-04-01T06:45:27.483650
{ "authors": [ "ruaq" ], "repo": "prevplan/prevhelp-status", "url": "https://github.com/prevplan/prevhelp-status/issues/398", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1160655611
[MT-4616] Refactor - Selection mark to runtime decorations Refactor Extension decorate property to be a factory (not an implementation). This allows extensions to define their decorators depending on the editor instance. This also matches Plugin decorate property from Plate: https://github.com/udecode/plate/blob/deeab96a077349ac607eaeffd7bb6f4f73d60832/packages/core/src/types/plugins/Decorate.ts#L5-L13 Rewrite text selection mark from being a data model to a runtime decorator. See https://docs.slatejs.org/concepts/09-rendering#decorations
gharchive/pull-request
2022-03-06T16:35:36
2025-04-01T06:45:27.491628
{ "authors": [ "aspirisen", "e1himself" ], "repo": "prezly/slate", "url": "https://github.com/prezly/slate/pull/124", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1604520163
[#24] 운영 서버에 배포하기 위한 안정화 작업 🌱 작업 사항 #46 이어서 작업 ❓ 리뷰 포인트 🦄 관련 이슈 #24 이것도 실패..ㅜㅜ
gharchive/pull-request
2023-03-01T08:40:24
2025-04-01T06:45:27.495201
{ "authors": [ "JoosungKwon" ], "repo": "prgrms-web-devcourse/Team-Kkini-Mukvengers-BE", "url": "https://github.com/prgrms-web-devcourse/Team-Kkini-Mukvengers-BE/pull/51", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
276264495
Fix "theme" warning in of Icon component Fix "theme" warning in of Icon component: Simply add the theme' prop to the prop definition in the function signature of the Basecomponent insideIcon`'s implementation. Run npm test -- -u to update all snapshots. const Base = ({ ... theme, ... }) => { ... } Codecov Report Merging #161 into master will not change coverage. The diff coverage is 100%. @@ Coverage Diff @@ ## master #161 +/- ## ===================================== Coverage 100% 100% ===================================== Files 31 31 Lines 296 296 ===================================== Hits 296 296 Impacted Files Coverage Δ src/Icon.js 100% <100%> (ø) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 1a49b16...111fdb8. Read the comment docs.
gharchive/pull-request
2017-11-23T04:35:07
2025-04-01T06:45:27.503201
{ "authors": [ "BeniCheni", "codecov-io" ], "repo": "pricelinelabs/design-system", "url": "https://github.com/pricelinelabs/design-system/pull/161", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
482902581
Fix/#544 modal style chaining related to ticket #544 Codecov Report Merging #563 into master will not change coverage. The diff coverage is 100%. @@ Coverage Diff @@ ## master #563 +/- ## ===================================== Coverage 100% 100% ===================================== Files 55 55 Lines 567 567 ===================================== Hits 567 567 Impacted Files Coverage Δ packages/modal/src/Modal.js 100% <100%> (ø) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update c2ad5c5...b0aaec4. Read the comment docs. @bertya ✅LGTM, but I don't see an approve button anywhere.
gharchive/pull-request
2019-08-20T14:38:45
2025-04-01T06:45:27.510283
{ "authors": [ "bertya", "codecov-io", "unscsprt" ], "repo": "pricelinelabs/design-system", "url": "https://github.com/pricelinelabs/design-system/pull/563", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1099486241
Fix #683 - use https protocol instead of git protocol https://github.blog/2021-09-01-improving-git-protocol-security-github/ Builds locally, let's see if it build via CI Thanks watching it! It just started failing today as all the builds were working as of yesterday. Worked!!
gharchive/pull-request
2022-01-11T18:20:06
2025-04-01T06:45:27.514101
{ "authors": [ "blutorange", "melloware" ], "repo": "primefaces-extensions/primefaces-extensions", "url": "https://github.com/primefaces-extensions/primefaces-extensions/pull/684", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2145932021
kz { "startsWith": "Басталады", "contains": "Қамтиды", "notContains": "Қамтымайды", "endsWith": "Аяқталады", "equals": "Тең", "notEquals": "Тең емес", "noFilter": "Фильтрсіз", "filter": "Фильтр", "lt": "Кем", "lte": "Кем немесе тең", "gt": "Көбірек", "gte": "Көбірек немесе тең", "dateIs": "Күніне тең ", "dateIsNot": "Күніне тең емес", "dateBefore": "Күніне дейін", "dateAfter": "Күнінен кейін", "custom": "Өзгеше", "clear": "Жою", "apply": "Қабылдау", "matchAll": "Барлығын сәйкестендіру", "matchAny": "Кез-келгенімен сәйкестендіру", "addRule": "Шартты қосу", "removeRule": "Шартты жою", "accept": "Иә", "reject": "Жоқ", "choose": "Таңдау", "upload": "Жүктеу", "cancel": "Болдырмау", "completed": "Аяқталды", "pending": "Күтілуде", "fileSizeTypes": ["Б", "Кб", "Мб", "Гб", "Тб", "Пб", "Эб", "Зб", "Йб"], "dayNames": ["Жексенбі", "Дүйсенбі", "Сейсенбі", "Сәрсенбі", "Бейсенбі", "Жұма", "Сенбі"], "dayNamesShort": ["Жсн", "Дсн", "Сей", "Ссн", "Бсн", "Жұм", "Снб"], "dayNamesMin": ["Же", "Дү", "Сс", "Сә", "Бс", "Жұ", "Сб"], "monthNames": [ "Қаңтар", "Ақпан", "Наурыз", "Сәуір", "Мамыр", "Маусым", "Шілде", "Тамыз", "Қыркүйек", "Қазан", "Қараша", "Желтоқсан" ], "monthNamesShort": [ "Қаң", "Ақп", "Нау", "Сәу", "Мам", "Мау", "Шіл", "Там", "Қыр", "Қаз", "Қар", "Жел" ], "chooseYear": "Жылды таңдау", "chooseMonth": "Айды таңдау", "chooseDate": "Күнін таңдау", "prevDecade": "Алдыңғы онжылдық", "nextDecade": "Келесі онжылдық", "prevYear": "Алдыңғы жыл", "nextYear": "Келесі жыл", "prevMonth": "Алдыңғы ай", "nextMonth": "Келесі ай", "prevHour": "Алдыңғы уақыт", "nextHour": "Келесі уақыт", "prevMinute": "Алдыңғы минута", "nextMinute": "Келесі минута", "prevSecond": "Алдыңғы секунд ", "nextSecond": "Келесі секунд", "am": "Түске дейін", "pm": "Түстен кейін", "today": "Бүгін", "now": "Қазір", "weekHeader": "Апт.", "firstDayOfWeek": 1, "showMonthAfterYear": false, "dateFormat": "dd.mm.yy", "weak": "Жеңіл", "medium": "Қалыпты", "strong": "Күрделі", "passwordPrompt": "Құпия сөзді енгізіңіз", "emptyFilterMessage": "Нәтижесіз", "searchMessage": "{0} нәтиже қолжетімді", "selectionMessage": "{0} дене таңдалынды", "emptySelectionMessage": "Таңдалынған элементтер жоқ", "emptySearchMessage": "Нәтиже табылмады", "emptyMessage": "Қолжетімді нұсқау жоқ", "aria": { "trueLabel": "Дұрыс", "falseLabel": "Дұрыс емес", "nullLabel": "Таңдалынбады", "star": "1 жұлдыз", "stars": "{star} жұлдыз", "selectAll": "Барлық денелер таңдалынды", "unselectAll": "Барлық денелер таңдалынбады ", "close": "Жабу", "previous": "Алдыңғы", "next": "Келесі", "navigation": "Навигация", "scrollTop": "Жоғарыға жылжыту", "moveTop": "Басына жылжыту", "moveUp": "Үстіне жылжыту", "moveDown": "Астына жылжыту", "moveBottom": "Соңына жылжыту", "moveToTarget": "Қабылдағышқа жылжыту", "moveToSource": "Дереккөзге жылжыту", "moveAllToTarget": "Барлығын қабылдағышқа жылжыту", "moveAllToSource": "Барлығын дереккөзге жылжыту", "pageLabel": "{page} бет", "firstPageLabel": "Бірінші бет", "lastPageLabel": "Соңғы бет", "nextPageLabel": "Келесі бет", "previousPageLabel": "Алдыңғы бет", "rowsPerPageLabel": "Әр беттегі жолдар", "jumpToPageDropdownLabel": "Беттің ашылмалы тізіміне өту", "jumpToPageInputLabel": "Енгізілген бетке өту", "selectRow": "Жолды таңдаңыз", "unselectRow": "Жол таңдалынбады", "expandRow": "Жол кеңейтілді", "collapseRow": "Жол жиылды", "showFilterMenu": "Фильтр мәзірін көрсету", "hideFilterMenu": "Фильтр мәзірін жасыру", "filterOperator": "Фильтр операторы", "filterConstraint": "Фильтрді шектеу", "editRow": "Жол өңдеу", "saveEdit": "Өңдеуді сақтау", "cancelEdit": "Өңдеуді болдырмау", "listView": "Тізім ретінде", "gridView": "Тор ретінде", "slide": "Слайд", "slideNumber": "{slideNumber}", "zoomImage": "Суретті үлкейту", "zoomIn": "Үлкейту", "zoomOut": "Кішірейту", "rotateRight": "Оңға бұру", "rotateLeft": "Солға бұру" } } "searchMessage": "{0} results are available" Is it correct to use the word "results" here? I think the result can't be plural. The result can be positive when something is found. Or negative when nothing was found. But the word is used in the singular. There is only one result. Like Search result: 0 elements found Instead of "elements" maybe "records"? OK I added your file as kk according to this ICU Locale “Kazakh (Kazakhstan)” (kk_KZ) country is KZ but the language code is kk If you are interested in helping out PrimeFaces JSF we have some additional text that needs translating... timeOnlyTitle: "Only Time", timeText: "Time", hourText: "Hour", minuteText: "Minute", secondText: "Second", millisecondText: "Millisecond", year: "Year", month: "Month", week: "Week", day: "Day", list: "Agenda", allDayText: "All Day", moreLinkText: "More...", noEventsText: "No Events", unexpectedError: "Unexpected error", aria: { "datatable.sort.ASC": "activate to sort column ascending", "datatable.sort.DESC": "activate to sort column descending", "datatable.sort.NONE": "activate to remove sorting on column", "colorpicker.OPEN": "Open color picker", "colorpicker.CLOSE": "Close color picker", "colorpicker.CLEAR": "Clear the selected color", "colorpicker.MARKER": "Saturation: {s}. Brightness: {v}.", "colorpicker.HUESLIDER": "Hue slider", "colorpicker.ALPHASLIDER": "Opacity slider", "colorpicker.INPUT": "Color value field", "colorpicker.FORMAT": "Color format", "colorpicker.SWATCH": "Color swatch", "colorpicker.INSTRUCTION": "Saturation and brightness selector. Use up, down, left and right arrow keys to select.", "spinner.INCREASE": "Increase Value", "spinner.DECREASE": "Decrease Value", "switch.ON": "On", "switch.OFF": "Off", "messages.ERROR": "Error", "messages.FATAL": "Fatal", "messages.INFO": "Information", "messages.WARN": "Warning", }, timeOnlyTitle: "Тек қана уақыт", timeText: "Уақыт", hourText: "Сағат", minuteText: "Минут", secondText: "Секунд", millisecondText: "Милисекунд", year: "Жыл", month: "Ай", week: "Апта", day: "Күн", list: "Күн тәртібі", allDayText: "Күні бойы", moreLinkText: "Көбірек...", noEventsText: "Оқиғалар жоқ", unexpectedError: "Күтпеген қате", aria: { "datatable.sort.ASC": "Қатарды өсуі бойынша сұрыптауды қосу", "datatable.sort.DESC": "Қатарды кемуі бойынша сұрыптауды қосу", "datatable.sort.NONE": "Қатар сұрыптауын жою", "colorpicker.OPEN": "Түс таңдауғышты ашу", "colorpicker.CLOSE": "Түс таңдауғышты жабу", "colorpicker.CLEAR": "Таңдалғаң түсті өшіру", "colorpicker.MARKER": "Қанықтылық: {s}. Жарықтық: {v}.", "colorpicker.HUESLIDER": "Реңк", "colorpicker.ALPHASLIDER": "Мөлдірлік", "colorpicker.INPUT": "Түс мәні өрісі", "colorpicker.FORMAT": "Түс пішімі", "colorpicker.SWATCH": "Түс үлгісі", "colorpicker.INSTRUCTION": "Қанықтылық пен жарықтық. Енгізу үшін жоғары, төмен, солға және оңға пернелерді пайдаланыңыз", "spinner.INCREASE": "Қосу", "spinner.DECREASE": "Азайту", "switch.ON": "Қосулы", "switch.OFF": "Өшірулі", "messages.ERROR": "Қате", "messages.FATAL": "Жазмыш", "messages.INFO": "Ақпарат", "messages.WARN": "Ескерту", },
gharchive/issue
2024-02-21T06:26:00
2025-04-01T06:45:27.553037
{ "authors": [ "melloware", "zhiga90" ], "repo": "primefaces/primelocale", "url": "https://github.com/primefaces/primelocale/issues/152", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2416562290
feat(ApplePay): handle emailAddress and phoneNumber from shippingContact Description Based on: https://github.com/primer-io/primer-sdk-ios/pull/936 This update clientSession with emailAddress and mobileNumber from the Apple Pay shippingContact. Those two fields are oftenly required by ecommerce app to build a fast checkout, so they can handle those informations about customer Other Notes Other changes that are not specifically related to the intent of the PR Manual Testing Add manual testing notes here if applicable, otherwise remove this section Screenshots If applicable, otherwise remove this section Contributor Checklist [ ] All status checks have passed prior to code review [ ] I have added unit tests to a reasonable level of coverage where suitable [ ] I have added UI tests to new user flows, if applicable [ ] I have manually tested newly added UX [ ] I have open a documentation PR, if applicable Reviewer Checklist [ ] I have verified that a suitable set of automated tests has been added [ ] I have verified that the title prefix aligns to the code changes + whether a release is expected after merging the PR [ ] I have verified the documentation PR aligns with this PR, if applicable Before Merging [ ] If introducing a breaking change, I have communicated it internally [ ] Any related documentation PRs are ready to merge Other Stuff You can find out more about our automation checks here Find out more about conventional commits here Closing as handled by https://github.com/primer-io/primer-sdk-ios/pull/985
gharchive/pull-request
2024-07-18T14:27:20
2025-04-01T06:45:27.579672
{ "authors": [ "NQuinn27", "huextrat" ], "repo": "primer-io/primer-sdk-ios", "url": "https://github.com/primer-io/primer-sdk-ios/pull/947", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1722045105
🐛 [BUG] - MinimalFooter alignment for Links and Buttons Describe the bug When using MinimalFooter.Link as a and button together, the links aren't aligned correctly. In the example below the last link is rendered as button. Reproduction steps Use `MinimalFooter.Link` as `a` and `button` together. Expected behavior All links in the footer should be aligned properly. Screenshots Browsers No response OS No response Can confirm this is fixed.
gharchive/issue
2023-05-23T13:04:04
2025-04-01T06:45:27.582727
{ "authors": [ "stefankp" ], "repo": "primer/brand", "url": "https://github.com/primer/brand/issues/286", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1293987755
docker-compose up 失败 OS:MacOS Monterey 12.3.1 Docker version 20.10.10, build b485636 (base) ✘ zane@m1  ~/data/primihub   master  docker-compose up Creating network "primihub_testing_net" with driver "bridge" Pulling simple_bootstrap_node (primihub/simple-bootstrap-node:1.0)... 1.0: Pulling from primihub/simple-bootstrap-node d07403d21a2a: Pull complete Digest: sha256:3115ca655bb8a45ad2b6a83d8227b9fde272584abcc74ed814c46d413679ebed Status: Downloaded newer image for primihub/simple-bootstrap-node:1.0 Pulling node2 (primihub/primihub-node:1.0.5)... 1.0.5: Pulling from primihub/primihub-node d7bfe07ed847: Pull complete 3e832b6dc085: Pull complete 6e2baf3e12f2: Pull complete b64c6dcc8c01: Pull complete 17b8d95d5b45: Pull complete 3145e5b0507f: Pull complete 9bbf991806d5: Pull complete 06de1ebb8232: Pull complete 8ad9830833d1: Pull complete Digest: sha256:f0b11ee5ac73b22d731f6685143fc2e8ec41867a108974aab9e2b9e9203c921e Status: Downloaded newer image for primihub/primihub-node:1.0.5 Creating simple_bootstrap_node ... done Creating node2_primihub ... done Creating node1_primihub ... done Creating node0_primihub ... done Attaching to simple_bootstrap_node, node0_primihub, node1_primihub, node2_primihub simple_bootstrap_node | [*] Listening on: 0.0.0.0 with port: 4001 node0_primihub | terminate called after throwing an instance of 'YAML::BadFile' node0_primihub | what(): bad file: /app/primihub_node0.yaml node0_primihub | qemu: uncaught target signal 6 (Aborted) - core dumped node2_primihub | terminate called after throwing an instance of 'YAML::BadFile' node2_primihub | what(): bad file: /app/primihub_node2.yaml node2_primihub | qemu: uncaught target signal 6 (Aborted) - core dumped node1_primihub | terminate called after throwing an instance of 'YAML::BadFile' node1_primihub | what(): bad file: /app/primihub_node1.yaml node1_primihub | qemu: uncaught target signal 6 (Aborted) - core dumped simple_bootstrap_node | 2022-07-05T08:39:52.689Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:39:52.743Z INFO bootsrap src/main.go:64 Host created. We are:QmdSyhb8eR9dDSR5jjnRoTDBwpBCSAjT7WueKJ9cQArYoA simple_bootstrap_node | 2022-07-05T08:39:52.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:39:52.747Z INFO bootsrap src/main.go:65 [/ip4/172.28.1.13/tcp/4001 /ip4/127.0.0.1/tcp/4001] simple_bootstrap_node | simple_bootstrap_node | [*] Your Bootstrap ID Is: /ip4/0.0.0.0/tcp/4001/ipfs/QmdSyhb8eR9dDSR5jjnRoTDBwpBCSAjT7WueKJ9cQArYoA simple_bootstrap_node | simple_bootstrap_node | 2022-07-05T08:39:52.757Z INFO dht/RtRefreshManager rtrefresh/rt_refresh_manager.go:279 starting refreshing cpl 0 with key CIQAAALIVMAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA (routing table size was 0) simple_bootstrap_node | 2022-07-05T08:39:52.758Z WARN dht/RtRefreshManager rtrefresh/rt_refresh_manager.go:136 failed when refreshing routing table2 errors occurred: simple_bootstrap_node | * failed to query for self, err=failed to find any peer in table simple_bootstrap_node | * failed to refresh cpl=0, err=failed to find any peer in table simple_bootstrap_node | simple_bootstrap_node | simple_bootstrap_node | 2022-07-05T08:39:52.760Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:39:52.762Z INFO dht/RtRefreshManager rtrefresh/rt_refresh_manager.go:279 starting refreshing cpl 0 with key CIQAAACPIYAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA (routing table size was 0) simple_bootstrap_node | 2022-07-05T08:39:52.762Z WARN dht/RtRefreshManager rtrefresh/rt_refresh_manager.go:199 failed when refreshing routing table {"error": "2 errors occurred:\n\t* failed to query for self, err=failed to find any peer in table\n\t* failed to refresh cpl=0, err=failed to find any peer in table\n\n"} simple_bootstrap_node | 2022-07-05T08:39:57.749Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:02.750Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:07.751Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:12.749Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:17.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:22.746Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:27.752Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:32.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:37.751Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:42.750Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:47.749Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:52.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:40:57.750Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:02.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:07.750Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:12.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:17.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:22.746Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:27.749Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:32.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:37.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:42.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:47.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:52.743Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:41:57.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:02.749Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:07.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:12.749Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:17.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:22.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:27.751Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:32.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:37.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:42.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:47.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:52.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:42:57.748Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:02.746Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:07.746Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:12.741Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:17.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:22.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:27.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:32.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:37.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:42.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:47.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:52.740Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:43:57.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:02.743Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:07.746Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:12.743Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:17.746Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:22.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:27.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:32.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:37.747Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:42.744Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:47.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:52.740Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:44:57.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:45:02.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:45:07.745Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:45:12.742Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:45:17.740Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} simple_bootstrap_node | 2022-07-05T08:45:22.740Z DEBUG basichost basic/basic_host.go:313 failed to fetch local IPv6 address {"error": "no route found for ::"} https://github.com/primihub/primihub/blob/ca7f51940f40093c9d2aa0f1208992534b3e7138/docker-compose.yml#L3 Change primihub/pimihub-node version to latest Has it been successful? I'll close this issue.
gharchive/issue
2022-07-05T08:47:18
2025-04-01T06:45:27.590141
{ "authors": [ "LyZane", "hobo0cn" ], "repo": "primihub/primihub", "url": "https://github.com/primihub/primihub/issues/45", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
315112660
Highlighting breaks down with See title. Usually these sorts of blocks are brief, but prevents the highlighting from easily being used for all HTML files which would be nice. <style type='text/css'></style> works fine for some reason, but any content breaks it. Sorry I missed this. Doesn't seem to be a problem of my package though.
gharchive/issue
2018-04-17T15:28:12
2025-04-01T06:45:27.595668
{ "authors": [ "MattTreichelYeah", "princemaple" ], "repo": "princemaple/ngx-html-syntax", "url": "https://github.com/princemaple/ngx-html-syntax/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2169049955
My results in open-domain QA are much lower using the given checkpoint for CEPE-LLaMA-2-7B. Could you provide some insights into the potential causes for this decline? I'm curious about the discrepancies between my results (in red font) and the results presented in your paper (in black font), both obtained using the default parameters with the run_qa.sh script. Could there be any potential errors on my end that could explain these differences? Hi, thanks for your interest in our work. For CEPE at k = 10, we only use and put all the passages in the decoder model, which should match the results for LLaMA-2. There might have been a mistake in the config file, which I will look into. Are you also using the QA files from the google drive? I have tried running it multiple times, but the results are still not superior to the basic llama-2-7b model.
gharchive/issue
2024-03-05T12:14:00
2025-04-01T06:45:27.598125
{ "authors": [ "howard-yen", "sunnynexus" ], "repo": "princeton-nlp/CEPE", "url": "https://github.com/princeton-nlp/CEPE/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1291079781
Fatal Logic Error found in trainer.py in the file: https://github.com/princeton-nlp/CoFiPruning/blob/main/trainer/trainer.py line 279 sepcifies following statement : if self.start_prune: zs = self.l0_module.forward(training=True) self.fill_inputs_with_zs(zs, inputs) only when this runs, we can get the gradient for the params in self.l0_optimizer. Only when the condiction satisfied as below (line 268): if self.prepruning_finetune_steps > 0 and self.global_step == self.prepruning_finetune_steps: self.start_prune = True However, line 301 just directly update the params without checking whether the grads are ready: if self.l0_module is not None and self.l0_optimizer is not None: self.l0_optimizer.step() self.lagrangian_optimizer.step() therefore, the adamw yields bugs for beta1/beta2 referred before define in its step method. As the the grad of the params are all None, the adamw implementation will skip define the hyper-params via the self.group dict. Hi, If the gradient is None, self.l0_optimizer.step() has no effect on the set of parameters passed to the optimizer. And it will start taking effect (updating the parameters) when gradients are no longer None after starting to prune. Okay, the torch I used is 1.8.2 version, where the adamw used in trainer is defined in the following file (the step method is given in line 55): https://github.com/pytorch/pytorch/blob/v1.8.2/torch/optim/adamw.py where it continue the loop in line 77 for p in group['params']: if p.grad is None: continue but the beta1, beta2 = group['betas'] is given in line 104 that are skipped by above. However, the latest torch version has different code-flow, the bug will not happen. For compatiability and robustness, I think it is necessary to check when applying the step method via the tag I proposed in pull request. Because the flexible pruning&distill method of your research is essential necessity, many torch users of all kinds of versions will refer the code. So we need to consider for such LTS-version (e.g. torch-1.8.x). The adamw used is defined in https://github.com/pytorch/pytorch/blob/master/torch/optim/adamw.py, the step method is defined in line 108, where the statements in line 130 beta1, beta2 = group['betas'] is before 'no-grad-skip' continue statement (line 132 ). Thanks very much for your replies. Okay, the torch I used is 1.8.2 version, where the adamw used in trainer is defined in the following file (the step method is given in line 55): https://github.com/pytorch/pytorch/blob/v1.8.2/torch/optim/adamw.py where it continue the loop in line 77 for p in group['params']: if p.grad is None: continue but the beta1, beta2 = group['betas'] is given in line 104 that are skipped by above. However, the latest torch version has different code-flow, the bug will not happen. For compatiability and robustness, I think it is necessary to check when applying the step method via the tag I proposed in pull request. Because the flexible pruning&distill method of your research is essential necessity, many torch users of all kinds of versions will refer the code. So we need to consider for such LTS-version (e.g. torch-1.8.x). The adamw used is defined in https://github.com/pytorch/pytorch/blob/master/torch/optim/adamw.py, the step method is defined in line 108, where the statements in line 130 beta1, beta2 = group['betas'] is before 'no-grad-skip' continue statement (line 132 ). Thanks very much for your replies. Thanks for identifying this! I merged the fix into the main branch.
gharchive/issue
2022-07-01T08:18:30
2025-04-01T06:45:27.607164
{ "authors": [ "xiamengzhou", "zhangzhenyu13" ], "repo": "princeton-nlp/CoFiPruning", "url": "https://github.com/princeton-nlp/CoFiPruning/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
927929454
Question about parameter fixing I notice that you add the whole word embedding to the optimizer instead of just prompts? # Add word embeddings to the optimizer optimizer = AdamW([{'params': model.base_model.embeddings.word_embeddings.parameters()}], lr=args.learning_rate, correct_bias=False) Would the embedding of other words also be updated? Hi @c-box, thanks for your interest! The embeddings of other words won't be updated, as we manually set the gradients of other word embeddings to be 0 here. Oh, I missed it. Thank you very much!
gharchive/issue
2021-06-23T07:04:56
2025-04-01T06:45:27.609611
{ "authors": [ "a3616001", "c-box" ], "repo": "princeton-nlp/OptiPrompt", "url": "https://github.com/princeton-nlp/OptiPrompt/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2228247485
Existing args.yaml with different arguments When calling run.py more than one time with the same command it complains that args.yaml already exists and quits. So I have to manually delete the relevant trajectories folder. Since I couldn't get any run going in #16 due to context length limits, I assume that the code stops at this try catch block sweagent/agent/model.py try: # Perform OpenAI API call print(self.history_to_messages(history)) response = self.client.chat.completions.create( messages=self.history_to_messages(history), model=self.api_model, temperature=self.args.temperature, top_p=self.args.top_p, ) except BadRequestError as e: raise CostLimitExceededError(f"Context window ({self.model_metadata['max_context']} tokens) exceeded") Altough I dont get the Context window ({self.model_metadata['max_context']} tokens) exceeded message in the console 🤔 run 1 ... 2024-04-05 16:43:55,156 - httpx - INFO - HTTP Request: POST <REDACTED>/chat/completions?api-version=<REDACTED> "HTTP/1.1 400 model_error" WARNING Cost limit exceeded INFO 💭 THOUGHT (primary) Exit due to cost limit INFO 🎬 ACTION (primary) exit_cost INFO Saved trajectory to trajectories/<REDACTED>/<REDACTED>__kim-borgen__swe-agent-test__testtt__t-0.20__p-0.95__c-2.00__install-1/kim-borgen__swe-agent-test-i1.traj INFO Saved predictions to trajectories/<REDACTED>/<REDACTED>__kim-borgen__swe-agent-test__testtt__t-0.20__p-0.95__c-2.00__install-1/all_preds.jsonl run 2 ... INFO 💽 Loaded dataset from https://github.com/kim-borgen/swe-agent-test/issues/1 INFO 🌱 Environment Initialized WARNING ************************************************** WARNING Found existing args.yaml with different arguments! WARNING ************************************************** INFO ⏭️ Skipping existing trajectory: trajectories/<REDACTED>/<REDACTED>__kim-borgen__swe-agent-test__testtt__t-0.20__p-0.95__c-2.00__install-1/kim-borgen__swe-agent-test-i1.traj When calling run.py more than one time with the same command it complains that args.yaml already exists and quits. So I have to manually delete the relevant trajectories folder. There's a --skip_existing=False flag for this
gharchive/issue
2024-04-05T15:02:09
2025-04-01T06:45:27.614211
{ "authors": [ "kim-borgen", "klieret" ], "repo": "princeton-nlp/SWE-agent", "url": "https://github.com/princeton-nlp/SWE-agent/issues/74", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2001531433
TypeError: 'numpy._DTypeMeta' object is not subscriptable Describe the bug After the installation is completed, execute $BLENDER -noaudio --background --python generate.py -- --seed 0 --task coarse -g desert.gin simple.gin --output_folder outputs/hello_world/coarse to report an error; the error message is as follows: TypeError: 'numpy._DTypeMeta' object is not subscriptable Steps to Reproduce What version of the code were you using? ''' What is the output of git log -n 1? ''' commit a8ba86a394f8757586b4fb15252a3282e56a91de (HEAD -> main, tag: v1.0.4, origin/main, origin/HEAD) Merge: 0f208a8 359f08e Author: Alex Raistrick <araistrick@princeton.edu> Date: Mon Oct 16 15:58:16 2023 -0400 Merge pull request #87 from princeton-vl/rc_1.0.4 v1.0.4 - Pregenerated download tools, ground truth updates, render throughput improvements What command did you run? $BLENDER -noaudio --background --python generate.py -- --seed 0 --task coarse -g desert.gin simple.gin --output_folder outputs/hello_world/coarse What are your FULL output logs? Blender 3.3.1 (hash b292cfe5a936 built 2022-10-05 00:14:35) Error: Python: Traceback (most recent call last): File "/home/yzhj/data/project/infinigen/worldgen/generate.py", line 20, in <module> from terrain import Terrain File "/home/yzhj/data/project/infinigen/worldgen/terrain/__init__.py", line 7, in <module> from .core import Terrain File "/home/yzhj/data/project/infinigen/worldgen/terrain/core.py", line 16, in <module> from terrain.scene import scene, transfer_scene_info File "/home/yzhj/data/project/infinigen/worldgen/terrain/scene.py", line 10, in <module> from terrain.elements.landtiles import LandTiles, Volcanos, FloatingIce File "/home/yzhj/data/project/infinigen/worldgen/terrain/elements/landtiles.py", line 12, in <module> from terrain.assets.landtiles import assets_to_data, landtile_asset File "/home/yzhj/data/project/infinigen/worldgen/terrain/assets/landtiles/__init__.py", line 7, in <module> from .core import assets_to_data, landtile_asset File "/home/yzhj/data/project/infinigen/worldgen/terrain/assets/landtiles/core.py", line 15, in <module> from .ant_landscape import ant_landscape_asset File "/home/yzhj/data/project/infinigen/worldgen/terrain/assets/landtiles/ant_landscape.py", line 13, in <module> from terrain.land_process.snowfall import run_snowfall File "/home/yzhj/data/project/infinigen/worldgen/terrain/land_process/snowfall.py", line 10, in <module> from landlab import RasterModelGrid File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/landlab/__init__.py", line 17, in <module> from .core.model_component import Component File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/landlab/core/model_component.py", line 36, in <module> from ..field import FieldError File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/landlab/field/__init__.py", line 2, in <module> from .graph_field import GraphFields File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/landlab/field/graph_field.py", line 2, in <module> import xarray as xr File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/__init__.py", line 3, in <module> from xarray import testing, tutorial File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/testing.py", line 11, in <module> from xarray.core.coordinates import Coordinates File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/core/coordinates.py", line 16, in <module> from xarray.core.alignment import Aligner File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/core/alignment.py", line 24, in <module> from xarray.core.variable import Variable, as_compatible_data, calculate_dimensions File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/core/variable.py", line 48, in <module> from xarray.namedarray.core import NamedArray File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/namedarray/core.py", line 24, in <module> from xarray.namedarray._typing import ( File "/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/xarray/namedarray/_typing.py", line 28, in <module> _DType = TypeVar("_DType", bound=np.dtype[Any]) TypeError: 'numpy._DTypeMeta' object is not subscriptable Blender quit If this is your first time running Infinigen, what are the full output logs of install.sh ?** ########### EXT COMPILER OPTIMIZATION ########### Platform : Architecture: x64 Compiler : gcc CPU baseline : Requested : 'min' Enabled : SSE SSE2 SSE3 Flags : -msse -msse2 -msse3 Extra checks: none CPU dispatch : Requested : 'max -xop -fma4' Enabled : SSSE3 SSE41 POPCNT SSE42 AVX F16C FMA3 AVX2 AVX512F AVX512CD AVX512_KNL AVX512_KNM AVX512_SKX AVX512_CLX AVX512_CNL AVX512_ICL Generated : none CCompilerOpt.cache_flush[809] : write cache to path -> /home/yzhj/data/project/infinigen/worldgen/terrain/build/temp.linux-x86_64-3.10/ccompiler_opt_cache_ext.py /home/yzhj/data/project/infinigen /home/yzhj/data/project/infinigen/worldgen/assets/creatures/geometry/cpp_utils/setup_linux.py:1: DeprecationWarning: The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives from distutils.core import setup, Extension Compiling bnurbs.pyx because it changed. [1/1] Cythonizing bnurbs.pyx /home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/Cython/Compiler/Main.py:369: FutureWarning: Cython directive 'language_level' not set, using 2 for now (Py2). This will change in a later release! File: /home/yzhj/data/project/infinigen/worldgen/assets/creatures/geometry/cpp_utils/bnurbs.pyx tree = Parsing.p_module(s, pxd, full_module_name) warning: bnurbs.pyx:86:7: Non-trivial type declarators in shared declaration (e.g. mix of pointers and values). Each pointer declaration should be on its own line. warning: bnurbs.pyx:86:14: Non-trivial type declarators in shared declaration (e.g. mix of pointers and values). Each pointer declaration should be on its own line. warning: bnurbs.pyx:99:8: Non-trivial type declarators in shared declaration (e.g. mix of pointers and values). Each pointer declaration should be on its own line. warning: bnurbs.pyx:99:17: Non-trivial type declarators in shared declaration (e.g. mix of pointers and values). Each pointer declaration should be on its own line. running build_ext building 'bnurbs' extension creating build creating build/temp.linux-x86_64-3.10 gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -I/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/sqlite/include -I/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/bzip2/include -I/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/lzma/include -I/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/zlib/include -fPIC -fPIC -fPIC -I/home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/numpy/core/include -I/home/yzhj/data/project/infinigen/blender/3.3/python/include/python3.10 -c bnurbs.c -o build/temp.linux-x86_64-3.10/bnurbs.o In file included from /home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/numpy/core/include/numpy/ndarraytypes.h:1969, from /home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/numpy/core/include/numpy/ndarrayobject.h:12, from /home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/numpy/core/include/numpy/arrayobject.h:4, from bnurbs.c:703: /home/yzhj/data/project/infinigen/blender/3.3/python/lib/python3.10/site-packages/numpy/core/include/numpy/npy_1_7_deprecated_api.h:17:2: warning: #warning "Using deprecated NumPy API, disable it with " "#define NPY_NO_DEPRECATED_API NPY_1_7_API_VERSION" [-Wcpp] 17 | #warning "Using deprecated NumPy API, disable it with " \ | ^~~~~~~ gcc -pthread -shared -L/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/ffi/lib -L/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/sqlite/lib -L/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/bzip2/lib -L/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/lzma/lib -L/home/sybren/buildbot-builder/linux_glibc217_x86_64_cmake/build_deps/deps/Release/zlib/lib build/temp.linux-x86_64-3.10/bnurbs.o -o /home/yzhj/data/project/infinigen/worldgen/assets/creatures/geometry/cpp_utils/bnurbs.cpython-310-x86_64-linux-gnu.so /home/yzhj/data/project/infinigen Platform OS & OS Version: Ubuntu 22.04.3 LTS GPU (?) : GeForce RTX 3070 Lite Hash Rate GPU Driver Version (?) : NVIDIA-SMI 525.147.05 Driver Version: 525.147.05 CUDA Version: 12.0 RAM (GB): 32G Additional context Add any other context about the problem here. Same issue but with Osx M1 Hello, I have updated our numpy version on #176 which should fix this issue.
gharchive/issue
2023-11-20T07:22:59
2025-04-01T06:45:27.621856
{ "authors": [ "MultiTrickFox", "araistrick", "shawh-sugon" ], "repo": "princeton-vl/infinigen", "url": "https://github.com/princeton-vl/infinigen/issues/172", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
498347239
feature request: create dynamic object types (not just type fields) following up on my comment from #98 regarding dynamic object types // cc @jasonkuhrt -- currently dynamicOutputMethod is more for creating dynamic field types (t.field) and there isn't currently a standard/approach for creating dynamic objects. example use-case: I need to filter an objects' field relations by the same parent status. specifically, if I query a notebook Notebook that has published: true then for it's Notes I'd like to ensure only published ones are shown as well. this could be solved if apollo-server gave us access to a grandparent in resolvers but short of that here's code for how I'm currently handling it with nexus: // create a notebook factory using `DynamicOutputMethod` dynamicOutputMethod({   name: 'notebook',   typeDefinition: `args: { typeName: string, fieldName: string, where: { published: boolean }   }`,  factory({ typeDef: t, args }) { args.forEach(({ typeName, fieldName, where: { published } }) => { const CustomNotebook = objectType({ name: typeName, definition: t => { t.string('name') t.field('notes', { type: 'Notes', resolve: async (notebook, _, { db }) => db.Note.findMany({ where: { notebookId: notebook.id, published }, }) }) }, }) t.field(fieldName, { type: CustomNotebook, args: { slugPath: stringArg() }, resolve: (authUser, { id }, { db }) => db.Notebook.find({ where: { id, ownerId: authUser.id, published } }) }) }) }) // instantiate notebook object type as custom field definition objectType({ type: 'Profile', definition: t => { t.notebook({ typeName: 'PublishedNotebook', fieldName: 'publishedNotebook', where: { published: true } }) } }) other possible solutions I've considered (not currently supported) is nesting field definitions t.field('publishedNotebook', ({ extend: 'Notebook', definition: t => { t.field('notes', {.. }) } }) just considered I can use info.path.prev to get the grandparent type definition, which might be the easiest approach This should be possible in the latest 0.12.0-rc.4, you may want to look at https://github.com/prisma-labs/nexus-prisma to see how they utilize dynamicOutputProperty.
gharchive/issue
2019-09-25T15:05:04
2025-04-01T06:45:27.631870
{ "authors": [ "alidcastano", "tgriesser" ], "repo": "prisma-labs/nexus", "url": "https://github.com/prisma-labs/nexus/issues/229", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1280507336
QuickStart Guide: Epic improvements Our QuickStart Guide could use some attention. [ ]` https://github.com/prisma/docs/issues/2812 [ ] Add a little more info about what exactly happened; the guide is almost too quick [ ] Collect feedback from others Alex R shared this: Inspiration for the Quickstart/ Getting Started https://learn.svelte.dev/tutorial/welcome-to-svelte This is done
gharchive/issue
2022-06-22T16:39:16
2025-04-01T06:45:27.634432
{ "authors": [ "ruheni", "tanberry" ], "repo": "prisma/docs", "url": "https://github.com/prisma/docs/issues/3311", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1334805770
(update) Update WebsiteHeader with new lens version Add WebsiteHeader to layout Describe this PR Update the WebsiteHeader on docs to use the one in lens for centralisation of components. Changes Update call on layout to point to WebsiteHeader with the new give props. Any other relevant information N/A @carlagn Can you check why blogpost images are not loaded here? Opening a new PR: https://github.com/prisma/docs/pull/3636
gharchive/pull-request
2022-08-10T15:30:03
2025-04-01T06:45:27.636612
{ "authors": [ "carlagn", "nilubava" ], "repo": "prisma/docs", "url": "https://github.com/prisma/docs/pull/3589", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
665548387
Fix 2 typos This PR fixes 2 typo errors that I found recently. Thanks for taking the time! 💚
gharchive/pull-request
2020-07-25T09:02:04
2025-04-01T06:45:27.637586
{ "authors": [ "mhwelander", "zardoy" ], "repo": "prisma/docs", "url": "https://github.com/prisma/docs/pull/702", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
354037061
move aws-sdk package into devDependencies $ git clone git@github.com:prisma/graphql-yoga.git $ cd graphql-yoga $ yarn $ du -sh node_modules 165M node_modules $ du -sh node_modules/aws-lambda 30M node_modules/aws-sdk The aws-sdk package is listed as a prod dependency but it is inexistent on src folder. As you can see from the script above, it takes a lot of disk space. How open are you guys into moving it into devDependencies? I can do the PR if you want. Besides taking space, it became an issue when I discovered the AWS Lambda only allows unzipped bundles to be smaller than 250MB and aws-sdk is the biggest chunk of it. Source: https://docs.aws.amazon.com/lambda/latest/dg/limits.html (see 'AWS Lambda Deployment Limits' section). Bump. This should be a dev dependency. Right now aws-sdk is about 40% of my application's package size (and this is not an application that even runs on aws infra). Maybe I'm missing something, but experiencing this as an issue because aws-lambda is a dependency while @types/aws-lambda is a devDependency @mponizil +1 on this, the dependencies should be switched - the @types/aws-lambda should be a prod dependency and aws-lambda should be a dev dependency. This is already auto-closed, but just for everybody's reference: this is now fixed (see #576)
gharchive/issue
2018-08-25T19:21:35
2025-04-01T06:45:27.640799
{ "authors": [ "lquixada", "mmkal", "pimterry" ], "repo": "prisma/graphql-yoga", "url": "https://github.com/prisma/graphql-yoga/issues/436", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
476191319
[Preview3.2]MySQL Lift Bug When using MySQL as datasource, executing prisma2 lift save --name 'init', no error executing prisma2 lift up --verbose with error: Specified key was too long; max key length is 767 bytes, which is usually caused by long varchar, usually more than varchar(255). After investigating, I believe it's a prisma2 bug, the generated README file in my migration file states that all my id fields decored with @id @default(cuid() or uuid()) are mapping to 'id' varchar(1000) NOT NULL, which will apparently cause the error I described at the beginning. My Schema File This is fixed in alpha. Will go to stable with preview6.
gharchive/issue
2019-08-02T05:40:58
2025-04-01T06:45:27.644564
{ "authors": [ "mavilein", "zzyyzz1992" ], "repo": "prisma/lift", "url": "https://github.com/prisma/lift/issues/93", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
679783840
where query not working as expected Bug description not filter not working as doc says here await prisma.post.findMany({ where: { content: { not: { contains: "sample" }, }, }, }); Error happened PrismaClientValidationError: Invalid `prisma.post.findMany()` invocation in /home/galala/projects/prisma-examples/typescript/rest-express/src/index.ts:24:23 { where: { content: { not: { contains: 'sample' } ~~~~~~~~~~~~~~~~~~~~ } } } Argument not: Got invalid value { contains: 'sample' } on prisma.findManyPost. Provided Json, expected String or null or NullableStringFilter. How to reproduce https://github.com/mohamedGamalAbuGalala/prisma-bugs link to the sandbox Expected behavior query with not is supposed to work as expected in docs and in typedef not?: string | StringFilter | null Do you want "NOT" as boolean expression? If yes then you need an array here. Hence, the right syntax will be await prisma.post.findMany({ where: { not: [ content:{ contains: "sample" }], }, }); For more refer this. @RitikDua Your example is not working with the typedef The problem here is that typedef has the definition of not as not?: string | StringFilter | null So this example should work and it is accepted by typedef but it doesn't in runtime and it throws the error I posted in the issue description. await prisma.post.findMany({ where: { content: { not: { contains: "sample" }, }, }, }); The problem here is that typedef has the definition of not as not?: string | StringFilter | null So this example should work and it is accepted by typedef but it doesn't in runtime and it throws the error I posted in the issue description. await prisma.post.findMany({ where: { content: { not: { contains: "sample" }, }, }, }); sorry i did a little mistake there you need not as Capital letters. Hence you need it as NOT I know that NOT will work for me. @pantharshit00 what kind of info do you need? Hey @mohamedGamalAbuGalala I was waiting for you to respond to @RitikDua's comment so that is why I added that label. Also, I am unable to reproduce this with 2.6.0, the provided example works. I was having trouble running it on codesandbox so I made a repl it with working example: https://repl.it/@HarshitPant/prisma-issue-833#src/index.ts Confirming that the bug is no longer exist in 2.6.0 (sandbox)
gharchive/issue
2020-08-16T16:31:37
2025-04-01T06:45:27.653740
{ "authors": [ "RitikDua", "mohamedGamalAbuGalala", "pantharshit00" ], "repo": "prisma/prisma-client-js", "url": "https://github.com/prisma/prisma-client-js/issues/833", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1447856183
Raise an error when no openssl version is detected Prisma can only work properly in presence of an openssl library. Some "slim" Docker images do not ship SSL libraries by default, as they trade off versatility for smaller disk memory footprint. In such "slim" Docker cases (here's an example),any prisma command using the engines fails with an error similar to: Error: Unknown binaryTarget linux-arm64-openssl-undefined and no custom engine files were provided This happens e.g. when using: prisma generate prisma init prisma db push prisma format Although prisma format itself uses a Wasm module that doesn't rely on library/binary engines, every CLI command is only executed after ensureBinariesExist() from @prisma/engines is called, which in turn calls the download() function from @prisma/fetch-engines, which triggers the Unknown binaryTarget error above. Rather than silently returning an undefined openssl version number, we should explicitly raise an error stating that openssl is missing, and should be installed. In practice, the code block to modify is likely in getPlatform.ts. For additional context: this comment on Github this internal thread on Slack In practice, the code block to modify is in getPlatform.ts. Not that it might be ok and expected for some OS and platforms for this to return undefined - so we need to be careful, and possibly handle that on higher levels (getPlatform()). Closing this in favor of https://github.com/prisma/prisma/issues/17154.
gharchive/issue
2022-11-14T11:10:33
2025-04-01T06:45:27.660534
{ "authors": [ "janpio", "jkomyno" ], "repo": "prisma/prisma", "url": "https://github.com/prisma/prisma/issues/16256", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2550526189
Add Support for Transient Property in Prisma Models Problem Problem Currently, Prisma enforces that all model fields are persisted in the database. This limitation can be problematic in cases where certain properties are required only at the application layer, such as computed values or temporary data used during processing. As a result, developers are forced to handle such properties separately, leading to more complex code and a lack of direct support in Prisma. Suggested solution Introduce support for transient properties in Prisma models that are not persisted to the database. These properties should: Be defined in the Prisma schema but excluded from database migrations. Be accessible via Prisma Client, similar to regular model fields. Allow developers to handle temporary or computed data more seamlessly within the Prisma model. Alternatives An alternative is to handle these transient fields outside of Prisma by manually adding them to the objects after retrieval from the database. However, this leads to scattered logic and decreases the overall maintainability of the codebase. Additional context Transient properties are common in other ORM frameworks and would offer more flexibility to Prisma users, making it easier to manage application-specific logic without impacting the database schema. This feature will streamline development, especially for use cases involving temporary or computed data. Suggested solution Alternatives Additional context Hey @jazzybruno. Would result extensions solve your usecases? Hi, Thank you for the response! I feel that this approach is a bit more complex compared to Spring Boot's @Transient property, which offers a more user-friendly solution. It would be great if Prisma could implement something similar to streamline the process for developers. Is there any timeline for this feature to be merged once implemented? On Fri, Sep 27, 2024 at 4:45 PM Serhii Tatarintsev @.***> wrote: Hey @jazzybruno https://github.com/jazzybruno. Would result extensions https://www.prisma.io/docs/orm/prisma-client/client-extensions/result solve your usecases? — Reply to this email directly, view it on GitHub https://github.com/prisma/prisma/issues/25297#issuecomment-2379448164, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWW6N44X676PZ6LGE652V3TZYVVO5AVCNFSM6AAAAABO47IRBKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNZZGQ2DQMJWGQ . You are receiving this because you were mentioned.Message ID: @.***> -- [image: photo] DUSHIMIYIMANA Jazzy Bruno Software Developer, Rwanda Coding Academy 0790059755 | 0790059755 | jazzybruno.vercel.app/ @.*** | Kigali https://maps.google.com/?q=Kigali | Jazzy Bruno https://wa.me/Jazzy+Bruno | jazzybruno jazzybruno https://msng.link/o/?jazzybruno=tg [image: facebook] https://www.facebook.com/jazzibruno.dushimiyimana/ [image: instagram] https://www.instagram.com/bruno_jazzy/ [image: linkedin] https://www.linkedin.com/in/jazzy-bruno-b9861a225/ Create your own email signature https://www.wisestamp.com/create-own-email-signature/?utm_source=promotion&utm_medium=signature&utm_campaign=create_your_own&srcid= Before we even start to talk about timelines, we have to agree on whether or not this is feature worth having. How would the field be accessible on the model? In JPA you would do something like this: @Transient public String getFullName() { return firstName + " " + lastName; } If we introduce similar @transient attribute to Prisma schema, where would return firstName + " " + lastName; part go? You'd still have to supply it on TS side somehow and probably write a code that is very similar to existing result extensions anyway. Does having an annotation really makes things simpler in that case? I’ve been trying to use Prisma’s $extends feature to add custom computed fields to a model. Specifically, I have a Category model that has an array of Fields, each with a price. My goal is to compute a totalPrice for the Category by summing the price values from the related Fields. However, I encountered a limitation where Prisma’s $extends functionality does not allow adding arrays or objects from relations to the needs object. Is there any plan to support this use case, or is there an alternative approach that could be suggested for computing fields based on related data arrays using $extends? On Mon, Sep 30, 2024 at 3:06 PM Serhii Tatarintsev @.***> wrote: Before we even start to talk about timelines, we have to agree on whether or not this is feature worth having. How would the field be accessible on the model? In JPA you would do something like this: @Transient public String getFullName() { return firstName + " " + lastName; } If we introduce similar @transient attribute to Prisma schema, where would return firstName + " " + lastName; part go? You'd still have to supply it on TS side somehow and probably write a code that is very similar to existing result extensions anyway. Does having an annotation really makes things simpler in that case? — Reply to this email directly, view it on GitHub https://github.com/prisma/prisma/issues/25297#issuecomment-2383138966, or unsubscribe https://github.com/notifications/unsubscribe-auth/AWW6N42XZVCN4PRTTZB2OMDZZFED3AVCNFSM6AAAAABO47IRBKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGOBTGEZTQOJWGY . You are receiving this because you were mentioned.Message ID: @.***> -- [image: photo] DUSHIMIYIMANA Jazzy Bruno Software Developer, Rwanda Coding Academy 0790059755 | 0790059755 | jazzybruno.vercel.app/ @.*** | Kigali https://maps.google.com/?q=Kigali | Jazzy Bruno https://wa.me/Jazzy+Bruno | jazzybruno jazzybruno https://msng.link/o/?jazzybruno=tg [image: facebook] https://www.facebook.com/jazzibruno.dushimiyimana/ [image: instagram] https://www.instagram.com/bruno_jazzy/ [image: linkedin] https://www.linkedin.com/in/jazzy-bruno-b9861a225/ Create your own email signature https://www.wisestamp.com/create-own-email-signature/?utm_source=promotion&utm_medium=signature&utm_campaign=create_your_own&srcid= Hi @SevInf, This looks like a great feature and I believe it’s essential. Here’s my use case: I have a login API that returns a User model. Now, I want to add tokens to the User model, but I won’t be storing them in the database. It would be ideal if I could do something like this: model User { id String @id @default(uuid()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt deletedAt DateTime? // Optional username String @unique password String email String? // Optional isAdmin Boolean @default(false) balance Float @default(0) token String @transient } The usage could remain straightforward. The token doesn't always need to be computed—its initial value can simply be the default for its data type.
gharchive/issue
2024-09-26T13:04:04
2025-04-01T06:45:27.683227
{ "authors": [ "SevInf", "hexdecimal16", "jazzybruno" ], "repo": "prisma/prisma", "url": "https://github.com/prisma/prisma/issues/25297", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
565227168
[DataModel] Compound Foreign Keys and Index Definitions The @@index definition currently has to refer to datamodel fields, but in the case of compound foreign keys two or more database fields are collapsed into one datamodel field. The indexes on the other hand can refer to the underlying database fields which are not in the datamodel (apart from being in the @map). The @@unique definition theoretically has the same issue, but since we do not render it in the case of relations but instead make the relation 1:1 this issue does not surface. There is a similar problem with @@id on the Mondial MySql schema: @do4gr : I don't think this is the same thing. The casing is just wrong in this case. The field name is province and it is being referred to as Province. @mavilein I don't think it's a casing issue. The field is called province because this is a compound foreign key pointing to the model Province. Introspection therefore gave the combined field the name province. The @@id refers to the database columns Name and Province of which Province gets hidden since it is combined with Country into one relation field named province. If the spelling had been the same it would have accidentally worked. Further examples: mysql_127\AdventureWorks2014.log model SalesOrderDetail { CarrierTrackingNumber String? LineTotal Float ModifiedDate DateTime OrderQty Int rowguid String @unique SalesOrderDetailID Int @default(autoincrement()) UnitPrice Float UnitPriceDiscount Float @default(0) /// This used to be part of a relation to SalesOrderHeader SalesOrderID Int specialOfferProduct SpecialOfferProduct @map(["SpecialOfferID", "ProductID"]) @@id([SalesOrderDetailID, SalesOrderID]) @@index([SalesOrderID], name: "FK_SalesOrderDetail_SalesOrderHeader_SalesOrderID") @@index([specialOfferProduct], name: "FK_SalesOrderDetail_SpecialOfferProduct_SpecialOfferIDProductID") @@index([rowguid], name: "GUID_SalesOrderDetail") @@index([ProductID], name: "IX_SalesOrderDetail_ProductID") } mysql_127\Basketball_men.log model awards_coaches { award String? id Int @default(autoincrement()) @id lgID String? note String? coaches coaches? @map(["coachID", "year"]) @relation(references: [coachID, year]) @@index([coachID], name: "coachID") @@index([coaches], name: "coachID_2") @@index([coaches], name: "coachID_3") @@index([coaches], name: "year") @@index([coaches], name: "year_2") @@index([year, coachID, lgID], name: "year_3") } mysql_127\tpcc.log model C_Order_Line { ol_amount Float? ol_delivery_d DateTime? ol_dist_info String? ol_number Int ol_quantity Int? c_Stock C_Stock? @map(["ol_i_id", "ol_supply_w_id"]) /// This used to be part of a relation to C_Order ol_o_id Int /// This used to be part of a relation to C_Order ol_d_id Int /// This used to be part of a relation to C_Order ol_w_id Int @@id([ol_o_id, ol_d_id, ol_w_id, ol_number]) @@index([ol_i_id], name: "ol_i_id") @@index([c_Stock], name: "ol_i_id_2") @@index([ol_o_id, ol_d_id, ol_w_id], name: "ol_o_id") @@unique([ol_w_id, ol_d_id, ol_o_id, ol_number], name: "C_Order_Line_I1") } mysql_127\tpce.log model E_HOLDING { H_DTS DateTime? H_PRICE Float H_QTY Float e_HOLDING_SUMMARY E_HOLDING_SUMMARY @map(["H_CA_ID", "H_S_SYMB"]) /// This used to be part of a relation to E_TRADE H_T_ID Float @id @@unique([H_CA_ID, H_S_SYMB, H_DTS, H_T_ID], name: "E_HOLDING_INDX1") } It is interesting to note, that the mysql_127\Basketball_men.log now looks slightly different from back when @do4gr posted his screenshot. The @@index now has a lot more coaches which I think might be a bug - the order of the fields matters for indexes so we can not just replace it with the relation name. There could/should also be cases where @@id is affected by this. Possibly this is not the case, as the errors are hidden by other errors (e.g. the one mentioned before in https://github.com/prisma/prisma2/issues/1606#issuecomment-586907016) - or it really is uncommon enough that none of our schemas have that. I talked the situation through with @sorenbs and he generally agrees that being able to use the underlying column/field of course makes sense because that is what is happening. It is just awkward right now because @map is usually not used on Prisma level, but database level. The problem is, that @map here is used in a different way than we usually use it (This is not documented in the specs, but implemented and there is issue about it at https://github.com/prisma/specs/issues/345 that describes something quite similar to what is implemented.) A suggestion for a different syntax that avoid the overloading of @map: Use @relation(fields: [...]) instead of @map([...]). Actually, my last message shows a problem with my suggestion: @map explicitly refers to database columns, and uses quotes to be able to refer to names PSL does not support. My suggestion fields does not, so we might have to have an additional @map in some cases which of course is quite ugly and messy. (This is not the case in any of the schemas we have I think, but could potentially be the case of course) Unfortunately that thought also shows why using the @map values in @@index etc is "wrong" as well: Indexes should use field names, but in @map we list column names. Another iteration then: model E_HOLDING { H_DTS DateTime? H_PRICE Float H_QTY Float + e_HOLDING_SUMMARY E_HOLDING_SUMMARY @map(["H_CA_ID", "H_S_SYMB"]) - e_HOLDING_SUMMARY E_HOLDING_SUMMARY @relation(fields: [H_CA_ID @map("_H_CA_ID"), H_S_SYMB @map("9H_S_SYMB")]) /// This used to be part of a relation to E_TRADE H_T_ID Float @id @@unique([H_CA_ID, H_S_SYMB, H_DTS, H_T_ID], name: "E_HOLDING_INDX1") } This example assumes the columns are called _H_CA_ID and 9H_S_SYMB which is unsupported because of the _ and the 9 at the beginning of the string. That requires an @map for the field. But as they are part of a relation, the renamed field is put into @relation(fields ...). The @@unique can the refer to the field name that is defined inside fields, but the engine can still know about the real column name via the @map. I think this is the right approach. There are quite a few things to consider: Prisma schema primitives. Almost everything in the schema should operate on or reference prisma schema primitives. I believe @map is the only exception. Therefore, the current implementation is too simplistic, and Jan's suggestion is an improvement. It introduces an extra layer, treating the fields that make up a compound relation as first class schema elements. Name conflicts? Assume one of the columns in the compound relation has the same name as the relation field. We would have to disallow this: model E_HOLDING { H_DTS DateTime? H_PRICE Float H_QTY Float + e_HOLDING_SUMMARY E_HOLDING_SUMMARY @map(["e_HOLDING_SUMMARY", "H_S_SYMB"]) - e_HOLDING_SUMMARY E_HOLDING_SUMMARY @relation(fields: [e_HOLDING_SUMMARY , H_S_SYMB @map("9H_S_SYMB")]) /// This used to be part of a relation to E_TRADE H_T_ID Float @id @@unique([H_CA_ID, H_S_SYMB, H_DTS, H_T_ID], name: "E_HOLDING_INDX1") } Terminology We now have two kinds of fields related to relations. What do we call them? relation field and sub relation field? model E_HOLDING { relation_field E_HOLDING_SUMMARY @relation(fields: [sub_relation_field @map("relation_column")]) } Exposing sub relation fields in client In the future we could choose to expose the sub relation fields in the client. This would resolve the request to include sub relation fields in the default selection set. We would probably want them to default to being read-only. We might want to also support filtering on them. because they share the namespace with normal fields, we don't need to wrap them or treat them specially in any way, which is nice. Syntax for @map on sub relation fields This would be the first time we allow nested attributes: model E_HOLDING { relation_field E_HOLDING_SUMMARY @relation(fields: [sub_relation_field @map("relation_column")]) } An alternative syntax could be: model E_HOLDING { relation_field E_HOLDING_SUMMARY @relation(fields: [sub_relation_field]) @map(["relation_column"]) } This works relatively well for relations with one column, but gets complicated Must the sub relation fields be explicitly spelled out in all cases? For relations with more than one column, probably yes. For relations Assume one of the columns in the compound relation has the same name as the relation field. We would have to disallow this: We would probably comment something out as a guardrail here. We now have two kinds of fields related to relations. What do we call them? relation field and sub relation field? relation sub field? In the future we could choose to expose the sub relation fields in the client. Let's consider this separately. An alternative syntax could be: ... That is much cleaner and I prefer it over mine. This works relatively well for relations with one column, but gets complicated with multiple columns: ... That is ok, as this really is an uncommon case and it still follow sound logic. For relations with just a single column, the following simplification could be allowed when the relation field and relation column has the same name: ... turns into ... Yes, in my understanding the relation would be named by the relation field anyway by default so no need to specify this again. If you want to customize this though, you will have to use the syntax. Indexes (and any other schema construct for that matter) must reference the fields, not the underlying columns. Correct. We could allow a shortcut to reference the relation field in addition to the sub relation field: Yes and no. It is probably much easier to standardize on the relation sub field name. Using the relation field name would not be explicit for compound indexes as the order of the columns in the index is not clearly spelled out in the index, just by looking at the relation definition which might change. Great. Seems like we have agreement on all points. (I agree with you on the last point, no reason to complicate things!) This works relatively well for relations with one column, but gets complicated with multiple columns: model E_HOLDING { relation_field E_HOLDING_SUMMARY @relation(fields: [sub_relation_field1, sub_relation_field1]) @map("relation_column1", "relation_column2") } In this case what happens if only some fields need to be renamed? For example the first and the last out of three. One would need to map all of them in this case right? @do4gr - One would need to map all of them, yes. I would like to suggest one further simplification to the suggested syntax. To me the current syntax proposal feels very heavy for those reasons: What's the purpose of the name of the sub relation field when it does not exposed in the client? E.g. for @relation(fields: [sub_relation_field]) @map(["relation_column"]) the name seems to be irrelevant. The syntax seems very dense. Name conflicts with other fields are very hard to spot. With this approach we have 3 types of fields: scalar, relation and sub relation fields. I would like to suggest an approach with which we could stick to the two types scalar and relation field. My suggestion is to turn relation fields into a completely virtual concept. Here are the rules i have in mind: Every column of an underlying table gets mapped to a simple scalar field. Even those ones that are part of a foreign key and therefore represent a relation. Relation Fields can be constructed by building upon those scalar fields. This way relation fields become very similar to foreign keys in SQL. A foreign key like FOREIGN KEY (a, b) REFERENCES OtherTable(col1, col2) is turned into relationField OtherTable @relation(fields: [a,b], references: [col1, col2]. Scalar fields that are used within a relation field become read only. (as already suggested by @sorenbs earlier) Here's a more detailed example: SQL schema CREATE TABLE post ( id serial NOT NULL, author_first_name text NOT NULL, author_last_name text NOT NULL, PRIMARY KEY (id), FOREIGN KEY (author_first_name, author_last_name) REFERENCES user(first_name, last_name) ); CREATE TABLE user ( id serial NOT NULL, first_name text NOT NULL, last_name text NOT NULL, PRIMARY KEY (id), UNIQUE(first_name, last_name) ); corresponding Prisma schema model Post { id Int @id authorFirstName String @map("author_first_name") authorLastName String @map("author_last_name") editorId Int @map("editor_id") // a relation field referencing the compound unique of `User` author User @relation(fields: [authorFirstName, authorLastName], references: [firstName, lastName]) // a relation field referencing the id of `User` editor User @relation(fields: [editorId], references: [id]) @@map("post") } model User { id Int @id firstName String @map("first_name") lastName String @map("last_name") @@unique([firstName, lastName]) @@map("user") } Should be fixed by https://github.com/prisma/prisma-engines/pull/621
gharchive/issue
2020-02-14T10:09:33
2025-04-01T06:45:27.713897
{ "authors": [ "do4gr", "janpio", "mavilein", "sorenbs" ], "repo": "prisma/prisma2", "url": "https://github.com/prisma/prisma2/issues/1606", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2558952160
feat: integrate contract registries Integrate contract registries into the interface. To fix: [x] Selected count doesn't change after approval in ApplicationsToApprove.tsx @kittybest can you help with this please? To finish: [ ] Project search [x] Ballot [x] Results re #263 Closing in favour of a new PR due to rebase not being the best option
gharchive/pull-request
2024-10-01T11:30:01
2025-04-01T06:45:27.726353
{ "authors": [ "ctrlc03" ], "repo": "privacy-scaling-explorations/maci-platform", "url": "https://github.com/privacy-scaling-explorations/maci-platform/pull/370", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2055000466
feat: prefer-arrow-functions I re-implemented the prefer-arrow-functions from scratch to handle more edge-cases and introduce minimal fixes. :tada: This PR is included in version 1.0.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2023-12-24T06:27:54
2025-04-01T06:45:27.752643
{ "authors": [ "privatenumber" ], "repo": "privatenumber/eslint-config", "url": "https://github.com/privatenumber/eslint-config/pull/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1137965920
Create TileCatalog class in place of using dictionaries This creates a TileCatalog class, which subclasses the dictionary. It adds additional validation and checking over a standard dictionary. Moreover, many functions that were solely concerned with converting tile catalogs into full catalogs can be made methods of this new class. Thanks for working on this @dereklhansen ! This is a great idea, feel free to ping me for a review if you feel that might be useful. Will likely wait to work on this until #458 is finished, since the data structure is still evolving [ ] Place data structures for TileCatalog and FullCatalog in bliss/catalog.py since they are more widely applicable than just the LocationEncoder.py @ismael-mendoza I know this is a big PR, but I think it would be good if you could skim over the changes. In particular, the changes in reporting.py should give you a good feel for the new FullCatalog class. Most of the assertions there are unnecessary now since they are checked upon initializing the FullCatalog. Thanks @dereklhansen - I'll take a look now :)
gharchive/pull-request
2022-02-14T23:21:19
2025-04-01T06:45:27.755580
{ "authors": [ "dereklhansen", "ismael-mendoza" ], "repo": "prob-ml/bliss", "url": "https://github.com/prob-ml/bliss/pull/432", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
953477247
Equation 9.4 Ver: June 5, 2021 Equation 9.4 is missing a constant which is to represent the normalization constant for the denominator p(x;theta). More of a mathematical correctness for using "=" instead of a "proportionality" as the constant is independet of class C fixed
gharchive/issue
2021-07-27T04:00:29
2025-04-01T06:45:27.759545
{ "authors": [ "murphyk", "nsanghi" ], "repo": "probml/pml-book", "url": "https://github.com/probml/pml-book/issues/152", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1217545845
Page 490, typo "We can do this be defining". by defining? fixed
gharchive/issue
2022-04-27T15:49:02
2025-04-01T06:45:27.760559
{ "authors": [ "boreshkinai", "murphyk" ], "repo": "probml/pml-book", "url": "https://github.com/probml/pml-book/issues/330", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
193541838
Use spiffs to store scenes etc. Let's use spiffs to store scenes etc. https://github.com/sticilface/Esp8266-Hue/tree/spiffs_branch is using spiffs. What's the status of this - definitely a good idea... To my knowledge no one has started implementing this so far. I will take a look at it. I've modded the main sketch quite a bit as I'm using a different neopixel library. Shouldn't matter as I would probably use another sketch that works with your lightservice classes Whatever you are doing, if the result is that it works better with your changes than before, I'd ask you to submit your changes as pull requests so that we can all benefit. Ideally you don't wait until many changes have accumulated, but send one separate PR per improvement at a time, as is best practice in any open source project. Thank you for your great work! Hi @Quanghoster any news on this? Sadly the sticilface rep is buggy (causes imediate hue app crash for me) and not maintained anymore looks like. I've been tied up with other work lately but will start spending a bit of time on this over the next few days. I'm thinking we have simply store the json as files, load on boot and save on change... OK, some progress. I now have spiffs support to store bulb groups which I have tested with the hue switcher app. I am basically saving each group slot as a file and loading them into the existing array on start up, if they exist. I've also made a start on spiffs support for scenes with the same approach the scene is being saved to the file system but I'm not seeing them listed in the app yet. I think there may be another Json format issue for the list scenes function which I hope to investigate this evening. Once I have groups and scenes working with spiffs I'll post a PR. Need some thoughts on how to handle formatting of fs initially... Is scenes working for anyone? I can create a scene which gets stored in a file with my code changes, but they aren't getting listed in the app when added. I've done some comparison with a bridge and see no difference in the json format so am not quite sure why this doesn't work for me. Here's what I'm getting back from an /api/user REST command, after adding a scene with the app. { "lights": { "1": { "type": "Extended color light", "manufacturername": "OpenSource", "swversion": "0.1", "name": "Hue LightStrips 1", "uniqueid": "2C:3A:E8:06:54:FB-1", "modelid": "LST001", "state": { "on": false, "bri": 0, "xy": [ 0.00000, 0.00000 ], "colormode": "hs", "effect": "none", "ct": 500, "hue": 0, "sat": 0, "alert": "none", "reachable": true } }, "2": { "type": "Extended color light", "manufacturername": "OpenSource", "swversion": "0.1", "name": "Hue LightStrips 2", "uniqueid": "2C:3A:E8:06:54:FB-2", "modelid": "LST001", "state": { "on": false, "bri": 0, "xy": [ 0.00000, 0.00000 ], "colormode": "hs", "effect": "none", "ct": 500, "hue": 0, "sat": 0, "alert": "none", "reachable": true } } }, "groups": { "1": { "name": "G1", "lights": [ "1", "2" ] } }, "config": { "name": "hue emulator", "swversion": "81012917", "bridgeid": "2C3AE8FFFE0654FB", "portalservices": false, "linkbutton": true, "mac": "2C:3A:E8:06:54:FB", "dhcp": true, "ipaddress": "192.168.1.25", "netmask": "255.255.255.0", "gateway": "192.168.1.1", "apiversion": "1.3.0", "timezone": "Europe/London", "whitelist": { "api": { "name": "clientname#devicename" } }, "swupdate": { "text": "", "notify": false, "updatestate": 0, "url": "" } }, "schedules": {}, "scenes": { "": { "name": "S1", "lights": [ "1", "2" ], "owner": "api", "recycle": false, "locked": false, "appdata": { "version": "1", "data": "" }, "picture": "", "lastupdated": "2017-11-04T10:17:15", "version": 2 }, "": { "name": "S2", "lights": [ "1", "2" ], "owner": "api", "recycle": false, "locked": false, "appdata": { "version": "1", "data": "" }, "picture": "", "lastupdated": "2017-11-04T10:17:15", "version": 2 } }, "rules": {}, "sensors": {}, "resourcelinks": {} } I don't think I have implemented scenes, at least not close to fully. Ah ok - it's not too far off as far as I can see. I just noticed that the index isn't being saved for each scene in the json above so will fix that. Your code is otherwise saving scenes and returning them. If I can fix this I can look at a PR for spiffs very soon Cool, looking forward to it.
gharchive/issue
2016-12-05T16:17:52
2025-04-01T06:45:27.777983
{ "authors": [ "Quanghoster", "mcer12", "probonopd" ], "repo": "probonopd/ESP8266HueEmulator", "url": "https://github.com/probonopd/ESP8266HueEmulator/issues/53", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1593249492
test(demo-monitoring): simplify design to do a SVG screenshot DISCLAIMER: this is a POC to create an SVG export of the demo to then add it to the PA website. Please don't merge this PR. It use the internal tools described in https://github.com/process-analytics/bpmn-visualization-tools/pull/18 Use the query parameters to display the related use-case and data-type. first url: https://cdn.statically.io/gh/process-analytics/bpmn-visualization-examples/070cfc9/demo/monitoring-all-process-instances/index.html?useCase=frequency&dataType=both final url: https://cdn.statically.io/gh/process-analytics/bpmn-visualization-examples/43c01e8/demo/monitoring-all-process-instances/index.html?useCase=frequency&dataType=both Best rendering when exportint as SVG is --width 2660 --height 1160 or larger otherwise the stroke width of edges is too large. Screenshots The page only displays the diagram that fit the whole page. https://github.com/process-analytics/process-analytics.dev/pull/899 has been merged, so closing this issue. The resources will stay available with statically.io 😄
gharchive/pull-request
2023-02-21T10:47:33
2025-04-01T06:45:27.864697
{ "authors": [ "tbouffard" ], "repo": "process-analytics/bpmn-visualization-examples", "url": "https://github.com/process-analytics/bpmn-visualization-examples/pull/464", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1857978919
feat(demo): better explain how to use the PathResolver demo Also improve the design of the page: use a flexible height for the bpmn-container to better adapt to various screen resolutions index page: move bpmn-visualization-examples live environments in the "available demos" section for consistency Development: simplify explanation about commands to start the demo while developing. ✔️ Tested with the surge preview environment
gharchive/pull-request
2023-08-20T05:53:21
2025-04-01T06:45:27.866525
{ "authors": [ "tbouffard" ], "repo": "process-analytics/bv-experimental-add-ons", "url": "https://github.com/process-analytics/bv-experimental-add-ons/pull/61", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1041622718
prevent compiler errors on SDK 10.0.19041.0 Libs prevent compiler errors in Virual Studio (Win10) by definition: PROCESS_MITIGATION_REDIRECTION_TRUST_POLICY on Windows 10 Systems with stable SDK : 10.0.19041.0. Missing in winnt.h OS < Win11: + typedef struct _PROCESS_MITIGATION_REDIRECTION_TRUST_POLICY { + __C89_NAMELESS union { + DWORD Flags; + __C89_NAMELESS struct { + DWORD EnforceRedirectionTrust : 1; + DWORD AuditRedirectionTrust : 1; + DWORD ReservedFlags : 30; + }; + }; + } PROCESS_MITIGATION_REDIRECTION_TRUST_POLICY, *PPROCESS_MITIGATION_REDIRECTION_TRUST_POLICY; Thanks for your fix. Could we upgrade CustomBuildTool to support VS 2022/Win 11 with latest SDK as well (CustomBuildTool/Utils.cs)? it would be better. support VS 2022 It has to actually be released. The directory paths and registry keys are all preview names. stable SDK : 10.0.19041.0 We shouldn't support outdated SDKs since these changes will just be added then deleted over and over? Thanks for your fix. Could we upgrade CustomBuildTool to support VS 2022/Win 11 with latest SDK as well (CustomBuildTool/Utils.cs)? it would be better. VS 2022 is released today. Do we have plan to support? There is already a fix with latest commits. There is already a fix with latest commits. Wow, so quick. 👍 Fixed with the latest SDK 👍 I struggled with this for an embarassingly long time, so I would like to add a footnote for other noobs like me: The "latest SDK" that @dmex mentions is Windows 10 SDK (10.0.20348.0). Install it from Visual Studio installer. I struggled with this for an embarassingly long time, so I would like to add a footnote for other noobs like me: The "latest SDK" that @dmex mentions is Windows 10 SDK (10.0.20348.0). Install it from Visual Studio installer. @Ishmaeel a tiny addition to your perfect answer # roll-back to vs2019 in all projects at once find ./processhacker -type f -iname '*.vcxproj' -exec sed -i 's|v143|v142|g' '{}' +; # disable spectre-mitigated libraries in all projects at once find ./processhacker -type f -iname '*.vcxproj' -exec sed -i 's|<SpectreMitigation>Spectre</SpectreMitigation>|<SpectreMitigation>false</SpectreMitigation>|g' '{}' +;
gharchive/pull-request
2021-11-01T20:53:38
2025-04-01T06:45:27.877410
{ "authors": [ "Biswa96", "Ishmaeel", "avoidik", "dmex", "suuhm", "xwnb" ], "repo": "processhacker/processhacker", "url": "https://github.com/processhacker/processhacker/pull/1024", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
115551166
loadJSON errors not making it to the IDE console Here's an example: var weather; function preload() { var url = 'http://api.openweathermap.org/data/2.5/weather?q=London,UK&APPID=XXXXXXX'; weather = loadJSON(url); } function setup() { } No error in the IDE, but I do see in the node webkit browser: Strange! Must be an issue with how I'm intercepting the error messages. Will investigate.
gharchive/issue
2015-11-06T17:29:21
2025-04-01T06:45:27.879271
{ "authors": [ "antiboredom", "shiffman" ], "repo": "processing/p5.js-editor", "url": "https://github.com/processing/p5.js-editor/issues/190", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
607002922
Fixed the getblob & http links in p5.SoundFile Fixes #456 The getblob and http links had an extra "/docs" in the href due to which they both gave a 404 error. This PR fixes it. Thanks ! @montoyamoraga Always happy to contribute 😊. Hoping to contribute more under GSOC ( if selected )
gharchive/pull-request
2020-04-26T12:39:54
2025-04-01T06:45:27.880890
{ "authors": [ "Rahulm2310" ], "repo": "processing/p5.js-sound", "url": "https://github.com/processing/p5.js-sound/pull/455", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
66399942
fixed the Particle System & Forces examples I fixed the Particle System and Forces examples. They are broken because they are using get() but the method name has been changed to copy(). I found commit 338b468 of p5.js which gave me the answer. thanks!!
gharchive/pull-request
2015-04-05T04:20:36
2025-04-01T06:45:27.893574
{ "authors": [ "lmccart", "zacharystenger" ], "repo": "processing/p5.js-website", "url": "https://github.com/processing/p5.js-website/pull/88", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2085615488
chore: upgrade backstage from 1.20.3 to 1.22.1 Checklist: [x] I have updated the necessary documentation [x] have signed off all my commits as required by DCO [x] My build is green Hey 👋🏼 we will need a patch changeset since we are bumping dependencies. Hello, done 👍🏾. I also took the opportunity to apply the latest Backstage patch update
gharchive/pull-request
2024-01-17T07:53:53
2025-04-01T06:45:27.958932
{ "authors": [ "gaelgoth", "kurtaking" ], "repo": "procore-oss/backstage-plugin-announcements", "url": "https://github.com/procore-oss/backstage-plugin-announcements/pull/184", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
440313338
Document creation of multiple channels per connection Docs: https://gvolpe.github.io/fs2-rabbit/connection-channel.html PR: https://github.com/gvolpe/fs2-rabbit/pull/187 Done in #195
gharchive/issue
2019-05-04T11:27:51
2025-04-01T06:45:28.014945
{ "authors": [ "gvolpe" ], "repo": "profunktor/fs2-rabbit", "url": "https://github.com/profunktor/fs2-rabbit/issues/194", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2361230408
🛑 Multi Hardware is down In 6d946b5, Multi Hardware (https://www.multi-hardware.com/retrogami) was down: HTTP code: 500 Response time: 589 ms Resolved: Multi Hardware is back up in 3caa86b after 9 minutes.
gharchive/issue
2024-06-19T04:13:45
2025-04-01T06:45:28.140041
{ "authors": [ "proger-su" ], "repo": "proger-su/upptime", "url": "https://github.com/proger-su/upptime/issues/782", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
180498345
Update app.js The new version of Oxford Dictionnaries has changed the link, please update this lovely Chrome Extension Thanks @OSryx for the help!
gharchive/pull-request
2016-10-02T06:35:06
2025-04-01T06:45:28.155068
{ "authors": [ "OSryx", "progmonster" ], "repo": "progmonster/oxford-dictionary-search.chrome-extensions", "url": "https://github.com/progmonster/oxford-dictionary-search.chrome-extensions/pull/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1129056622
Fix test_suite so access control is off or ACLs are installed Operations in test_suite will fail once Access Control is enforced (issue #14454) so either "real" access control can't be enabled for the test suite, or the test suite needs to install ACLs so the test operations pass, or some combination of both (as appropriate). I don't know enough about how the test suite is set up and run and what is being tested to have an opinion on disable access control vs. install ACLs. FYI I went through most of the tests today and didn't see others that need fixing, but I'm waiting for CI to confirm, as I do have some hangs when running so it's difficult to be certain. Redux on this one: spent about a week fixing tests (unit tests and YAML tests) and also adding a YAML test for access control cluster (already have unit tests for system module) discovered several glitches with the test system, YAML, codegen etc. which got resolved tweaked logging so it appears on more devices/setups (not all have "detail" level) added ACL to some test setups (e.g. group messaging YAML tests) so they could succeed refactored some testing configuration so it uses a "permissive" access control (if it is testing other things) and so that setup doesn't share/bleed from one test suite to another (an issue on some platforms/setups)
gharchive/issue
2022-02-09T21:50:51
2025-04-01T06:45:28.171898
{ "authors": [ "mlepage-google" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/14985", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1185957484
[TC-IDM-6.2] Chip-tool needs to provide a way to set optional Path Field "IsUrgent flag" for a particular event path in the EventPathIB. Test-Case ID:- [TC-IDM-6.2] Events Subscribe Interaction from TH to DUT Issue As per Test-Plan test step 9 says "Activate Event subscription from TH to DUT and change the attribute value to create subscribed event before minimum interval with 'IsUrgent' flag specified for a particular event path in the EventPathIB". Chip-tool needs to provide a way to set optional Path Field "IsUrgent flag" for a particular event path in the EventPathIB. PFA Test-Plan screenshot:- Spec screenshot:- @Rajashreekalmane This is not a chip-tool issue. You need to trigger an event that is "urgent". So the event will contains the IsUrgent flag.
gharchive/issue
2022-03-30T06:34:15
2025-04-01T06:45:28.175384
{ "authors": [ "Rajashreekalmane", "vivien-apple" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/16816", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1313412137
./chip-tool basic write node-label '"newnode"' 1 0 fails Problem Test case [TC-BINFO-2.1] uses commands like ./chip-tool basic write node-label '"newnode"' 1 0 to try to write to readonly attributes which have string type (this is a 'negative' test) The command as given does not work and yields an error message. (For the integer attributes it works as expected) Proposed Solution provide an alternative that works in chip-tool so we can execute these tests update - this is depending on the quotes which is VERY hard to see in the verification steps document
gharchive/issue
2022-07-21T15:08:06
2025-04-01T06:45:28.177564
{ "authors": [ "leorozendaal" ], "repo": "project-chip/connectedhomeip", "url": "https://github.com/project-chip/connectedhomeip/issues/21052", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }