added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:34:47.177375
2022-11-28T17:08:35
1466774092
{ "authors": [ "gdalle", "mmcmanus1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8734", "repo": "mmcmanus1/HashBlotto.jl", "url": "https://github.com/mmcmanus1/HashBlotto.jl/issues/6" }
gharchive/issue
Make printing optional / use logging instead https://github.com/mmcmanus1/HashBlotto.jl/blob/7d629ded147f02d7ef63fabea604640404dde890/src/greed_test.jl#L19 Some users don't want their code to print messages, or they might want to log stuff into a txt file instead. You can make it optional with a kwarg or use https://docs.julialang.org/en/v1/stdlib/Logging/ Thanks!!
2025-04-01T04:34:47.178660
2017-02-14T21:49:39
207646669
{ "authors": [ "Jacob-Lane", "colatkinson" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8735", "repo": "mmetro/WeirdSideofYouTube", "url": "https://github.com/mmetro/WeirdSideofYouTube/pull/60" }
gharchive/pull-request
Update packages and improve package.json The package.json hadn't been updated from the template it was taken from. I updated the dependencies to the latest versions, and changed the metadata to be about this project. I also removed the method-override package, since we weren't actually using it. Looks good!
2025-04-01T04:34:47.180837
2022-01-07T03:47:04
1095954155
{ "authors": [ "mmgoodnow", "newadventure079" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8736", "repo": "mmgoodnow/cross-seed", "url": "https://github.com/mmgoodnow/cross-seed/issues/187" }
gharchive/issue
All cross-seed to search based on Jackett tags, type, or test status You can search in Jackett for private trackers only, for example, and Jackett will append &filter=type%3Aprivate to its search string. It'd be nice if we could have cross-seed use this functionality and query based on test status (passed, failed), tracker type (private, public, semi-private) and indexers that have tags associated with them Any functionality specific to Jackett is not going to be worked on as of https://github.com/mmgoodnow/cross-seed/pull/179
2025-04-01T04:34:47.184778
2015-04-03T13:04:53
66142957
{ "authors": [ "cordoval", "mmoreram" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8737", "repo": "mmoreram/SimpleDoctrineMapping", "url": "https://github.com/mmoreram/SimpleDoctrineMapping/pull/18" }
gharchive/pull-request
fix docblock for mapping compiler pass methods Q A Bug Fix? n New Feature? n BC Breaks? n Deprecations? n Tests Pass? n Fixed Tickets License MIT Doc PR Hmmm, @cordoval @hacfi Why this change? Enabled should be always a bool, never a string... acording to the code it has double usage Absolutely... you can pass a boolean or a parameter name ;) Thanks! Good point! :D
2025-04-01T04:34:47.202668
2023-12-07T23:33:45
2031685146
{ "authors": [ "drammock", "hoechenberger", "larsoner" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8738", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/issues/12276" }
gharchive/issue
non-interactive brain figure Description of the problem brain windows are not interactable, don't respond to their close buttons, and crash sometimes when calling brain.close() Steps to reproduce import mne data_path = mne.datasets.sample.data_path() subjects_dir = data_path / "subjects" sample_dir = data_path / "MEG" / "sample" brain_kwargs = dict(alpha=0.1, background="white", cortex="low_contrast") brain = mne.viz.Brain("sample", subjects_dir=subjects_dir, **brain_kwargs) Link to data No response Expected results an interactive window. Actual results a non-interactable window showing a brain that is zoomed in to the level where it shows about 3 gyri. Closing the window manually doesn't work ("the program isn't responding, wait or terminate?"). However: $ mamba list pyvistaqt # packages in environment at /opt/mambaforge/envs/mnedev: # # Name Version Build Channel pyvistaqt 0.11.0 pyhd8ed1ab_0 conda-forge I can't replicate at least on macOS arm64 with: ├☑ pyvista 0.43.dev0 (OpenGL 4.1 Metal - 88 via Apple M1 Pro) ├☑ pyvistaqt 0.12.0.dev55+gee4bbf9 ├☑ vtk 9.2.6 ├☑ qtpy 2.4.0 (PyQt5=5.15.8) Window shows up fine and I can interact: And brain.close is okay. I should be able to test on Linux tomorrow. Notice that pyvistaqt is marked as version 0.0.0. However: If you use build 1 instead of build 0 of pyvistaqt 0.11 on conda-forge the version should be fixed, see https://github.com/conda-forge/pyvistaqt-feedstock/pull/18 From a quick look at https://app.circleci.com/pipelines/github/mne-tools/mne-python/22194/workflows/a09f0e0f-0303-4214-b67e-14537775d0b6/jobs/62159 which passed and the resulting doc build things are at least not totally broken there. But CircleCI uses PyQt6 I think. Might be worth seeing if install PyQt6 from pip fixes things for you, that would tell us we have some PyQt5 compat problem I can't replicate on Linux with PyQt6 6.6.0 or PyQt5 5.15.10 or 5.15.8 :( Works for me on macOS with Qt5 something seems to be deeply wrong with my system. I can't get it to work with pyqt6 nor pyside6 either: $ QT_API=pyqt6 ipython In [1]: import mne ...: ...: data_path = mne.datasets.sample.data_path() ...: subjects_dir = data_path / "subjects" ...: sample_dir = data_path / "MEG" / "sample" ...: brain_kwargs = dict(alpha=0.1, background="white", cortex="low_contrast") ...: brain = mne.viz.Brain("sample", subjects_dir=subjects_dir, **brain_kwargs) Using pyvistaqt 3d backend. malloc(): invalid size (unsorted) Aborted (core dumped) $ QT_API=pyside6 ipython In [1]: import mne ...: ...: data_path = mne.datasets.sample.data_path() ...: subjects_dir = data_path / "subjects" ...: sample_dir = data_path / "MEG" / "sample" ...: brain_kwargs = dict(alpha=0.1, background="white", cortex="low_contrast") ...: brain = mne.viz.Brain("sample", subjects_dir=subjects_dir, **brain_kwargs) Using pyvistaqt 3d backend. Installed qt6 event loop hook. In [2]: Killed with QT_API=pyside6 I got the same behavior as with pyqt5: namely the image was way zoomed in and non-interactive, and the window had to be force-killed as its close button didn't work. :scream: Maybe you could try with a Ubuntu-python-based virtualenv instead of conda, just to rule out some badness there? And if that by some miracle works, try a fresh conda env after conda deactivateing your current env, etc. doing this: $ /usr/bin/python3 -m venv mnetest $ source mnetest/bin/activate $ pip install mne[dev] followed by running the MWE in the PR description, yields this: RuntimeError: Could not load any valid 3D backend pyvistaqt: No Qt bindings could be found notebook: No module named 'ipyevents' install pyvistaqt, using pip or conda: 'pip install pyvistaqt' 'conda install -c conda-forge pyvistaqt' or install ipywidgets, if using a notebook backend 'pip install ipywidgets' 'conda install -c conda-forge ipywidgets' So our pip install mne[dev] apparently doesn't install either pyvistaqt nor ipywidgets? IIRC it installs pyqt but does not specify a Qt binding so you also need for example pip install PyQt6 ok, I've got things working again. Here's what I did: download our environment.yml file, and remove pyqt!=5.15.3,!=5.15.4 and replace it with pip: - PyQt6 a fresh conda env based on that file worked. We should probably add ipyevents to the full variant, WDYT? We should probably add ipyevents to the full variant, WDYT? it's in there, I just made the mistake of installing pip install mne[dev] instead of mne[full,dev]. Remember when I said I found it unintuitive/confusing if the dev variant didn't install everything? I wasn't joking 😆 I find it not great either Can we not change it?
2025-04-01T04:34:47.206161
2015-07-14T10:05:57
94911264
{ "authors": [ "Eric89GXL", "dengemann", "mainakjas" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8739", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/issues/2289" }
gharchive/issue
document related functions by populating "See Also" We should list related function in the documentation of all functions a la numpy/scipy. cc @dengemann @agramfort That's kind of what our main api page does, right? Maybe we can just make it better. We should start populating See Also in docstrings, which would also help. I think I meant populating the See Also :) I'll edit the description of the PR I am taking this. YEAHH great call!
2025-04-01T04:34:47.209053
2018-03-03T12:51:38
301995973
{ "authors": [ "agramfort", "dokato", "drammock" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8740", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/issues/4982" }
gharchive/issue
mne anonymize command we should have the command equivalent of $ mne_anonymize usage: mne_anonymize [options] Remove subject information from a fif file. --his remove the HIS ID tag as well --file name the file to modify. --help print this info. --version print version info. I'm happy to take it, if no one started. go for it. it would be nice to allow be able to operate inplace too but fine to do it in a subsequent PR. okay, but please help me understand --his parameter. Since in mne.io.meas_info.anonymize_info we remove whole subject_info key so his_id as well by default. Should I modify anonymize_info function to preserve this depending on some his flag? removing it all the time is good. Forget the --his param. this was closed in #6892
2025-04-01T04:34:47.216558
2015-09-03T17:42:49
104747216
{ "authors": [ "Eric89GXL", "agramfort", "wmvanvliet" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8741", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/pull/2452" }
gharchive/pull-request
MRG: Check info consistency Closes #2218. Added some consistency checking of info regarding bads and found a few minor bugs along the way. CIs are happy, ready for review/merge from my end. Sorry, but this is a -1 from me. Calling _check_info every time an info object is used is bad design and clutters the code. Moreover, it doesn't solve the problem of allowing inconsistencies in the Info object. This would do nothing to protect the user from making mistakes in their own code. An actual solution to the problem would be to make Info immutable. Every time you make a change (through an update() function or _setitem__ for example), a new instance of Info is created through it's constructor, which is where all sanity checking is performed. Or don't solve the problem and leave it be. We could just make sure that we use the proper pick_channels family of functions consistently throughout the MNE code and don't rely on things like [info['ch_names'].index(ch) for ch in channel_names_list]. See discussion of the relevant issue, we can't practically make Info immutable because of how users have been using it. I think this is the best option we have available. Yes it clutters the code a bit and is not 100% DRY, but it should help avoid errors (and even exposed some in our code). To me those tradeoffs are worth the extra bit of clutter. I should say that if you can flesh out the immutable idea a bit and take care of the examples mentioned in the issue above am open to the idea, but I don't think we can do it in practice. A competing PR would be the best way to demonstrate that such an approach works for our code base and examples without breaking anything... And FWIW while this won't prevent users from making a mistake in the first place (i.e., adding a bad channel that doesn't exist), it should prevent it from hurting them because as soon as it would matter / be used, they should get an error. So in that sense it is a step forward from before, too. @wmvanvliet I am open to alternative options as @Eric89GXL but this already fixes quite some bugs as shown by the diff @agramfort done with the last couple of commits I included a couple of minor docstring fixes I noticed we needed @wmvanvliet actually pick_channels can't be used because it doesn't take info as an argument :( Pushed another commit with some unification/completion of pick_info, which wasn't quite complete before (only dealt with comps in the raw-writing function, nowhere else). besides +1 for merge I think the idea is that people shouldn't ever manually mess with their info structures, other than via the public functions. I agree that it should not be necessary to mess with the info dictionary directly, mostly because its such a PITA to do so. @agramfort done thanks @Eric89GXL
2025-04-01T04:34:47.220984
2017-08-10T07:00:33
249251017
{ "authors": [ "Eric89GXL", "agramfort", "codecov-io", "jdue", "jona-sassenhagen" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8742", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/pull/4479" }
gharchive/pull-request
Fix to retain xlabel when plotting evoked using spatial_colors=True This concerns #4472. Codecov Report Merging #4479 into master will decrease coverage by 2.45%. The diff coverage is 100%. @@ Coverage Diff @@ ## master #4479 +/- ## ========================================== - Coverage 83.55% 81.09% -2.46% ========================================== Files 349 349 Lines 64848 64849 +1 Branches 10026 10027 +1 ========================================== - Hits 54182 52588 -1594 - Misses 7835 9377 +1542 - Partials 2831 2884 +53 LGTM Thanks for fix @jdue ! Merged with a minor cosmit https://github.com/mne-tools/mne-python/commit/2428dddd397421732d52a735579bd54c5054b39a
2025-04-01T04:34:47.243686
2020-08-11T18:19:58
677092989
{ "authors": [ "GuillaumeFavelier", "agramfort", "larsoner" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8743", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/pull/8116" }
gharchive/pull-request
WIP: Recreate our helmet graphic I tried to recreate our helmet: Using Mayavi I get this output for the new example: If I switch to PyVista, I get: Problems thus far: The coincident topology is not resolved for mayavi or PyVista (if you rotate the scene, PyVista also cuts off parts of the yellow). Perhaps if plot_evoked_field could set the polygon_offset it could be fixed? The coincident topology for the field lines is wrong for PyVista -- they should be on top. The camera systems are not equivalent between PyVista and Mayavi. @GuillaumeFavelier can you look into these when you get a chance? Feel free to push directly here I use a fixed width for the tubes, it would be better to find a better value automatically. The coincident topology still needs to be fixed. Indeed it is getting closer: For coindicent topology, can we have something like polygon_offset=0 in some mesh call? Then it sets the appropriate Mayavi and PyVista properties under the hood (use polygon offset resolution + set polygon offset)? Assuming it fixes the problem then it's okay to make use of this in the example, and we can use it in _Brain, too. (I've noticed some coincident topology problems there, too, from time to time when rotating a brain.) We probably need to enable the polygon offset for all meshes for it to work (?). I added polygon_offset in mesh() and surface() and updated the docstrings. It also replaces _resolve_coincident_topology in _Brain. I'm not sure what is the best strategy to resolve those issues. For now, I just set some fixed values that give the best result based on what I observe. We probably need to enable the polygon offset for all meshes for it to work (?) That I don't know. I did some experiments locally and it does not seem necessary but it may be in the future once we have a definitive solution. Also about Mayavi, I did not find any alternative to SetRelativeCoincidentTopologyPolygonOffsetParameters. Hera are some relevant default values that I found: ... 'resolve_coincident_topology': 'polygon_offset', 'resolve_coincident_topology_': 1, 'resolve_coincident_topology_polygon_offset_faces': 1, 'resolve_coincident_topology_z_shift': 0.01, ... This is the helmet I have with the new changes: By letting the contour filter detect the range of scalar values, I get the same result: Can you try playing with the Mayavi parameters window (leftmost icon in the toolbar) to see if you can find the right button to tweak there to fix the coincident topology? If you can't get it to work I can try. Should we add a parameter for the line width of the contours? It looks like mayavi defaults to 1 and PyVista doesn't I can try For PyVista these are tubes now, I will render again using just the lines. To tweak that, there is the width parameter. This is what I use: # And the field lines on top renderer.contour(surface=surf, scalars=data, contours=n_contours, vmin=-vmax, vmax=vmax, opacity=alpha, colormap=colormap_lines, kind='tube', width=0.0005) Do the lines also require a polygon_offset tweak? (The tubes presumably do not because they actually protrude from the surface; the lines OTOH will be coincident) Seems like it does not matter it it's just a screenshot. The difference is more visible when you interact with the scene: line tube Do the lines also require a polygon_offset tweak? Apparently they don't the modification. Okay, feel free to choose either one based on which one you think looks better / more like the original. I guess the only remaining issue is the Mayavi coincident topology, right? I guess the only remaining issue is the Mayavi coincident topology, right? Yes indeed I have pretty much the same mapper options: I tried z-shift but there is some noise remaining near the cell edges: The best solution I found was to "Force translucent" the white surface (that way, the colors blend): Result: pip-pre errors/segfaults are https://github.com/numpy/numpy/issues/17198 and can be ignored for now Update after #8182 : Mayavi PyVista Also we should consider making PyVista the default backend. It's an item of #7162 Also we should consider making PyVista the default backend. It's an item of #7162 https://github.com/mne-tools/mne-python/issues/7162 +10 ! can you start a PR for this? @agramfort I opened https://github.com/mne-tools/mne-python/pull/8220 :) ok merge when green then ! I pushed a commit to tweak the distance too: Before After WDYT @larsoner ? perfect on my side ! Thanks @GuillaumeFavelier !
2025-04-01T04:34:47.248141
2020-11-30T17:04:55
753648511
{ "authors": [ "cbrnr", "larsoner" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8744", "repo": "mne-tools/mne-python", "url": "https://github.com/mne-tools/mne-python/pull/8584" }
gharchive/pull-request
Detect/correct invalid HP/LP settings in EDF/BDF/GDF Fixes #8582. EDF/BDF/GDF files can contain invalid LP/HP filter settings (cutoff frequencies for HP can be greater than the one for LP). This PR (1) issues a warning and (2) sets HP/LP values for affected channels to None. @apospraf loading files from BCI Competition IV 2B works as expected with this PR. NumPy installation fails in one job? Please go ahead and add a test, I won't have time until tomorrow. I don't think the current approach quite works properly, for example it doesn't handle '0.0' and '0.000' because they are strings >>> '0.0' < '0.000' True Moreover sometimes edf_info['highpass'] can have length zero and edf_info['lowpass'] can have length n_channels, which fails the NumPy < comparison on latest NumPy dev. It seems safer to deal with these being reversed when actually setting info['highpass'] and info['lowpass'], but maybe not. I'm pushing a commit with this approach and a test that asserts it works for the monkeypatch code at least. @cbrnr all good now? Yes, perfect!
2025-04-01T04:34:47.249688
2019-05-26T06:58:32
448535670
{ "authors": [ "mnogueron", "n1ru4l" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8745", "repo": "mnogueron/react-easy-panzoom", "url": "https://github.com/mnogueron/react-easy-panzoom/pull/5" }
gharchive/pull-request
feat: add disableDoubleClickZoom property closes #4 Adds a new property disableDoubleClickZoom to the PanZoom component that allows disabling the doubleClick zoom. I also added a new storybook example for testing this feature. @n1ru4l Thank you for your PR! It's neat, I'll merge it 👍
2025-04-01T04:34:47.264500
2021-10-08T02:41:56
1020621999
{ "authors": [ "zenhack" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8746", "repo": "mnutt/davros", "url": "https://github.com/mnutt/davros/issues/140" }
gharchive/issue
Drag & dropping directories into davros creates empty regular files I just discovered that if you try to add a directory to Davros by dragging & dropping it from the file manager, Davros creates an empty, regular file by the same name, instead of a directory. I can reproduce this as of 0.31.1 It looks like this is an issue with the upstream file upload library that davros is using: https://github.com/adopted-ember-addons/ember-file-upload/issues/2
2025-04-01T04:34:47.312444
2014-12-11T23:34:38
51753014
{ "authors": [ "donnielrt", "hora" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8747", "repo": "mobify/hijax", "url": "https://github.com/mobify/hijax/pull/7" }
gharchive/pull-request
Adds onerror proxy option Status: Opened for visibility Reviewers: @donnielrt Changes adds a proxy function for XHR.onerror To do [ ] write tests [ ] update README [ ] bump version number [ ] build dist @hora lemme know if you want to fast track this! :) We can sit down next week and move this along. Closing for now, since the stakeholder's not available anymore. We can reopen this if anybody needs it!
2025-04-01T04:34:47.351860
2017-03-21T18:23:16
215834750
{ "authors": [ "leader22", "mweststrate", "yeze322" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8751", "repo": "mobxjs/mobx-react", "url": "https://github.com/mobxjs/mobx-react/issues/224" }
gharchive/issue
How to make the store injected by Provider read-only? Assuming we are developing a website with login function. We have a store called userStore which is used to storage userId, userName, etc. The userStore is passed through <Provider> like this: <Provider store={userStore}> <Router> <TabA /> <TabB /> </Router> </Provider> What I want is, only share userStore among tabs but forbid modifying its property in some children such as TabA @inject('store') @observer class TabA extends Component { render () { // want to forbid operations like this <button onClick={() => {this.store.userName = 'newUserName'}}>Change global store</button> } } I'm wondering if MobX provides any native solutions to implement a read-only store, such as @injectReadonly('store') or should I use other libraries like immutable.js to do this? Thanks :) How about useStrict()? Use action()(or @action) with useStrict(true). https://github.com/mobxjs/mobx/blob/gh-pages/docs/refguide/api.md#usestrict https://mobx.js.org/refguide/action.html Besides strict mode Mobx doesn't offer a build in abstraction to make something read only. intercept can be used for that, but does not make something deeply read-only. If you are using typescript, it is easy though; just expose your store as interface with readonly properties and the compiler will do it for you.
2025-04-01T04:34:47.354132
2017-12-11T20:23:42
281155104
{ "authors": [ "RafalFilipek", "mweststrate" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8752", "repo": "mobxjs/mobx-react", "url": "https://github.com/mobxjs/mobx-react/pull/376" }
gharchive/pull-request
Update Observer component definition Hi! Right now we have to wrap observerd value with HTML tag like: const C => <Observer>{() => <div>{obser.vable}</div>}</Observer> With this change you can just return observed value const C => <Observer>{() => obser.vable}</Observer> Thank you! Great, thanks! @FredyC yeah, I want to release if #403 is in. If that is not today, I'll probably cut a release at the end of the day anyway
2025-04-01T04:34:47.532263
2019-06-29T05:29:53
462267695
{ "authors": [ "ParkSB", "coveralls", "outsideris" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8753", "repo": "mochajs/mocha", "url": "https://github.com/mochajs/mocha/pull/3962" }
gharchive/pull-request
Add prefix list- to --interfaces and --reporters options Description of the Change Rename --interfaces to --list-interfaces. Rename --reporters to --list-reporters. Add --reporters alias for --reporter. Alternate Designs N/A Benefits It makes guessing a behavior of the options easier. Possible Drawbacks N/A Applicable issues Closes #3777 Coverage increased (+0.07%) to 92.487% when pulling 3f407eee2c30d6c3349640bd22d73f2fb94ead7d on ParkSB:issue/3777 into 5d4dd98747637d0e7ed3007328ec9627dd7eda41 on mochajs:master. Changing document should be included into this PR.
2025-04-01T04:34:47.565366
2024-03-26T23:01:39
2209492684
{ "authors": [ "charlesfrye", "ekzhang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8754", "repo": "modal-labs/modal-examples", "url": "https://github.com/modal-labs/modal-examples/pull/669" }
gharchive/pull-request
remove many old examples, move some into misc with reduced QoS Some of the examples are getting a bit dated. Additionally, we'd like to focus on a smaller set of high quality examples so that we can be sure folks' first experience of Modal is good without creating too high of a maintenance burden. This PR removes a number of dated or superseded examples. It further moves several of them into the misc folder, which now has a README that identifies it as a set of examples with lower QoS. It also removes misc from the oversight of the synthetic monitoring system. Type of Change [ ] New example [x] Example updates (Bug fixes, new features, etc.) [x] Other (changes to the codebase, but not to examples) It also removes misc from the oversight of the synthetic monitoring system. Great idea! Sounds like it would help us focus on monitoring examples that matter that are more widely used. converting to draft to prevent accidental merge before the frontend is ready
2025-04-01T04:34:47.575648
2023-05-29T19:11:23
1731149631
{ "authors": [ "celinval", "tautschnig" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8755", "repo": "model-checking/kani", "url": "https://github.com/model-checking/kani/pull/2486" }
gharchive/pull-request
Simplify compiler queries structure (internal) Description of changes: Move the query database out of its own package. Remove UserInput trait for now. Remove unnecessary functions. Resolved issues: N/A Related RFC: Call-outs: The query database was initially introduced to eventually help us to communicate between all Kani components, but it became just a compiler internal communication mechanism. Thus, move it to be inside compiler. Testing: How is this change tested? No user behavior changes is expected. Is this a refactor change? Yes Checklist [ ] Each commit message has a non-empty body, explaining why the change was made [ ] Methods or procedures are documented [ ] Regression or unit tests are included, or existing tests cover the modified code [ ] My PR is restricted to a single feature or bugfix By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 and MIT licenses. @celinval This one requires a rebase or local conflict resolution (given the version and dependency information has changed in the meantime). @celinval This one requires a rebase or local conflict resolution (given the version and dependency information has changed in the meantime). Nevermind, I was able to push to your branch. @tautschnig thank you!
2025-04-01T04:34:47.578175
2017-01-15T11:11:33
200862567
{ "authors": [ "beutlich", "dietmarw", "modelica-trac-importer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8756", "repo": "modelica/Modelica", "url": "https://github.com/modelica/Modelica/issues/2108" }
gharchive/issue
MSL uses ModelicaReference Reported by beutlich on 11 Nov 2016 13:30 UTC Should the MSL have an uses annotation on ModelicaReference since there are modelica URIs inside MSL referring to it? Migrated-From: https://trac.modelica.org/Modelica/ticket/2108 Question is if the URI dependency in the documentation is enough to have a uses clause. So you could argue that technically the MSL does not depend on Modelica Reference to function the documentation might be targeting some dead links. This is currently not an issue since Modelica Reference is shipped side by side with Modelica. So I like to postpone this until we have a modular version of Modelica Reference. Not a relevant issue.
2025-04-01T04:34:47.581092
2017-01-14T12:04:46
200799441
{ "authors": [ "modelica-trac-importer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8757", "repo": "modelica/Modelica", "url": "https://github.com/modelica/Modelica/issues/875" }
gharchive/issue
Physical quantity "Angles" nowhere defined Reported by anonymous on 15 Oct 2012 08:22 UTC The physical quantity Angles is used (8 times) in Modelica.Mechanics.MultiBody.Sensors but it is never defined. It should be fixed to Angle. Migrated-From: https://trac.modelica.org/Modelica/ticket/875 Comment by dietmarw on 15 Oct 2012 08:49 UTC FIXED in e5e8c0a800240b74cac7180f7b6b230770408d2d (although I could only find 4 occurrences). Modified by dietmarw on 15 Oct 2012 08:49 UTC Comment by anonymous on 15 Oct 2012 09:02 UTC Yes, 4 occurrences was right. Thank you for the quick fix.
2025-04-01T04:34:47.615243
2015-04-10T16:41:26
67634288
{ "authors": [ "arinhouck", "mikepack" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8759", "repo": "modeset/teaspoon", "url": "https://github.com/modeset/teaspoon/issues/345" }
gharchive/issue
Random Test Failures Rails 4.1.0 Teaspoon ~> 0.9.1 Using Qunit Ember JS I have manually tested over and over in my teaspoon environment with same data after failures. The coverage of operations in these tests are functional. However at random, our teaspoon environment will send a failure on a couple individual tests within a specific folder or pass them all. This also happens when using codeship as well. I know you can't really track down the issue, but even providing some advice or possible problems that could cause these failures would be helpful. I'm sure lots of developers ran into these issues before. Thank you, Arin Hi Arin. My gut tells me it is not Teaspoon that is causing the random failures, but the tests within your application. If Teaspoon reported a failure and you're seeing a failure in your suite, Teaspoon has done its job at that point and the test is truly failing. If you're seeing a dropped or missing test, then it could be Teaspoon. Random failures are often due to the order in which the tests run. When trying to debug them, I usually look for some global state that could be corrupt when the tests run in a specific order. Things like storing data on classes instead of instances, or on the prototype that is shared between objects, or some underlying persistent store, like Local Storage. Can you confirm that you are seeing a test failure and not a dropped test? Are other tests passing? Closing. Will happily reopen if you can provide some insight as to whether the tests are truly dropped or just failing.
2025-04-01T04:34:47.627244
2021-01-15T14:44:02
786948357
{ "authors": [ "anton-malakhov", "pyrito" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8760", "repo": "modin-project/modin", "url": "https://github.com/modin-project/modin/issues/2621" }
gharchive/issue
Get omniscidb compilation phase out of critical path System information OS Platform and Distribution (e.g., Linux Ubuntu 16.04): all Modin installed from (source or binary): all Modin version: future Python version: all supported Exact command to reproduce: Describe the problem This is NYCTaxi Q4 repeated 10 times: i.e. query compilation is serial and is on critical path. There should be one or another way to hide or eliminate it: introduce explicit query compilation API (with parametrization) so Modin could control when and what to compile, executing only previously compiled queries implement pipeline using TBB FlowGraph or/and implement better caching on the omnisci side only What is the status of this now?
2025-04-01T04:34:47.637822
2018-10-06T23:05:57
367503713
{ "authors": [ "AmplabJenkins", "devin-petersohn", "simon-mo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8761", "repo": "modin-project/modin", "url": "https://github.com/modin-project/modin/pull/115" }
gharchive/pull-request
[WIP] Freeing ObjectID when python frees RemotePartition objects. cc @robertnishihara What do these changes do? Attempt to free the remote object when python garbage collects the wrapper object. Currently I am not seeing the memory get freed, but I may be doing something wrong. Related issue number #98 [x] passes git diff upstream/master -u -- "*.py" | flake8 --diff Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin/36/ Test FAILed. Merged build finished. Test FAILed. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin/126/ Test FAILed. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin/165/ Test FAILed. Merged build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin/168/ Test FAILed. Build finished. Test FAILed. Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin-Performance-Testing/3/ Test FAILed. I will keep working on this. In particular: [ ] verify setting Plasma Store's free cache size to 0 will actually reduce the memory footprint. [ ] make ray's initialization of plasma store configurable: https://github.com/ray-project/ray/blob/29c182d44908100f5fd56be90b14ad7eb7482a88/python/ray/worker.py#L2037 [ ] once the PR above is merged, set cache size to 0 on Modin's init. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin-Performance-PRB/7/ Test PASSed. Merged build finished. Test PASSed. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Modin-PRB/70/ Test PASSed.
2025-04-01T04:34:47.641720
2022-02-21T21:32:59
1146243858
{ "authors": [ "Garra1980", "YarShev" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8762", "repo": "modin-project/modin", "url": "https://github.com/modin-project/modin/pull/4240" }
gharchive/pull-request
FIX-#4197: Remove redundant warnings Signed-off-by: izamyati<EMAIL_ADDRESS> What do these changes do? [x] commit message follows format outlined here [ ] passes flake8 modin/ asv_bench/benchmarks scripts/doc_checker.py [ ] passes black --check modin/ asv_bench/benchmarks scripts/doc_checker.py [x] signed commit with git commit -s [x] Resolves #4197 [ ] tests added and passing [ ] module layout described at docs/development/architecture.rst is up-to-date [ ] added (Issue Number: PR title (PR Number)) and github username to release notes for next major release @Garra1980, I forgot to mention that we needed to add a release note. Please, do that as part of a new PR.
2025-04-01T04:34:47.646533
2019-08-29T20:22:25
487158374
{ "authors": [ "deadok22", "ovesh", "shoenig" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8763", "repo": "modprox/mp", "url": "https://github.com/modprox/mp/issues/172" }
gharchive/issue
Incorrect import of minimock in go.mod file Figure out where the incorrect import is coming from and get rid of it [p1 modprox (master)] $ grep minimock go.mod github.com/gojuno/minimock v3.0.0 // indirect github.com/gojuno/minimock/v3 v3.0.1 I can't reproduce this. What's the problem exactly? @ovesh github.com/gojuno/minimock v3.0.0 is incorrect. The module name should be github.com/gojuno/minimock/v3.
2025-04-01T04:34:47.651449
2016-08-22T03:29:28
172363214
{ "authors": [ "Aiyubi", "TristanHM" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8764", "repo": "modrzew/pokeminer", "url": "https://github.com/modrzew/pokeminer/pull/265" }
gharchive/pull-request
Multiple Grids for 0.5.4 Adapted the multiple grids code with some debugging from #94, full credit to @traversc Code really needs optimization but I'm not familiar enough with python. why are there label and discord change in this pullrequest? Because I'm a newb and didn't think changes after the request would be included. I should have made a separate branch, sorry.
2025-04-01T04:34:47.676190
2024-01-26T03:47:35
2101553291
{ "authors": [ "2heal1", "ScriptedAlchemy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8765", "repo": "module-federation/universe", "url": "https://github.com/module-federation/universe/issues/2026" }
gharchive/issue
No remotes runtime code inject while there is no async code Describe the bug When users use AsyncBoundaryPlugin , there will not use async entry // bootstrap.tsx import('./index.tsx') And if the project not use dynamic import as well , the webpack runtimeRequirementInTree will not have ensureChunkHandlers , so the remotes runtime module will not be added The reproduce steps: # install deps npm install # build npm run build Then see the dist/main.js , it will not find .f.remotes . But if remove this comment , use dynamic import , and execute npm run build , the dist will have .f.remotes Reproduction https://github.com/2heal1/webpack-demo/tree/chore/normalize-mf Used Package Manager npm System Info System: OS: macOS 13.0.1 CPU: (10) arm64 Apple M1 Pro Memory: 151.36 MB / 32.00 GB Shell: 5.8.1 - /bin/zsh Binaries: Node: 16.17.0 - ~/.nvm/versions/node/v16.17.0/bin/node Yarn: 1.22.19 - ~/.nvm/versions/node/v16.17.0/bin/yarn npm: 6.14.18 - ~/.nvm/versions/node/v16.17.0/bin/npm pnpm: 8.11.0 - ~/Library/pnpm/pnpm Browsers: Chrome: 120.0.6099.234 Safari: 16.1 Validations [X] Read the docs. [X] Read the common issues list. [X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate. [X] Make sure this is a Module federation issue and not a framework-specific issue. [X] The provided reproduction is a minimal reproducible example of the bug. When using AsyncBoundaryPlugin, you may need to flag the federation runtime module in the entrypoint as eager. You will also need to exclude the remote entry from being wrapped in async boundary new AsyncBoundaryPlugin({ eager: (module) => /\.federation/.test(module?.request || ''), excludeChunk: (chunk) => chunk.name === mfOptions.name }), Hmm i add it , but the dist/main.js will execute webpack_require.f.remotes() which not defined
2025-04-01T04:34:47.721691
2020-04-21T00:54:09
603627051
{ "authors": [ "LiTszOn", "mofanv", "ychen404" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8766", "repo": "mofanv/darknetz", "url": "https://github.com/mofanv/darknetz/issues/12" }
gharchive/issue
Not able to run darknetz Hi, I am having problem integrating darknetz with op-tee. I am able to run op-tee and pass all the test item of xtest without integrating darknetz. But once I copy the darknetz folder and the tzdatasets folder to op-tee using the following two commands provided on this repo. cp -a darknetp/. optee/optee_examples/darknetp/ cp -a tz_datasets/. optee/out-br/target/root/ Then I run into kernel panic and the following is the error message from the normal world terminal. Unpacking initramfs... Initramfs unpacking failed: write error Freeing initrd memory: 328932K workingset: timestamp_bits=30 max_order=19 bucket_order=0 squashfs: version 4.0 (2009/01/31) Phillip Lougher 9p: Installing v9fs 9p2000 file system support io scheduler mq-deadline registered io scheduler kyber registered physmap-flash 4000000.flash: physmap platform flash device: [mem 0x04000000-0x07ffffff] 4000000.flash: Found 2 x16 devices at 0x0 in 32-bit bank. Manufacturer ID 0x000000 Chip ID 0x000000 Intel/Sharp Extended Query Table at 0x0031 Using buffer write method libphy: Fixed MDIO Bus: probed usbcore: registered new interface driver usb-storage rtc-pl031 9010000.pl031: registered as rtc0 ledtrig-cpu: registered to indicate activity on CPUs usbcore: registered new interface driver usbhid usbhid: USB HID core driver optee: probing for conduit method from DT. optee: revision 3.8 (aeb5ba43) optee: dynamic shared memory is enabled optee: initialized driver oprofile: no performance counters oprofile: using timer interrupt. NET: Registered protocol family 17 9pnet: Installing 9P2000 support Registering SWP/SWPB emulation handler rtc-pl031 9010000.pl031: setting system clock to 2020-04-21T00:42:18 UTC<PHONE_NUMBER>) ALSA device list: No soundcards found. Freeing unused kernel memory: 1024K Run /init as init process Failed to execute /init (error -2) Run /sbin/init as init process Run /etc/init as init process Run /bin/init as init process Run /bin/sh as init process Kernel panic - not syncing: No working init found. Try passing init= option to kernel. See Linux Documentation/admin-guide/init.rst for guidance. CPU: 1 PID: 1 Comm: swapper/0 Not tainted 5.5.0 #1 Hardware name: Generic DT based system [<801114f8>] (unwind_backtrace) from [<8010c8b8>] (show_stack+0x10/0x14) [<8010c8b8>] (show_stack) from [<8076c45c>] (dump_stack+0x84/0x98) [<8076c45c>] (dump_stack) from [<80121e48>] (panic+0x11c/0x31c) [<80121e48>] (panic) from [<80783fd8>] (kernel_init+0x108/0x110) [<80783fd8>] (kernel_init) from [<801010e8>] (ret_from_fork+0x14/0x2c) Exception stack(0xc00c9fb0 to 0xc00c9ff8) 9fa0: 00000000 00000000 00000000 00000000 9fc0: 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000000 9fe0: 00000000 00000000 00000000 00000000 00000013 00000000 CPU0: stopping CPU: 0 PID: 0 Comm: swapper/0 Not tainted 5.5.0 #1 Hardware name: Generic DT based system [<801114f8>] (unwind_backtrace) from [<8010c8b8>] (show_stack+0x10/0x14) [<8010c8b8>] (show_stack) from [<8076c45c>] (dump_stack+0x84/0x98) [<8076c45c>] (dump_stack) from [<8010ffb4>] (handle_IPI+0x3f4/0x41c) [<8010ffb4>] (handle_IPI) from [<8045e534>] (gic_handle_irq+0x8c/0x90) [<8045e534>] (gic_handle_irq) from [<80101a8c>] (__irq_svc+0x6c/0xa8) Exception stack(0x80b01f10 to 0x80b01f58) 1f00: 00000000 00003280 c20cf034 8011a000 1f20: ffffe000 00000000 80b04c28 80b04c64 00000001 00000000 80a6ab88 8094e260 1f40: 00000000 80b01f60 80109360 80109364 60000113 ffffffff [<80101a8c>] (__irq_svc) from [<80109364>] (arch_cpu_idle+0x38/0x3c) [<80109364>] (arch_cpu_idle) from [<801515ec>] (do_idle+0x1a4/0x260) [<801515ec>] (do_idle) from [<801519a4>] (cpu_startup_entry+0x18/0x1c) [<801519a4>] (cpu_startup_entry) from [<80a00eb4>] (start_kernel+0x484/0x4b4) ---[ end Kernel panic - not syncing: No working init found. Try passing init= option to kernel. See Linux Documentation/admin-guide/init.rst for guidance. ]--- Following is the error message from Secure World D/TC:0 0 get_aslr_seed:1304 Cannot find valid kaslr-seed D/TC:0 0 add_phys_mem:586 VCORE_UNPG_RX_PA type TEE_RAM_RX 0x0e100000 size 0x00075000 D/TC:0 0 add_phys_mem:586 VCORE_UNPG_RW_PA type TEE_RAM_RW 0x0e175000 size 0x0018b000 D/TC:0 0 add_phys_mem:586 TA_RAM_START type TA_RAM 0x0e300000 size 0x00d00000 D/TC:0 0 add_phys_mem:586 TEE_SHMEM_START type NSEC_SHM 0x7fe00000 size 0x00200000 D/TC:0 0 add_phys_mem:586 ROUNDDOWN(0x09040000, CORE_MMU_PGDIR_SIZE) type IO_SEC 0x09000000 size 0x00100000 D/TC:0 0 add_phys_mem:586 ROUNDDOWN(0x0e000000, CORE_MMU_PGDIR_SIZE) type IO_SEC 0x0e000000 size 0x00100000 D/TC:0 0 add_phys_mem:586 ROUNDDOWN((0x08000000 + 0), CORE_MMU_PGDIR_SIZE) type IO_SEC 0x08000000 size 0x00100000 D/TC:0 0 add_phys_mem:586 ROUNDDOWN((0x08000000 + 0x10000), CORE_MMU_PGDIR_SIZE) type IO_SEC 0x08000000 size 0x00100000 D/TC:0 0 add_phys_mem:599 Physical mem map overlaps 0x8000000 D/TC:0 0 verify_special_mem_areas:524 No NSEC DDR memory area defined D/TC:0 0 add_va_space:625 type RES_VASPACE size 0x00a00000 D/TC:0 0 add_va_space:625 type SHM_VASPACE size 0x02000000 D/TC:0 0 dump_mmap_table:737 type TEE_RAM_RX va 0x0e100000..0x0e174fff pa 0x0e100000..0x0e174fff size 0x00075000 (smallpg) D/TC:0 0 dump_mmap_table:737 type TEE_RAM_RW va 0x0e175000..0x0e2fffff pa 0x0e175000..0x0e2fffff size 0x0018b000 (smallpg) D/TC:0 0 dump_mmap_table:737 type SHM_VASPACE va 0x0e300000..0x102fffff pa 0x00000000..0x01ffffff size 0x02000000 (pgdir) D/TC:0 0 dump_mmap_table:737 type RES_VASPACE va 0x10300000..0x10cfffff pa 0x00000000..0x009fffff size 0x00a00000 (pgdir) D/TC:0 0 dump_mmap_table:737 type IO_SEC va 0x10d00000..0x10dfffff pa 0x08000000..0x080fffff size 0x00100000 (pgdir) D/TC:0 0 dump_mmap_table:737 type IO_SEC va 0x10e00000..0x10efffff pa 0x09 000000..0x090fffff size 0x00100000 (pgdir) D/TC:0 0 dump_mmap_table:737 type IO_SEC va 0x10f00000..0x10ffffff pa 0x0e000000..0x0e0fffff size 0x00100000 (pgdir) D/TC:0 0 dump_mmap_table:737 type TA_RAM va 0x11000000..0x11cfffff pa 0x0e300000..0x0effffff size 0x00d00000 (pgdir) D/TC:0 0 dump_mmap_table:737 type NSEC_SHM va 0x11d00000..0x11efffff pa 0x7fe00000..0x7fffffff size 0x00200000 (pgdir) D/TC:0 0 core_mmu_alloc_l2:265 L2 table used: 1/5 I/TC: D/TC:0 0 init_canaries:164 #Stack canaries for stack_tmp[0] with top at 0xe19ae38 D/TC:0 0 init_canaries:164 watch *0xe19ae3c D/TC:0 0 init_canaries:164 #Stack canaries for stack_tmp[1] with top at 0xe19b778 D/TC:0 0 init_canaries:164 watch *0xe19b77c D/TC:0 0 init_canaries:164 #Stack canaries for stack_tmp[2] with top at 0xe19c0b8 D/TC:0 0 init_canaries:164 watch *0xe19c0bc D/TC:0 0 init_canaries:164 #Stack canaries for stack_tmp[3] with top at 0xe19c9f8 D/TC:0 0 init_canaries:164 watch *0xe19c9fc D/TC:0 0 init_canaries:165 #Stack canaries for stack_abt[0] with top at 0xe19d238 D/TC:0 0 init_canaries:165 watch *0xe19d23c D/TC:0 0 init_canaries:165 #Stack canaries for stack_abt[1] with top at 0xe19da78 D/TC:0 0 init_canaries:165 watch *0xe19da7c D/TC:0 0 init_canaries:165 #Stack canaries for stack_abt[2] with top at 0xe19e2b8 D/TC:0 0 init_canaries:165 watch *0xe19e2bc D/TC:0 0 init_canaries:165 #Stack canaries for stack_abt[3] with top at 0xe19eaf8 D/TC:0 0 init_canaries:165 watch *0xe19eafc D/TC:0 0 init_canaries:167 #Stack canaries for stack_thread[0] with top at 0xe1a0b38 D/TC:0 0 init_canaries:167 watch *0xe1a0b3c D/TC:0 0 init_canaries:167 #Stack canaries for stack_thread[1] with top at 0xe1a2b78 D/TC:0 0 init_canaries:167 watch *0xe1a2b7c I/TC: Non-secure external DT found D/TC:0 0 carve_out_phys_mem:286 No need to carve out 0xe100000 size 0x200000 D/TC:0 0 carve_out_phys_mem:286 No need to carve out 0xe300000 size 0xd00000 D/TC:0 0 dt_add_psci_node:729 PSCI Device Tree node already exists! I/TC: Switching console to device: /pl011@9040000 I/TC: OP-TEE version: 3.8.0-263-gaeb5ba4 (gcc version 8.3.0 (GNU Toolchain for the A-profile Architecture 8.3-2019.03 (arm-rel-8.36))) #1 Sat Apr 18 07:30:35 UTC 2020 arm D/TC:0 0 init_primary_helper:1171 Executing at offset 0 with virtual load address 0xe100000 D/TC:0 0 check_ta_store:635 TA store: "Secure Storage TA" D/TC:0 0 check_ta_store:635 TA store: "REE" D/TC:0 0 mobj_mapped_shm_init:447 Shared memory address range: e300000, 10300000 D/TC:0 0 gic_it_set_cpu_mask:251 cpu_mask: writing 0xff to 0x10d00828 D/TC:0 0 gic_it_set_cpu_mask:253 cpu_mask: 0xff D/TC:0 0 gic_it_set_prio:266 prio: writing 0x1 to 0x10d00428 I/TC: Initialized D/TC:0 0 init_primary_helper:1184 Primary CPU switching to normal world boot D/TC:0 psci_cpu_on:211 core pos: 1: ns_entry 0x40102620 D/TC:1 init_secondary_helper:1208 Secondary CPU Switching to normal world boot D/TC:1 tee_entry_exchange_capabilities:102 Dynamic shared memory is enabled D/TC:1 0 core_mmu_alloc_l2:265 L2 table used: 2/5 D/TC:? 0 tee_ta_init_pseudo_ta_session:284 Lookup pseudo TA 7011a688-ddde-4053-a5a9-7b3c4ddf13b8 D/TC:? 0 tee_ta_init_pseudo_ta_session:297 Open device.pta D/TC:? 0 tee_ta_init_pseudo_ta_session:311 device.pta : 7011a688-ddde-4053-a5a9-7b3c4ddf13b8 D/TC:? 0 tee_ta_close_session:499 csess 0xe191dd0 id 1 D/TC:? 0 tee_ta_close_session:518 Destroy session I have no idea how integrating darknetz can lead to booting issue. Any help is appreciated! Hi @ychen404 , Seems there is a problem when packing initramfs. Can you please try to remove the data folder from optee/out-br/target/root/, and only keep darknetp/ in the optee_examples/ folder? Another issue could be the OPTEE version is not compatible. Please try to downgrade it to 3.4.0 following https://github.com/OP-TEE/optee_docs/blob/master/building/gits/build.rst Hi @mofanv I removed the data folder from opt/out-br/target/root/ and only kept darknetp in the optee_examples folder. Now the OS can boot without getting initramfs issue. But when i tried to run darknetp in the Normal World, I got the following error message. darknetp Prepare session with the TA darknetp: TEEC_Opensession failed with code 0xffff000c origin 0x3 I will try to downgrade op-tee to 3.4.0. Do I need to replace the build folder using your link, https://github.com/OP-TEE/optee_docs/blob/master/building/gits/build.rst Or, it is fine to use the build folder from git as follows? ./repo init -u https://github.com/OP-TEE/manifest.git -m default.xml -b 3.4.0 Thanks! I'd suggest removing the OPTEE project and do all things from the beginning to avoid any possible issues. Could you try to clone the OPTEE again, and then use the following command to initialize the branch (same in the link I sent above)? repo init -u https://github.com/OP-TEE/manifest.git -m rpi3.xml -b 3.4.0 Please let me know whether version 3.4.0 fix errors or not. Thanks. Hi Fan, I am using QEMU v7. Can I use the branch 3.4.0 for default.xml instead? On Thu, Apr 30, 2020 at 2:34 AM Mo, Fan Vincent<EMAIL_ADDRESS>wrote: I'd suggest removing the OPTEE project and do all things from the beginning to avoid any possible issues. Could you try to clone the OPTEE again, and then use the following command to initialize the branch (same in the link I sent above)? repo init -u https://github.com/OP-TEE/manifest.git -m rpi3.xml -b 3.4.0 Please let me know whether version 3.4.0 fix errors or not. Thanks. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/mofanv/darknetz/issues/12#issuecomment-621723948, or unsubscribe https://github.com/notifications/unsubscribe-auth/AFJO3TW4FFI2MIPDH7H4GDTRPFA23ANCNFSM4MM4LGKA . Yes. go ahead. QEMU v7 should work with default.xml Hi! I encountered the same problem as @ychen404, that I also got the darknetp: TEEC_Opensession failed with code 0xffff000c origin 0x3 error message after removing data folder from opt/out-br/target/root/. I am using QEMU v7 with branch 3.4.0. Please let me know about any idea of how it can be fixed. Thanks in advance! @LiTszOn Hi, is there any error when you run the hello_world example? @mofanv Many thanks for your swift reply! No, there is no error when I run hello_world, which outputs Invoking TA to increment 42 TA incremented value to 43 There is also no error when I run random,hotp etc. @mofanv Still not working with 3.4.0. I have no error running hello world. # darknetp Prepare session with the TA darknetp: TEEC_Opensession failed with code 0xffff0008 origin 0x3 @mofanv It is working now after reducing the required memory size to 2 MB. Thanks. @ychen404 Thanks for your feedback. May I know if you constrain the required memory size to 2 MB by specifying -m 2M in the Makefile under optee/build? No, the memory size is defined in a header file, which locates at: ta/include/user_ta_header_defines.h You only need to change #define TA_DATA_SIZE To 2 MB. @ychen404 Thank you for your prompt reply!! It works now, thanks a lot!! @mofanv I am having trouble running optee 3.4.0 on Raspberry Pi 3b+. I used the following command to init the repo, as you recommended. repo init -u https://github.com/OP-TEE/manifest.git -m rpi3.xml -b 3.4.0 But I am getting the following invalid path error. ManifestInvalidPathError: invalid "src": ../toolchains/aarch64/bin/aarch64-linux-gnu-gdb: bad component: Can you point me some directions. Thanks. It is highly probable that this is an OPTEE problem, I'd suggest you ask the OPTEE team because they know this better.
2025-04-01T04:34:47.729968
2024-01-16T10:37:52
2083598215
{ "authors": [ "ivanborisof", "jamesonfajardo", "juliansteenbakker", "vitoramaral10" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8767", "repo": "mogol/flutter_secure_storage", "url": "https://github.com/mogol/flutter_secure_storage/issues/672" }
gharchive/issue
Asymmetric cipher with insecure padding used Hi Team, a security audit flagged this issue on our app. May I request for more info on this. flutter_secure_storage: ^4.2.0 The code you threw is most likely obtained after OWASP security analysis in the web application appsweep: protected Cipher i() { String string; if (Build.VERSION.SDK_INT < 23) { string = "AndroidOpenSSL"; return Cipher.getInstance("RSA/ECB/PKCS1Padding", string); } string = "AndroidKeyStoreBCWorkaround"; return Cipher.getInstance("RSA/ECB/PKCS1Padding", string); } This code was obtained during reverse engineering and it matches the code in the library file (RSACipher18Implementation.java): protected Cipher getRSACipher() throws Exception { if (Build.VERSION.SDK_INT < Build.VERSION_CODES.M) { return Cipher.getInstance("RSA/ECB/PKCS1Padding", "AndroidOpenSSL"); // error in android 6: InvalidKeyException: Need RSA private or public key } else { return Cipher.getInstance("RSA/ECB/PKCS1Padding", "AndroidKeyStoreBCWorkaround"); // error in android 5: NoSuchProviderException: Provider not available: AndroidKeyStoreBCWorkaround } } The getRSAChiper method is used in only two places in the file (RSACipher18Implementation.java) of the RSACipher18Implementation class: @Override public byte[] wrap(Key key) throws Exception { PublicKey publicKey = getPublicKey(); Cipher cipher = **getRSACipher();** cipher.init(Cipher.WRAP_MODE, publicKey, getAlgorithmParameterSpec()); return cipher.wrap(key); } @Override public Key unwrap(byte[] wrappedKey, String algorithm) throws Exception { PrivateKey privateKey = getPrivateKey(); Cipher cipher = **getRSACipher();** cipher.init(Cipher.UNWRAP_MODE, privateKey, getAlgorithmParameterSpec()); return cipher.unwrap(wrappedKey, algorithm, Cipher.SECRET_KEY); } Further if we trace where the RSACipher18Implementation class is used we will see enum: enum KeyCipherAlgorithm { RSA_ECB_PKCS1Padding(RSACipher18Implementation::new, 1), @SuppressWarnings({"UnusedDeclaration"}) RSA_ECB_OAEPwithSHA_256andMGF1Padding(RSACipherOAEPImplementation::new, Build.VERSION_CODES.M); final KeyCipherFunction keyCipher; final int minVersionCode; KeyCipherAlgorithm(KeyCipherFunction keyCipher, int minVersionCode) { this.keyCipher = keyCipher; this.minVersionCode = minVersionCode; } } This enum is used in the StorageCipherFactory class: public class StorageCipherFactory { private static final String ELEMENT_PREFERENCES_ALGORITHM_PREFIX = "FlutterSecureSAlgorithm"; private static final String ELEMENT_PREFERENCES_ALGORITHM_KEY = ELEMENT_PREFERENCES_ALGORITHM_PREFIX + "Key"; private static final String ELEMENT_PREFERENCES_ALGORITHM_STORAGE = ELEMENT_PREFERENCES_ALGORITHM_PREFIX + "Storage"; private static final KeyCipherAlgorithm DEFAULT_KEY_ALGORITHM = KeyCipherAlgorithm.RSA_ECB_PKCS1Padding; private static final StorageCipherAlgorithm DEFAULT_STORAGE_ALGORITHM = StorageCipherAlgorithm.AES_CBC_PKCS7Padding; private final KeyCipherAlgorithm savedKeyAlgorithm; private final StorageCipherAlgorithm savedStorageAlgorithm; private final KeyCipherAlgorithm currentKeyAlgorithm; private final StorageCipherAlgorithm currentStorageAlgorithm; public StorageCipherFactory(SharedPreferences source, Map<String, Object> options) { savedKeyAlgorithm = KeyCipherAlgorithm.valueOf(source.getString(ELEMENT_PREFERENCES_ALGORITHM_KEY, DEFAULT_KEY_ALGORITHM.name())); savedStorageAlgorithm = StorageCipherAlgorithm.valueOf(source.getString(ELEMENT_PREFERENCES_ALGORITHM_STORAGE, DEFAULT_STORAGE_ALGORITHM.name())); final KeyCipherAlgorithm currentKeyAlgorithmTmp = KeyCipherAlgorithm.valueOf(getFromOptionsWithDefault(options, "keyCipherAlgorithm", DEFAULT_KEY_ALGORITHM.name())); currentKeyAlgorithm = (currentKeyAlgorithmTmp.minVersionCode <= Build.VERSION.SDK_INT) ? currentKeyAlgorithmTmp : DEFAULT_KEY_ALGORITHM; final StorageCipherAlgorithm currentStorageAlgorithmTmp = StorageCipherAlgorithm.valueOf(getFromOptionsWithDefault(options, "storageCipherAlgorithm", DEFAULT_STORAGE_ALGORITHM.name())); currentStorageAlgorithm = (currentStorageAlgorithmTmp.minVersionCode <= Build.VERSION.SDK_INT) ? currentStorageAlgorithmTmp : DEFAULT_STORAGE_ALGORITHM; } private String getFromOptionsWithDefault(Map<String, Object> options, String key, String defaultValue) { final Object value = options.get(key); return value != null ? value.toString() : defaultValue; } public boolean requiresReEncryption() { return savedKeyAlgorithm != currentKeyAlgorithm || savedStorageAlgorithm != currentStorageAlgorithm; } public StorageCipher getSavedStorageCipher(Context context) throws Exception { final KeyCipher keyCipher = savedKeyAlgorithm.keyCipher.apply(context); return savedStorageAlgorithm.storageCipher.apply(context, keyCipher); } public StorageCipher getCurrentStorageCipher(Context context) throws Exception { final KeyCipher keyCipher = currentKeyAlgorithm.keyCipher.apply(context); return currentStorageAlgorithm.storageCipher.apply(context, keyCipher); } public void storeCurrentAlgorithms(SharedPreferences.Editor editor) { editor.putString(ELEMENT_PREFERENCES_ALGORITHM_KEY, currentKeyAlgorithm.name()); editor.putString(ELEMENT_PREFERENCES_ALGORITHM_STORAGE, currentStorageAlgorithm.name()); } public void removeCurrentAlgorithms(SharedPreferences.Editor editor) { editor.remove(ELEMENT_PREFERENCES_ALGORITHM_KEY); editor.remove(ELEMENT_PREFERENCES_ALGORITHM_STORAGE); } } NOTE THERE IS A METHOD getFromOptionsWithDefault which is involved in selecting the enum value: private String getFromOptionsWithDefault(Map<String, Object> options, String key, String defaultValue) { final Object value = options.get(key); return value != null ? value.toString() : defaultValue; } Conclusion Yes, there is indeed a warning "Asymmetric cipher with insecure addition is used" on the piece of code that you threw. Well you should realise that from API version greater than 23 the good encryption algorithm RSA_ECB_OAEPwithSHA_256andMGF1Padding will be used. This can also be set manually when declaring SecureStorage: static const _secureStorage = FlutterSecureStorage( aOptions: AndroidOptions( encryptedSharedPreferences: true, keyCipherAlgorithm: KeyCipherAlgorithm.RSA_ECB_OAEPwithSHA_256andMGF1Padding, ), ); Or it is easier via encryptedSharedPreferences: true, because when true, keyCipherAlgorithm will not be passed the default parameter of the old encryption algorithm: static const _secureStorage = FlutterSecureStorage( aOptions: AndroidOptions( encryptedSharedPreferences: true, ), ); Comment from the documentation for the keyCipherAlgorithm parameter: /// If EncryptedSharedPrefences is set to false, you can select algorithm /// that will be used to encrypt secret key. /// By default RSA/ECB/PKCS1Padding if used. /// Newer RSA/ECB/OAEPWithSHA-256AndMGF1Padding is available from Android 6. /// Plugin will fall back to default algorithm in previous system versions. I'm sorry, but I'm having this problem with my app as well. I've tried to make the settings as mentioned here, but it seems that AppSweep doesn't understand this and keeps accusing the security flaw. I think this can only be solved by removing this part of the package code. This issue will be tracked in #769
2025-04-01T04:34:47.734759
2023-12-10T17:06:41
2034453857
{ "authors": [ "mogzol", "tofsjonas" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8768", "repo": "mogzol/prisma-generator-typescript-interfaces", "url": "https://github.com/mogzol/prisma-generator-typescript-interfaces/issues/1" }
gharchive/issue
It crashes on Mac M1 :( The error I am getting: import: delegate library support not built-in '' (X11) @ error/import.c/ImportImageCommand/1302. import: delegate library support not built-in '' (X11) @ error/import.c/ImportImageCommand/1302. import: delegate library support not built-in '' (X11) @ error/import.c/ImportImageCommand/1302. /api/node_modules/.bin/prisma-generator-typescript-interfaces: line 4: //: is a directory /api/node_modules/.bin/prisma-generator-typescript-interfaces: line 5: const: command not found /api/node_modules/.bin/prisma-generator-typescript-interfaces: line 6: syntax error near unexpected token `=' /api/node_modules/.bin/prisma-generator-typescript-interfaces: line 6: ` String: () => "string",' Node version: v18.17.1 What does your schema.prisma file look like? (at least the generator block) I have used this on an M1 mac so it should work, it looks like maybe it's not being run properly though. My schema looks like this: datasource db { provider = "mysql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" } generator typescriptInterfaces { provider = "prisma-generator-typescript-interfaces" } But I suspect it is a wayland vs x11 issue, ie an OS problem 😢 (I use wayland) Ah, I found the issue, dumb mistake, I forgot to include a shebang on the script so it wasn't running via node for you. Apparently this isn't an issue with pnpm (which is all I've tested with), but is an issue with npm. Sorry! I've just released v1.2.0 which should fix it. Also fixes a compatibility issue with v5.7.0 of Prisma. It works! Thank you 🙏
2025-04-01T04:34:47.769592
2022-10-26T17:43:30
1424456985
{ "authors": [ "ajarmoniuk" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8769", "repo": "mojohaus/versions-maven-plugin", "url": "https://github.com/mojohaus/versions-maven-plugin/pull/799" }
gharchive/pull-request
#794: SetMojo would always change the version of the POM, regardless if a match was found Probably an old technical debt of sorts in SetMojo: the mojo would normally change the files matching oldVersion, groupId, or ArtifactId. However, if no match was found in the reactor, then the root model is changed anyway! https://github.com/mojohaus/versions-maven-plugin/blob/b75a5656113f8974e7bb26f2cd0cdeacff59800e/src/main/java/org/codehaus/mojo/versions/SetMojo.java#L384 I don't understand the reason for it. When checked git blame, the original author added those lines to fix some breaking integration tests. However, it looks like no tests fail even with those lines removed. Anyhow. I don't think that behaviour was intuitive and it is certainly very unexpected. A side effect of that behaviour was that, with "always", the timestamp of the file was updated regardless if a match was found. So I had to explicitly add all reactor files if the update timestamp switch is set to "always". @slawekjaranowski Please review and merge. Please merge unless you have any more remarks. Rebased
2025-04-01T04:34:47.788921
2017-01-26T04:28:33
203290357
{ "authors": [ "Deester4x4jr", "alinchican" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8770", "repo": "mojotech/jeet", "url": "https://github.com/mojotech/jeet/issues/522" }
gharchive/issue
shift seems to have been changed to move Just noticing, as I recently installed jeet with npm... looks like shift has been renamed move? Didn't find it anywhere on the website, so I was a little puzzled when I kept trying to shift columns and was getting nowhere. Maybe some documentation is in order? Your grid system is awesome, and I think documentation would really help adoption. I did find this, but it seems a bit outdated (mainly because of the shift/move thingy): https://www.sitepoint.com/super-fast-responsive-layouts-jeet/ Thanks for all the hard work thus far! -Josh- It is documented in the migration guide. Check out the README. shift() and unshift() have been renamed to move() and unmove() to prevent potential clashes with Stylus built-in shift() function Go through projects and rename shift() and unshift() to move() and unmove().
2025-04-01T04:34:47.792513
2024-09-02T07:10:51
2500182576
{ "authors": [ "PluginYML" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8771", "repo": "mokhosh/filament-kanban", "url": "https://github.com/mokhosh/filament-kanban/issues/55" }
gharchive/issue
[Bug]: Cannot move single entries. What happened? I have installed the whole thing, but I cannot move my entries. I use Firefox, but have also tried Edge. How to reproduce the bug As described above Package Version ^2.8 PHP Version 8.2.4 Laravel Version ^11.9 Which operating systems does with happen with? Windows Which browsers does with happen with? Firefox Notes No response Here is the code to follow: class OrderKanbanBoard extends KanbanBoard { protected static string $model = Order::class; protected static ?string $navigationGroup = 'Bestellungen'; protected static ?string $navigationLabel = 'Bestellübersicht'; protected static ?string $title = 'Bestellübersicht'; protected static ?string $navigationIcon = null; protected static ?int $navigationSort = 2; protected static string $recordTitleAttribute = 'treatmentNr'; protected static string $recordStatusAttribute = 'production_step_id'; public bool $disableEditModal = true; public function onStatusChanged(int $recordId, string $status, array $fromOrderedIds, array $toOrderedIds): void { Order::whereId($recordId)->update([ 'production_step_id' => $status ]); } public function onSortChanged(int $recordId, string $status, array $orderedIds): void { } protected function records(): Collection { return Order::latest('updated_at')->get(); } protected function statuses(): Collection { $collection = new Collection(); $collection->add([ 'id' => null, 'title' => 'Nicht zugewiesen' ]); ProductionStep::orderBy('sort')->get()->each(function (ProductionStep $step) use ($collection) { $collection->add([ 'id' => $step->id, 'title' => $step->name, ]); }); return $collection; } } Yes, you were right. Now everything works perfectly, thank you!
2025-04-01T04:34:47.872195
2016-01-16T19:19:46
127046951
{ "authors": [ "farlop", "mj1856" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8772", "repo": "moment/moment", "url": "https://github.com/moment/moment/issues/2900" }
gharchive/issue
Since 2.11.0 week numbers are not calculated correctly I noticed that since updated to 2.11.0 week numbers (more precisely using format('w')) are not calculated correctly on my fullCalendar. I'm using spanish locale. For example, the week number for date 28/12/2015 should be 53 which is right in version 2.10.6. Starting 2.11.0, moment says it's 44. Hmmm. I tried: moment('2015-12-28').locale('es').format('w') Under both 2.11.0 and 2.11.1, I get week 53. Please test again, and provide a code snippet that reproduces the problem. Thanks. FYI, Week 44 by the Spanish locale is from 2015-10-26 through 2015-11-01. OK. I managed to reproduce it, but it involves fullCalendar, so I doubt if it's a moment.js bug or a fullCalendar... In this jsfiddle: https://jsfiddle.net/xwyqq4rg/4/ If the 'firstDay' argument is provided as a string, the issue occurs, but if it is provided as int it works fine. I thought it was a moment.js because rolling back the version solved the problem, but now I don't know if I should create a new issue in fullCalendar... I'll investigate and get back to you. Thanks. Seems like the easy answer is to always pass it as a number. I'm not sure precisely what allowed you to pass it as a string in previous versions. If fullcalendar wants to allow a string in their API, they can do that. XRef fullcalendar/fullcalendar#3014
2025-04-01T04:34:47.877564
2018-09-25T22:20:44
363788273
{ "authors": [ "ashsearle", "marwahaha", "samit4me" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8773", "repo": "moment/moment", "url": "https://github.com/moment/moment/issues/4788" }
gharchive/issue
moment().isSame(undefined, 'day') === true moment().isSame(undefined, 'day') === true If you remove the second parameter it works as intended (e.g. false), but if you change day to year it also returns true. I've tested this using version 2.22.2. Found the following: moment().isAfter(undefined, 'day') === false (correct) moment().isBefore(undefined, 'day') === false (correct) moment().isSameOrAfter(undefined, 'day') === true (broken) moment().isSameOrBefore(undefined, 'day') === true (broken) I understand the docs say the first argument should be of type Moment|String|Number|Date|Array but it would be nice if undefined was also supported like in the other methods. The documentation for isBefore has some hints about this: If nothing is passed to moment#isBefore, it will default to the current time. NOTE: moment().isBefore() has undefined behavior and should not be used! If the code runs fast the initial created moment would be the same as the one created in isBefore to perform the check, so the result would be false. But if the code runs slower it's possible that the moment created in isBefore is measurably after the one created in moment(), so the call would return true. Passing undefined is the closest you can get to passing nothing. The same issue described above applies to isAfter, isSameOrAfter and isSameOrBefore. The reason you're reliably seeing true when adding a unit is that you're very unlikely to bridge a day or year boundary while creating two moment instances. But, if you omit the unit you're effectively doing: moment().isSame(undefined, 'millisecond'), which is unstable. Nice find. But Ash is right, this doesn't always work. Perhaps the documentation can be improved on this.
2025-04-01T04:34:47.880301
2018-05-22T05:04:19
325141984
{ "authors": [ "coveralls", "ichernev", "nisha-kaushik" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8774", "repo": "moment/moment", "url": "https://github.com/moment/moment/pull/4636" }
gharchive/pull-request
#3263 : moment isValid() issue with 12 hour format and AM.../PM #3263 Coverage increased (+0.2%) to 94.693% when pulling d6a96b84856b4813745ad189b8d65dbd4d7b7e50 on nisha-kaushik:develop into 3062ca13cd1cfbcef1c4a0cd5543f5fcd329c1d9 on moment:develop. This is already handled in strict mode. bigHour parsing flag is used for that purpose. https://github.com/moment/moment/blob/develop/src/lib/create/valid.js#L30 If you really want you can add a separate function (or monkey patch isValid) to return isValid + bigHour === undefined even in non-strict mode. But I won't merge that :)
2025-04-01T04:34:47.888798
2022-11-29T17:15:16
1468429569
{ "authors": [ "nicolevasilevsky", "pnrobinson" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8776", "repo": "monarch-initiative/mondo", "url": "https://github.com/monarch-initiative/mondo/issues/5665" }
gharchive/issue
[Revise subclass] orofacial clefting syndrome Mondo term (ID and Label) orofacial clefting syndrome Suggested revision and reasons All of the subclasses are genetic and yet there is no corresponding parentage Your nano-attribution (ORCID) If you don't have an ORCID, you can sign up for one here this seems like an orphanet grouping class that is grouping terms based on phenotypes. is there a true disease that is called orofacial clefting syndrome? Maybe this grouping class should be obsoleted? Here is a related article https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2825080/ talking to Ada, this could be renamed to 'mendelian syndromes with cleft lip/palate' and make it a child of Mendelian disease
2025-04-01T04:34:47.890726
2023-02-06T13:05:36
1572534718
{ "authors": [ "yaseminbridges" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8777", "repo": "monarch-initiative/pheval", "url": "https://github.com/monarch-initiative/pheval/pull/54" }
gharchive/pull-request
#53 implement methods for assigning ranks to sorted result output (WIP) These methods include sorting tool-specific output that has been post-processed to the PhEval gene/variant dataclasses in descending order with "best" scores first and assigning ranks to these results. Looks great. Remember, when I approve something, I didn't do an in-depth review, only an architectural one, so you should get a proper review from @souzadevinicius and/or @julesjacobsen if you want to be sure. No problem, I've requested both of their reviews for this as well Just added 2 small methods to write the pheval gene and variant .tsv results
2025-04-01T04:34:47.949317
2016-12-26T11:52:58
197580030
{ "authors": [ "Apoorva333", "ShaneHarvey" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8778", "repo": "mongodb-labs/mongo-connector", "url": "https://github.com/mongodb-labs/mongo-connector/issues/618" }
gharchive/issue
MongoDB custom filter I am using mongo connector to map data from mongodb to elasticsearch. Is there any property available in the mongo-connector through which I would be able to index a collection(for the very first time) consisting of documents of only a particular date range(which can be identified by "date" field in each document) from mongodb ? mongo-connector version: 2.5.0.dev0 Source MongoDB version: 3.0.12 Elasticsearch version: 1.4.4 Closing as duplicate of https://github.com/mongodb-labs/mongo-connector/issues/304.
2025-04-01T04:34:47.960682
2023-05-30T11:36:12
1732077635
{ "authors": [ "jmikola" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8779", "repo": "mongodb/mongo-php-library", "url": "https://github.com/mongodb/mongo-php-library/pull/1086" }
gharchive/pull-request
Remove phpcs:ignore comments for unused use statements The line in DropEncryptedCollection was never necessary. The line in CreateEncryptedCollection may have been used at one point since Binary is only referenced in a psalm-var annotation, but phpcs no longer complains. Reported the unrelated test failure on MongoDB 4.0 shard cluster in PHPLIB-1151. Other test failures are covered by PHPLIB-1115 and PHPLIB-1071.
2025-04-01T04:34:47.997233
2017-01-20T16:36:27
202185143
{ "authors": [ "afouchet", "edengreen", "srinivasreddy" ], "license": "ISC", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8780", "repo": "mongomock/mongomock", "url": "https://github.com/mongomock/mongomock/pull/289" }
gharchive/pull-request
Enable pymongo bulk_write Mongomock did not have the proper behavior when using collection.bulk_write(operations) which allows to make several operations in one call. The operations are (InsertOne, DeleteOne, DeleteMany, ReplaceOne, UpdateOne, UpdateMany), in pymongo. In this PR, the bulk_write behavior + the tests for each operation. Here is a gist with the usage of bulk_write: https://gist.github.com/afouchet/6af2fe7e1010a7197e8e545549f45464 Hi @srinivasreddy! Any comment or feedback on this? Thanks!! my pleasure Thanks for the great work (looking forward to see it officially in pypi). One small issue - could you please add support for pymongo's BulkWriteResult? Should be something like replacing (in mongomock.collection.BulkOperationBuilder.execute): return result with return BulkWriteResult(result, True) Thanks! @edengreen Are you interested in making a PR for that change? @edengreen Also, it is very difficult to track these requests in comments. Could you please raise this as an issue ? People interested can take this up. Added a PR for the change - #299
2025-04-01T04:34:48.007636
2020-09-07T09:41:46
694919604
{ "authors": [ "benclapp", "paulfantom" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8783", "repo": "monitoring-mixins/website", "url": "https://github.com/monitoring-mixins/website/pull/12" }
gharchive/pull-request
Add cert-manager mixin As in title /cc @benclapp Awesome, looks great! Thanks for kicking this off :) Quick question - is it common for people to add an MR here if there updates to alerts/rules/dashboards in a mixin? Happy to do as we make changes just wanting to get an idea. is it common for people to add an MR here if there updates to alerts/rules/dashboards in a mixin? This repository has a daily cronjob which regenerates everything, so no need for a manual updates. Anything I can help with to get this merged? @benclapp I would like to get some feedback on https://gitlab.com/uneeq-oss/cert-manager-mixin/-/issues/1 cert-manager mixin is now included on the site at https://monitoring.mixins.dev/cert-manager/. Thanks @benclapp for the patience and quick fixes in the mixin :100:
2025-04-01T04:34:48.028742
2019-09-11T09:55:32
492134953
{ "authors": [ "Avasil", "Krever", "alexandru", "jvican", "mdedetrich" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8784", "repo": "monix/monix", "url": "https://github.com/monix/monix/issues/1024" }
gharchive/issue
Monix vNext — future plans Series 3.0.0 is now stable and for its duration we can no longer break binary compatibility — at least not on top of Scala 2.12+ — and I predict more features being built on top of what we have. But we should also talk about a wish list for future versions, that can happen both soon (without breaking binary compatibility) or in a far future major version that can break binary compatibility. This list is a work in progress, items may be added or removed, suggestions welcome. Safe (non-binary breaking) changes Provide bi-functor Task and Coeval implementations, possibly in their own sub-project? Not sure if we should, this needs to be discussed — I don't believe in the approach, but there are some people that do As an alternative, maybe a nicer EitherT would be better (e.g. one that is covariant, maybe with better performance as the one in Cats is not representative of what can be done) Make Iterant covariant, for improving type inference (e.g. Iterant[+F[_], +A]) Better stack traces, preferably opt-in, possible solutions need to be investigated ... Binary breaking changes Note that these aren't necessarily plans that will happen in less than a year, or even more — we are talking about a long time frame: Rebuild Observable to use the Reactive Streams protocol internally, instead of the Future-based protocol that we have — but keep the Future-based one for consumption of streams and for ease of use — this might yield better performance in some cases, even if it's more challenging, but we can take inspiration from RxJava, while coming up with our own innovations and trying hard to keep source compatibility right now we piggyback the Reactive Streams on top of our Future-based one, but maybe it should be the other way around what I like about Reactive Streams is that the protocol for cancelation is clearer, whereas with Future you can end up with leaks due to not being able to unregister an "onComplete" one thing I really want, that RxJava doesn't do, is a stack safe flatMap implementation ... I'm not sure right now if this is a limitation of the underlying protocol or not if we can make a stack safe flatMap, we can also implement Sync[Observable] and Async[Observable] at the very least 😉 we might be able to implement that stack safe flatMap even sooner, without breaking binary compatibility, but I remember trying it and being very challenging this needs experimentation before deciding whether it's a good idea or not Refactor Iterant to not have so many operators that depend so much on Sync — e.g. operators should depend on Monad, MonadError, Defer, etc, as Sync should only be needed when we're suspending actual side effects I'm not sure if we can remove some of that dependence on Sync right now, without breaking binary compatibility whatever we can do sooner, we'll do sooner given the long time frame we are talking about, we might be talking about refactoring Iterant to use the future Cats-Effect 3.0 type class hierarchy Remove all @deprecated symbols ... My wishes are modest and could target 3.x.x line. I'm not in rush for 4.0.0 considering how long it took us to get to 3.0.0. And Cats-Effect 3.0 will most likely drain our resources. stack traces (already mentioned) Polymorphic Thread.interrupt support ( #983 ) Overhaul https://monix.io/ Documentation - I feel like it is harder to navigate and looks worse than github sites or whatever is Akka doing. For instance, everything about Task is on one page and it's pretty "narrow" so there's a lot of scrolling. Might be nice to read on a phone but as our docs grow I think we should be able to split it into different sections without requiring a lot of clicks to get there. Font in code snippets could probably be better too. I'm not really knowledgeable about design but that's my subjective feelings. :D Better stack traces, preferably opt-in, possible solutions need to be investigated I'd say it should be opt-out. If there is a performance impact, most people won't notice it but the feature will be useful for everyone. but that's something we can discuss when we have anything Provide bi-functor Task and Coeval implementations, possibly in their own sub-project? Not sure if we should, this needs to be discussed — I don't believe in the approach, but there are some people that do I've grown to really like it. Well, I didn't actually use it in a project yet but I usually follow a standard pattern of F[Either[E, A]] where E is an error which is recoverable and F.raiseError is either not recoverable or when I handle all errors the same way. To me, Bifunctor is just taking this common pattern and embedding it. I don't buy performance argument but ergonomics are better (it's covariant and there is no need to wrap to EitherT to get shorcircuiting on E) and it allows cool stuff like returning UIO in an attempt. ZIO is providing type aliases, something like Task[A] = ZIO[Any, Throwable, A] and IO[E, A] = ZIO[Any, E, A] to not require dealing with an error all the time. I don't know how well it does in practice, I assume it often infers to ZIO anyway. Different type aliases and return types in methods are extra complexity and something you need to learn even if you just want monofunctor. I wonder how could we provide it without making it the default and still be compatible with "normal" Task. As an alternative, maybe a nicer EitherT would be better (e.g. one that is covariant, maybe with better performance as the one in Cats is not representative of what can be done) Covariant EitherT would be cool as well. Not sure if it's good enough in comparison to Bifunctor but it should be useful even if we have both. I'm attending ZIO Hackaton next week. Maybe there will be someone interested in contributing to Monix as well so I could guide this someone to create a prototype. It should be relatively simple if we disregard performance for now. I'm not in a hurry for a binary breaking 4.0.0 either, this is the far future we're talking about. However we might prepare at least some stuff for it, on a separate branch maybe, because opportunities present themselves — for example if Cats or Cats-Effect break binary compatibility, then we can do so too. I don't predict that will happen soon though and it is my wish for Monix to be fairly stable going forward. Good points on the monix.io documentation and integrating with Java's Thread.interrupt. We'll open separate issues for these and for bifunctors and discuss implementation ideas there. I will jump in only as a user but bifunctor and stacktraces are two game changers I would like to have. And similar to @Avasil, in my usecases performance is the last concern. Hey there! Great to see this discussion has started. I'd like to say that performance is an important quality of Monix and it's quite important for me! I believe some of the items in this list are great (better stack traces are important), but I would love to see more PRs to make Monix even more performant than it is today. I use Monix heavily in Bloop for fast scheduling of various tasks on machines with lots of cores. For example, we're heavy user of observer/observables to model many of the performance-sensitive build server internals. I haven't seen obvious performance bottlenecks with the current implementation --I haven't really profiled it either-- but the faster scheduling, context management and task creation is the faster we will build Scala code. Building code fast is pretty much all Bloop needs to do and we rely heavily on Monix to not get in the way of that (and we're really satisfied with it so far!). I'm planning to run some benchmarks when we upgrade to 3.0.0 to see if the latest merged optimizations have an effect in our particular use of Monix APIs. That should shed more light on whether 3.0.0 is fast enough and we would/would not benefit from further performance work. Not sure if it'll be an apple-to-apples comparison but will offer us some insights. Until I run those experiments though, performance work always gets a :+1: from me -- I love it that the use of powerful Monix abstractions doesn't come with a runtime cost. It allows people to use Monix in areas where it has traditionally not been able to model programs with such a level of abstraction (e.g. developer tools/compilers/who knows what else). Thanks for all of the work you guys put into this library and hope that whatever you end up doing you keep pushing the limits of what Monix currently enables Scala developers to do and how they do it. Thanks @jvican, performance is always a priority. And in case you see any regressions after an upgrade, it's a mistake that can be rectified. Personal priority (in order) for our use case at work is Stack traces that work with Task/Coeval Bifunctor (not EitherT, using that is quite painful) @mdedetrich 👍
2025-04-01T04:34:48.043588
2024-03-05T23:50:57
2170362985
{ "authors": [ "monkeymademe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8785", "repo": "monkeymademe/picamera2-WebUI", "url": "https://github.com/monkeymademe/picamera2-WebUI/issues/27" }
gharchive/issue
New Resolutions don't account for square formats The latest (v0.0.4) version of the webui added in selectable resolutions which are based off the resolution of the v3 camera. While this is ok for the v3 and likely the v2. There is squishing issues with v1, HQ and GS that needs to be addressed. I think a more dynamic resolution selection setting needs to be made and a custom resolution input so that a user can force what they need. Raw 'should' remain unaffected as the resolution is set by the sensor mode Fixed in v1.0.0
2025-04-01T04:34:48.051298
2024-01-31T07:04:39
2109424357
{ "authors": [ "magency-prod", "monkoose" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8786", "repo": "monkoose/nvlime", "url": "https://github.com/monkoose/nvlime/issues/4" }
gharchive/issue
localleader cc does not work the built in keybind localleader cc connect to repl does not work at all press the button and nothing happens however, localleader rr works, and other keybinds i dug in the sourcecode and found the line :call nvlime#plugin#ConnectREPL() this line however works for some reason after I entered it in nvim, it askes me the server ip and port, and it connects my config for lazy: { "monkoose/nvlime", ft = { "lisp" }, dependencies = { "monkoose/parsley", }, init = function() vim.g.nvlime_config = { leader = "<LocalLeader>", implementation = "sbcl", user_contrib_initializers = nil, autodoc = { enabled = false, max_level = 5, max_lines = 50, }, main_window = { position = "right", size = "", }, floating_window = { border = "single", scroll_step = 3, }, cmp = { enabled = true }, arglist = { enabled = true }, } end, }, Can't reproduce. For me it works as expected. Make sure that any other plugin isn't overwriting this map. Check it with :nmap <localleader>cc. oh right something was overriding the mapping i though the project was dead for a second It isn't dead, but development happens with some gaps. And for some reason I do not receive notifications from this repo sometimes, so I have missed this issue.
2025-04-01T04:34:48.055918
2015-07-07T18:47:30
93601402
{ "authors": [ "ddobrev", "genuinelucifer", "xistoso" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8787", "repo": "mono/CppSharp", "url": "https://github.com/mono/CppSharp/issues/514" }
gharchive/issue
Passing null to a function pointer causes a crash We don't check what we pass to Marshal.GetFunctionPointerForDelegate. To reproduce, set CSharpTemp.Foo.AttributedFunctionPtr to null. @ddobrev Please clarify a bit more. It is a typedef. SO how to set it to null? @genuinelucifer it's not a typedef - that's how one defines a field with a function pointer type in C++. So that field is wrapped as a property in C# - that's what I meant by setting it to null. @genuinelucifer in CSharpTemp.Tests.cs use this test and you can see it explode [Test] public void TestNullAttributedFunctionPtr() { var foo = new Foo(); foo.AttributedFunctionPtr = null; } @ddobrev @xistoso Thanks.
2025-04-01T04:34:48.061780
2017-04-27T21:37:36
224920576
{ "authors": [ "dalexsoto", "monojenkins", "rolfbjarne" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8788", "repo": "mono/Embeddinator-4000", "url": "https://github.com/mono/Embeddinator-4000/pull/203" }
gharchive/pull-request
[objc][tests] Add objcgen warning and error tests. Fixes issue #104 Adds support for testing error conditions in Xcode projects see XcodeBuildErrorTest inside ObjCGenErrWarnTests.cs Adds a library to create warning conditions to test objcgen and its correponding test (GenWarningTest inside ObjCGenErrWarnTests.cs). Build and tests passed. Build and tests passed. Build and tests passed. Build and tests passed. Build and tests passed. @spouliot My reasoning behind keeping the Xcode project is to delegate the task of figuring out all clang arguments to it, by doing so we can support tvOS, iOS, watchOS, macOS configs, header search paths, etc. Without worrying about all different configs in our unit tests and also another point is that this is closer to how our customers will integrate our E4K bits. But I am open to suggestions 😄 Build and tests passed. We'll need a new Xcode project for every platform (one for iOS, one for tvOS, one for macOS, one for watchOS...), so I'm not sure it's easier than invoking clang directly. Also personally I find it much harder to figure out how to massage an Xcode project into giving clang the argument I want to give it than constructing the clang command-line manually... IMHO the only time it makes sense to have an Xcode project is when we want to test the entire integration with Xcode (in this case we're only interested in what clang says about the generated code), or for when we want to be able to debug the generated code with lldb (in this case we won't, because we're not executing the generated code). Since this is two opinions vs one on removing the Xcode project and just use clang 😛 IMHO the only time it makes sense to have an Xcode project is when we want to test the entire integration with Xcode (in this case we're only interested in what clang says about the generated code), or for when we want to be able to debug the generated code with lldb (in this case we won't, because we're not executing the generated code). Question: do we really need the Xcode project ? or could we just call clang and get the warnings out of stdout ? and I agree on @rolfbjarne's arguments I'll remove the Xcode project and will use Clang instead. Build and tests passed. @rolfbjarne fixed, using xcrun clang now Build and tests passed.
2025-04-01T04:34:48.070099
2024-04-27T15:54:00
2267071590
{ "authors": [ "dsyno", "mattleibow" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8789", "repo": "mono/SkiaSharp", "url": "https://github.com/mono/SkiaSharp/issues/2849" }
gharchive/issue
[BUG] SKBitmap.ColorSpace is always null. Unable to check for CMYK, RGB, etc. Description Our code needs to check whether an image is CMYK or RGB (and reject any image that is CMYK). However, @mattleibow "explicitly set the color space to null for backwards compatibility". That is breaking things, since it now seems to be impossible (without jumping through major hoops) to check whether an image is CMYK or not. Solution: Allow ColorSpace to return the correct value. Or provide another method for checking the color space. Code SKBitmap.ColorSpace Expected Behavior SKBitmap.ColorSpace should return correct value, instead of null. Actual Behavior SKBitmap.ColorSpace always returns null. Version of SkiaSharp 2.88.3 (Current) Last Known Good Version of SkiaSharp Other (Please indicate in the description) IDE / Editor Visual Studio (Windows) Platform / Operating System All Platform / Operating System Version All Devices All Relevant Screenshots No response Relevant Log Output No response Code of Conduct [x] I agree to follow this project's Code of Conduct We could remove that for 3.0. You could use SKImage as well - or use SKCodec directly and SKBitmap.Decode. That will not remove the colorspace. I just checked 3.0 and that already has the comapt line removed. Have you tried 3.x at all? I just checked 3.0 and that already has the comapt line removed. Have you tried 3.x at all? I haven't because 3.0 is still in a preview release. We can try it once 3.0 becomes a stable version. Any ETA for a 3.0 stable?
2025-04-01T04:34:48.085649
2023-09-04T22:53:34
1880937248
{ "authors": [ "ToolmakerSteve", "TopperDEL", "bogdan799", "levinli303", "marcojak", "mattleibow", "r2d2Proton" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8790", "repo": "mono/SkiaSharp", "url": "https://github.com/mono/SkiaSharp/pull/2598" }
gharchive/pull-request
Implement SKGLView for .NET MAUI Description of Change Implemented SKGLView for Android and iOS Mac Catalyst will not get support - a new Metal-based view needs to be created Tizen needs the support in the native view Windows does not yet have GPU support Bugs Fixed Fixes #2050 API Changes None. Behavioral Changes None. Required skia PR None. PR Checklist [ ] Has tests (if omitted, state reason in description) [ ] Rebased on top of main at time of PR [ ] Merged related skia PRs [ ] Changes adhere to coding standard [ ] Updated documentation Hi Matthew, are there any plans to have a common hw accelerated view for a MAUI app for Android, iOS, Windows and Mac? If not possible, how could I use an hw accelerated view on Windows? (and maybe a new one for Mac)? @marcojak This PR might be related for you: https://github.com/mono/SkiaSharp/issues/1893 Definitely would like this to work on all the platforms being targeted. Windows, Mac, etc. @mattleibow I've sort of tested implementation of SKGLView on ios and android. I needed metal on iOS, so I've just replaced native view for SKGLViewHandler.iOS and seem to be fine, could be considered while implementing universal HW accelerated view. All seems to be much smoother than CPU rendering for my usecase, thank you for the solution. However I've noticed following issue on android: Here when setting Rendermode to RenderMode.WhenDirty, repaint doesn't happen althouth I'm invalidating view many times: SKGLViewHandler.Android.cs public static void MapHasRenderLoop(SKGLViewHandler handler, ISKGLView view) { handler.PlatformView.RenderMode = view.HasRenderLoop ? Rendermode.Continuously : Rendermode.WhenDirty; } I checked following handler function, and seems you're using wrong method here: // Mapper actions / properties public static void OnInvalidateSurface(SKGLViewHandler handler, ISKGLView view, object? args) { handler.PlatformView?.Invalidate(); } According to MSDN: Continuously 1 The renderer is called continuously to re-render the scene. WhenDirty 0 The renderer only renders when the surface is created, or when RequestRender() is called. Here's the change I've applied according to MSDN and it seem to work fine: public static void OnInvalidateSurface(SKGLViewHandler handler, ISKGLView view, object? args) { handler.PlatformView?.RequestRender(); } Thanks @bogdan799! Your findings helped. Mac Catalyst has CAOpenGLLayer. You can have a UIView backed by CAOpenGLLayer and render OpenGL in the layer so that you don't have to mess with Metal. I am not sure if it is exposed in the .NET bindings. They left out all the deprecated APIs for Mac Catalyst :( @mattleibow I am getting the exception "Handler not found for view SkiaSharp.Views.Maui.Controls.SKGLView." in a fresh .Net Maui App project, targetting "Windows Machine", with an SKGLView added. Tested with Microsoft.Maui.Controls 8.0.40 + SkiaSharp.Views & SkiaSharp.Views.Maui Controls latest stable 2.88.8, then tried latest preview 3.0.0-preview.3.1 Adding these two lines in appropriate places in MainPage.xaml: xmlns:sk="clr-namespace:SkiaSharp.Views.Maui.Controls;assembly=SkiaSharp.Views.Maui.Controls" <sk:SKGLView WidthRequest="500" HeightRequest="300" BackgroundColor="Green" /> Is this expected to work in Maui 8, or do I need a preview version of .Net Maui 9 Preview? Should I create a new issue and/or upload a sample project? Please open a new issue with a repro sample. GL should be working. https://github.com/mono/SkiaSharp/issues/2860 @mattleibow - is there a plan to backport #2598 (Maui SKGLView fix) on Windows to 2.88.x? Updating my client from .Net Framework 4.5 to .Net 8. We've successfully updated on iOS and Android. (WITHOUT Maui; these are "platform native" .Net 8 apps, with GUI rendered via SkiaSharp. Plus some OpenGL.) On Windows .Net 8, SkiaSharp 2.88.8 or .9, we've got everything EXCEPT OpenGL working. Using Maui SKCanvasView. I made a new Maui app. Added SkiaSharp 2.88.9. Remembered to do "UseSkiaSharp()". Still get "Handler not found for view SkiaSharp.Views.Maui.Controls.SKGLView.". Updating to SkiaSharp <IP_ADDRESS> preview, it works. BUT this app is used in a production situation where failures must be very rare. So I can't use anything that isn't considered a stable release.
2025-04-01T04:34:48.135661
2019-11-11T19:28:48
521143286
{ "authors": [ "monperrus", "therrick" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8791", "repo": "monperrus/crawler-user-agents", "url": "https://github.com/monperrus/crawler-user-agents/pull/252" }
gharchive/pull-request
add Bytespider We see a large amount of bot traffic with agents ending with 'Bytespider'. One example: "Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/55.0.5939.1534 Mobile Safari/537.36; Bytespider". I started to include an instances list, but then realized it presents with literally thousands of different user agent strings. see: https://stackoverflow.com/questions/57908900/what-is-the-bytespider-user-agent https://udger.com/resources/ua-list/bot-detail?bot=ByteDance+crawler I couldn't find any official web page for this crawler to include in the JSON for this crawler. Thanks a lot for your contribution.
2025-04-01T04:34:48.138230
2024-12-24T04:27:31
2757146351
{ "authors": [ "bintanghd", "monstrenyatko" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8792", "repo": "monstrenyatko/ArduinoMqtt", "url": "https://github.com/monstrenyatko/ArduinoMqtt/issues/35" }
gharchive/issue
qos0 send more than one, up to 5x I am using ethernet modul w5500 and arduinomqtt with default qos (which is 0). but my broker sometime receive upto 5x same message. any ideas how this could happen?and to resolve this. thank you Try enabling logging #define MQTT_LOG_ENABLED 1 to check the library behavior. The library publish implementation is super simple: no retries - just minimal message processing for QoS > 0. See publish will do, but unfortunately this happens just occasional. not every time. so it would be bit difficut to capture anomali There are no retries in the library publish implementation. It could be a retry caused by your network stack (see ethernet module w5500). If message duplication is the problem, try publishing using QoS1, in this case each message has ID that potentially can be used for de-duplication by the broker or receiver.
2025-04-01T04:34:48.153855
2016-09-16T21:27:22
177524193
{ "authors": [ "JeroenVinke", "adriatic" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8793", "repo": "monterey-framework/monterey", "url": "https://github.com/monterey-framework/monterey/issues/314" }
gharchive/issue
running npm start for "skeleton-esnext-aspnetcore" shows four errors Intention I'm submitting a bug report Environment Project name skeleton-typescript-webpack Project path: C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack NodeJS: 6.5.0 NPM: 3.10.7 Electron: 1.4.0 Chrome: 53.0.2785.113 Operating system: win32 Project is using NPM Package.json path: C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack\package.json Project is using Webpack Webpack.config.js path: C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack\webpack.config.js Current behavior: Observe the checking errors at the end of the terminal output (below) [16:58:41] -----STARTED----- [16:58:45] ><EMAIL_ADDRESS>start C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack [16:58:45] > npm run server:dev [16:58:46] ><EMAIL_ADDRESS>server:dev C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack [16:58:46] > cross-env NODE_ENV=development node ./node_modules/webpack-dev-server/bin/webpack-dev-server --inline --progress --profile --watch [16:58:57] http://localhost:9000/ [16:58:57] webpack result is served from / [16:58:57] content is served from C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack [16:58:57] 404s will fallback to /index.html [16:59:16] Hash: 775037560e5b9c027bbc [16:59:16] Version: webpack 2.1.0-beta.22 [16:59:16] Time: 19010ms [16:59:16] Asset Size Chunks Chunk Names [16:59:16] c8ddf1e5e5bf3682bc7bebf30f394148.woff 90.4 kB [emitted] [16:59:16] f4769f9bdb7466be65088239c12046d1.eot 20.1 kB [emitted] [16:59:16] e18bbf611f2a2e43afc071aa2f4e1512.ttf 45.4 kB [emitted] [16:59:16] fa2772327f55d8198301fdb8bcfc8158.woff 23.4 kB [emitted] [16:59:16] 448c34a56d699c29117adc64c43affeb.woff2 18 kB [emitted] [16:59:16] 25a32416abee198dd821b0b17a198a8f.eot 76.5 kB [emitted] [16:59:16] d7c639084f684d66a1bc66855d193ed8.svg 392 kB [emitted] [16:59:16] 1dc35d25e61d819a9c357074014867ab.ttf 153 kB [emitted] [16:59:16]<PHONE_NUMBER>7bd7575d6327160d64e760.svg 109 kB [emitted] [16:59:16] e6cf7c6ec7c2d6f670ae9d762604cb0b.woff2 71.9 kB [emitted] [16:59:16] app.bundle.js 1.04 MB 0 [emitted] app [16:59:16] aurelia.bundle.js 1.46 MB 1 [emitted] aurelia [16:59:16] aurelia-bootstrap.bundle.js 1.35 MB 2 [emitted] aurelia-bootstrap [16:59:16] styles.css 196 kB 0 [emitted] app [16:59:16] favicon.ico 2.25 kB [emitted] [16:59:16] index.html 775 bytes [emitted] [16:59:16] chunk {0} app.bundle.js, styles.css (app) 385 kB {1} [initial] [rendered] [61] ./src/blur-image.ts 13.7 kB {0} [optional] [built] [] -> factory:2688ms building:6961ms dependencies:0ms = 9649ms [81] ./src/app.ts 831 bytes {0} [optional] [built] [] -> factory:2687ms building:6983ms = 9670ms [82] ./src/child-router.ts 885 bytes {0} [optional] [built] [] -> factory:2688ms building:6613ms = 9301ms [83] ./src/main.ts 3.07 kB {0} [optional] [built] [] -> factory:2688ms building:7039ms dependencies:2ms = 9729ms [84] ./src/users.ts 4.39 kB {0} [optional] [built] [] -> factory:2688ms building:6745ms dependencies:1ms = 9434ms [85] ./src/welcome.ts 2.38 kB {0} [optional] [built] [] -> factory:2688ms building:7074ms = 9762ms [87] ./~/bootstrap/js/affix.js 4.84 kB {0} [built] [] -> factory:7045ms building:150ms dependencies:2ms = 7197ms [88] ./~/bootstrap/js/alert.js 2.28 kB {0} [built] [] -> factory:7042ms building:133ms dependencies:21ms = 7196ms [89] ./~/bootstrap/js/button.js 3.82 kB {0} [built] [] -> factory:7043ms building:115ms dependencies:38ms = 7196ms [90] ./~/bootstrap/js/carousel.js 7.14 kB {0} [built] [] -> factory:7043ms building:141ms dependencies:13ms = 7197ms [91] ./~/bootstrap/js/collapse.js 5.99 kB {0} [built] [] -> factory:7043ms building:147ms dependencies:7ms = 7197ms [92] ./~/bootstrap/js/dropdown.js 4.74 kB {0} [built] [] -> factory:7043ms building:137ms dependencies:16ms = 7196ms [93] ./~/bootstrap/js/modal.js 9.99 kB {0} [built] [] -> factory:7043ms building:135ms dependencies:18ms = 7196ms [94] ./~/bootstrap/js/popover.js 3.16 kB {0} [built] [] -> factory:7044ms building:137ms dependencies:16ms = 7197ms [95] ./~/bootstrap/js/scrollspy.js 4.71 kB {0} [built] [] -> factory:7044ms building:148ms dependencies:5ms = 7197ms [96] ./~/bootstrap/js/tab.js 3.9 kB {0} [built] [] -> factory:7044ms building:149ms dependencies:4ms = 7197ms [97] ./~/bootstrap/js/tooltip.js 16.7 kB {0} [built] [] -> factory:7044ms building:145ms dependencies:8ms = 7197ms [98] ./~/bootstrap/js/transition.js 1.83 kB {0} [built] [] -> factory:7042ms building:118ms dependencies:36ms = 7196ms [101] ./~/expose-loader?jQuery!./~/jquery/dist/jquery.js 160 bytes {0} [built] [] -> factory:2152ms building:0ms dependencies:1ms = 2153ms [102] ./~/bootstrap/dist/css/bootstrap.css 41 bytes {0} [built] [] -> factory:241ms building:4441ms = 4682ms [103] ./~/font-awesome/css/font-awesome.css 41 bytes {0} [built] [] -> factory:242ms building:4487ms = 4729ms [104] ./styles/styles.css 41 bytes {0} [built] [] -> factory:206ms building:4236ms = 4442ms [105] ./src/app.html 216 bytes {0} [optional] [built] [] -> factory:535ms building:2184ms = 2719ms [106] ./src/child-router.html 533 bytes {0} [optional] [built] [] -> factory:2681ms building:38ms = 2719ms [107] ./src/nav-bar.html 1.27 kB {0} [optional] [built] [] -> factory:2682ms building:38ms = 2720ms [108] ./src/users.html 808 bytes {0} [optional] [built] [] -> factory:2682ms building:38ms = 2720ms [109] ./src/welcome.html 815 bytes {0} [optional] [built] [] -> factory:2682ms building:38ms = 2720ms [110] ./~/jquery/dist/jquery.js 264 kB {0} [built] [] -> factory:0ms building:7195ms = 7195ms [152] ./~/whatwg-fetch/fetch.js 11.6 kB {0} [built] [] -> factory:7166ms building:52ms dependencies:38ms = 7256ms [153] ./src ^\.\/.*$ 2.43 kB {0} [built] [] -> factory:177ms building:44ms dependencies:26ms = 247ms [154] multi app 28 bytes {0} [built] [16:59:16] factory:0ms building:0ms = 0ms [aurelia-fetch-client] ./~/aurelia-fetch-client/dist/commonjs/aurelia-fetch-client.js 7.56 kB {0} [built] [] -> factory:2683ms building:41ms dependencies:6946ms = 9670ms [bootstrap] ./~/bootstrap/dist/js/npm.js 484 bytes {0} [built] [] -> factory:2683ms building:42ms = 2725ms [isomorphic-fetch] ./~/isomorphic-fetch/fetch-npm-browserify.js 233 bytes {0} [built] [] -> factory:505ms building:2209ms = 2714ms [jquery] ./~/jquery/dist/jquery.js 273 bytes {0} [built] [] -> factory:532ms building:0ms dependencies:1ms = 533ms [16:59:16] chunk {1} aurelia.bundle.js (aurelia) 554 kB {2} [initial] [rendered] [4] ./~/aurelia-templating/dist/commonjs/aurelia-templating.js 148 kB {1} [built] [] -> factory:222ms building:441ms dependencies:55ms = 718ms [5] ./~/aurelia-dependency-injection/dist/commonjs/aurelia-dependency-injection.js 20.7 kB {1} [built] [] -> factory:200ms building:148ms dependencies:322ms = 670ms [6] ./~/aurelia-binding/dist/commonjs/aurelia-binding.js 166 kB {1} [built] [] -> factory:200ms building:287ms dependencies:184ms = 671ms [10] ./~/aurelia-logging/dist/commonjs/aurelia-logging.js 2.11 kB {1} [built] [] -> factory:203ms building:170ms = 373ms [14] ./~/aurelia-task-queue/dist/commonjs/aurelia-task-queue.js 3.21 kB {1} [built] [] -> factory:222ms building:288ms dependencies:168ms = 678ms [15] ./~/aurelia-templating-resources/dist/commonjs/repeat-utilities.js 3 kB {1} [built] [] -> factory:172ms building:132ms dependencies:0ms = 304ms [26] ./~/aurelia-history/dist/commonjs/aurelia-history.js 810 bytes {1} [built] [] -> factory:201ms building:488ms = 689ms [28] ./~/aurelia-templating-resources/dist/commonjs/aurelia-hide-style.js 829 bytes {1} [built] [] -> factory:172ms building:133ms dependencies:0ms = 305ms [29] ./~/aurelia-templating-resources/dist/commonjs/binding-mode-behaviors.js 1.35 kB {1} [built] [] -> factory:169ms building:103ms dependencies:0ms = 272ms [30] ./~/aurelia-templating-resources/dist/commonjs/compose.js 6.22 kB {1} [built] [] -> factory:167ms building:89ms dependencies:2ms = 258ms [31] ./~/aurelia-templating-resources/dist/commonjs/debounce-binding-behavior.js 1.73 kB {1} [built] [] -> factory:170ms building:103ms dependencies:0ms = 273ms [32] ./~/aurelia-templating-resources/dist/commonjs/focus.js 2.1 kB {1} [built] [] -> factory:168ms building:101ms dependencies:0ms = 269ms [33] ./~/aurelia-templating-resources/dist/commonjs/hide.js 1.38 kB {1} [built] [] -> factory:168ms building:98ms dependencies:0ms = 266ms [34] ./~/aurelia-templating-resources/dist/commonjs/if.js 2.99 kB {1} [built] [] -> factory:167ms building:92ms dependencies:1ms = 260ms [35] ./~/aurelia-templating-resources/dist/commonjs/repeat.js 11.2 kB {1} [built] [] -> factory:168ms building:95ms dependencies:1ms = 264ms [36] ./~/aurelia-templating-resources/dist/commonjs/replaceable.js 1.14 kB {1} [built] [] -> factory:168ms building:99ms dependencies:0ms = 267ms [37] ./~/aurelia-templating-resources/dist/commonjs/sanitize-html.js 983 bytes {1} [built] [] -> factory:168ms building:97ms dependencies:0ms = 265ms [38] ./~/aurelia-templating-resources/dist/commonjs/show.js 1.38 kB {1} [built] [] -> factory:168ms building:96ms dependencies:1ms = 265ms [39] ./~/aurelia-templating-resources/dist/commonjs/signal-binding-behavior.js 1.88 kB {1} [built] [] -> factory:170ms building:104ms dependencies:0ms = 274ms [40] ./~/aurelia-templating-resources/dist/commonjs/throttle-binding-behavior.js 1.92 kB {1} [built] [] -> factory:169ms building:104ms dependencies:0ms = 273ms [41] ./~/aurelia-templating-resources/dist/commonjs/update-trigger-binding-behavior.js 1.88 kB {1} [built] [] -> factory:170ms building:118ms dependencies:0ms = 288ms [42] ./~/aurelia-templating-resources/dist/commonjs/with.js 1.55 kB {1} [built] [] -> factory:167ms building:93ms dependencies:1ms = 261ms [43] ./~/aurelia-templating-router/dist/commonjs/route-href.js 2.37 kB {1} [built] [] -> factory:182ms building:78ms dependencies:3ms = 263ms [44] ./~/aurelia-templating-router/dist/commonjs/router-view.js 8.75 kB {1} [built] [] -> factory:182ms building:88ms dependencies:1ms = 271ms [51] ./~/aurelia-templating-resources/dist/commonjs/abstract-repeater.js 1.96 kB {1} [built] [] -> factory:170ms building:120ms = 290ms [52] ./~/aurelia-templating-resources/dist/commonjs/analyze-view-factory.js 1.79 kB {1} [built] [] -> factory:172ms building:133ms = 305ms [53] ./~/aurelia-templating-resources/dist/commonjs/array-repeat-strategy.js 8.31 kB {1} [built] [] -> factory:171ms building:127ms dependencies:0ms = 298ms [54] ./~/aurelia-templating-resources/dist/commonjs/binding-signaler.js 583 bytes {1} [built] [] -> factory:170ms building:119ms dependencies:0ms = 289ms [55] ./~/aurelia-templating-resources/dist/commonjs/html-sanitizer.js 397 bytes {1} [built] [] -> factory:169ms building:102ms = 271ms [56] ./~/aurelia-templating-resources/dist/commonjs/map-repeat-strategy.js 3.7 kB {1} [built] [] -> factory:171ms building:129ms dependencies:0ms = 300ms [57] ./~/aurelia-templating-resources/dist/commonjs/null-repeat-strategy.js 473 bytes {1} [built] [] -> factory:171ms building:127ms = 298ms [58] ./~/aurelia-templating-resources/dist/commonjs/number-repeat-strategy.js 1.8 kB {1} [built] [] -> factory:171ms building:131ms dependencies:0ms = 302ms [59] ./~/aurelia-templating-resources/dist/commonjs/repeat-strategy-locator.js 1.76 kB {1} [built] [] -> factory:171ms building:120ms dependencies:1ms = 292ms [60] ./~/aurelia-templating-resources/dist/commonjs/set-repeat-strategy.js 3.1 kB {1} [built] [] -> factory:171ms building:130ms dependencies:1ms = 302ms [77] ./~/aurelia-templating-resources/dist/commonjs/css-resource.js 3.76 kB {1} [built] [] -> factory:169ms building:100ms dependencies:2ms = 271ms [78] ./~/aurelia-templating-resources/dist/commonjs/dynamic-element.js 834 bytes {1} [built] [] -> factory:316ms building:2270ms dependencies:0ms = 2586ms [79] ./~/aurelia-templating-resources/dist/commonjs/html-resource-plugin.js 1.2 kB {1} [built] [] -> factory:171ms building:122ms dependencies:0ms = 293ms [80] ./~/aurelia-templating-router/dist/commonjs/route-loader.js 2.49 kB {1} [built] [] -> factory:182ms building:82ms dependencies:2ms = 266ms [155] multi aurelia 256 bytes {1} [built] [16:59:16] factory:0ms building:0ms dependencies:150ms = 150ms [aurelia-event-aggregator] ./~/aurelia-event-aggregator/dist/commonjs/aurelia-event-aggregator.js 3.3 kB {1} [built] [] -> factory:201ms building:150ms dependencies:319ms = 670ms [aurelia-framework] ./~/aurelia-framework/dist/commonjs/aurelia-framework.js 16.8 kB {1} [built] [] -> factory:202ms building:164ms dependencies:352ms = 718ms [aurelia-history-browser] ./~/aurelia-history-browser/dist/commonjs/aurelia-history-browser.js 9.37 kB {1} [built] [] -> factory:202ms building:116ms dependencies:352ms = 670ms [aurelia-logging-console] ./~/aurelia-logging-console/dist/commonjs/aurelia-logging-console.js 1.67 kB {1} [built] [] -> factory:203ms building:288ms dependencies:181ms = 672ms [aurelia-route-recognizer] ./~/aurelia-route-recognizer/dist/commonjs/aurelia-route-recognizer.js 11.6 kB {1} [built] [] -> factory:220ms building:288ms dependencies:170ms = 678ms [aurelia-router] ./~/aurelia-router/dist/commonjs/aurelia-router.js 54.3 kB {1} [built] [] -> factory:223ms building:365ms dependencies:130ms = 718ms [aurelia-templating-binding] ./~/aurelia-templating-binding/dist/commonjs/aurelia-templating-binding.js 24.5 kB {1} [built] [] -> factory:221ms building:303ms dependencies:156ms = 680ms [aurelia-templating-resources] ./~/aurelia-templating-resources/dist/commonjs/aurelia-templating-resources.js 5.28 kB {1} [built] [] -> factory:224ms building:304ms dependencies:155ms = 683ms [aurelia-templating-router] ./~/aurelia-templating-router/dist/commonjs/aurelia-templating-router.js 862 bytes {1} [built] [] -> factory:223ms building:289ms dependencies:166ms = 678ms [16:59:16] chunk {2} aurelia-bootstrap.bundle.js (aurelia-bootstrap) 500 kB [entry] [rendered] [0] ./~/inherits/inherits_browser.js 672 bytes {2} [built] [] -> factory:352ms building:551ms = 903ms [1] (webpack)/buildin/global.js 506 bytes {2} [built] [] -> factory:530ms building:387ms = 917ms [2] ./~/debug/browser.js 3.76 kB {2} [built] [] -> factory:386ms building:543ms = 929ms [3] ./~/aurelia-pal/dist/commonjs/aurelia-pal.js 1.97 kB {2} [built] [] -> factory:198ms building:56ms dependencies:507ms = 761ms [7] ./~/sockjs-client/lib/event/emitter.js 1.27 kB {2} [built] [] -> factory:486ms building:73ms dependencies:0ms = 559ms [8] ./~/sockjs-client/lib/utils/url.js 975 bytes {2} [built] [] -> factory:17ms building:343ms dependencies:36ms = 396ms [9] ./~/aurelia-metadata/dist/commonjs/aurelia-metadata.js 6.67 kB {2} [built] [] -> factory:203ms building:176ms dependencies:292ms = 671ms [11] ./~/sockjs-client/~/json3/lib/json3.js 43.3 kB {2} [built] [] -> factory:880ms building:203ms dependencies:2ms = 1085ms [12] ./~/sockjs-client/lib/utils/event.js 2 kB {2} [built] [] -> factory:17ms building:349ms dependencies:1ms = 367ms [13] ./~/aurelia-path/dist/commonjs/aurelia-path.js 5.46 kB {2} [built] [] -> factory:204ms building:291ms = 495ms [16] ./~/sockjs-client/lib/transport/lib/ajax-based.js 1.31 kB {2} [built] [] -> factory:56ms building:185ms dependencies:2ms = 243ms [17] ./~/sockjs-client/lib/utils/random.js 746 bytes {2} [built] [] -> factory:16ms building:343ms = 359ms [18] ./~/aurelia-loader/dist/commonjs/aurelia-loader.js 4.58 kB {2} [built] [] -> factory:202ms building:152ms dependencies:316ms = 670ms [19] (webpack)-dev-server/client?http://localhost:9000 4.18 kB {2} [built] [] -> factory:150ms building:94ms = 244ms [20] ./~/sockjs-client/lib/transport/sender/xhr-local.js 352 bytes {2} [built] [] -> factory:499ms building:174ms dependencies:26ms = 699ms [21] ./~/sockjs-client/lib/utils/browser.js 560 bytes {2} [built] [] -> factory:18ms building:354ms dependencies:0ms = 372ms [22] ./~/sockjs-client/lib/utils/iframe.js 5.06 kB {2} [built] [] -> factory:521ms building:79ms dependencies:1ms = 601ms [23] ./~/sockjs-client/lib/transport/receiver/xhr.js 1.58 kB {2} [built] [] -> factory:65ms building:159ms dependencies:18ms = 242ms [24] ./~/sockjs-client/lib/transport/sender/xhr-cors.js 343 bytes {2} [built] [] -> factory:499ms building:185ms dependencies:40ms = 724ms [25] ./~/timers-browserify/main.js 2.1 kB {2} [built] [] -> factory:259ms building:478ms dependencies:2186ms = 2923ms [27] ./~/regenerator-runtime/runtime-module.js 1.01 kB {2} [built] [] -> factory:172ms building:59ms = 231ms [45] ./~/process/browser.js 5.3 kB {2} [built] [] -> factory:415ms building:2284ms = 2699ms [46] ./~/sockjs-client/lib/event/event.js 477 bytes {2} [built] [] -> factory:18ms building:356ms = 374ms [47] ./~/sockjs-client/lib/transport/lib/iframe-wrap.js 981 bytes {2} [built] [] -> factory:269ms building:622ms dependencies:22ms = 913ms [48] ./~/sockjs-client/lib/transport/sender/xdr.js 2.46 kB {2} [built] [] -> factory:498ms building:172ms dependencies:28ms = 698ms [49] ./~/sockjs-client/lib/utils/object.js 532 bytes {2} [built] [] -> factory:18ms building:349ms = 367ms [50] ./~/url-parse/index.js 8.55 kB {2} [built] [] -> factory:352ms building:560ms = 912ms [62] ./~/sockjs-client/lib/event/eventtarget.js 1.85 kB {2} [built] [] -> factory:19ms building:355ms = 374ms [63] ./~/sockjs-client/lib/info-ajax.js 1.03 kB {2} [built] [] -> factory:492ms building:193ms dependencies:0ms = 685ms [64] ./~/sockjs-client/lib/info-iframe-receiver.js 791 bytes {2} [built] [] -> factory:521ms building:89ms dependencies:0ms = 610ms [65] ./~/sockjs-client/lib/location.js 177 bytes {2} [built] [] -> factory:12ms building:275ms dependencies:0ms = 287ms [66] ./~/sockjs-client/lib/transport/browser/abstract-xhr.js 4.8 kB {2} [built] [] -> factory:51ms building:45ms dependencies:4ms = 100ms [67] ./~/sockjs-client/lib/transport/browser/eventsource.js 37 bytes {2} [built] [] -> factory:62ms building:163ms dependencies:0ms = 225ms [68] ./~/sockjs-client/lib/transport/eventsource.js 766 bytes {2} [built] [] -> factory:267ms building:593ms dependencies:51ms = 911ms [69] ./~/sockjs-client/lib/transport/htmlfile.js 710 bytes {2} [built] [] -> factory:267ms building:587ms dependencies:20ms = 874ms [70] ./~/sockjs-client/lib/transport/iframe.js 3.83 kB {2} [built] [] -> factory:37ms building:164ms dependencies:3ms = 204ms [71] ./~/sockjs-client/lib/transport/lib/sender-receiver.js 1.15 kB {2} [built] [] -> factory:36ms building:175ms dependencies:1ms = 212ms [72] ./~/sockjs-client/lib/transport/xdr-streaming.js 984 bytes {2} [built] [] -> factory:267ms building:589ms dependencies:54ms = 910ms [73] ./~/sockjs-client/lib/transport/xhr-polling.js 894 bytes {2} [built] [] -> factory:268ms building:607ms dependencies:46ms = 921ms [74] ./~/sockjs-client/lib/version.js 26 bytes {2} [built] [] -> factory:12ms building:276ms = 288ms [75] (webpack)/buildin/module.js 548 bytes {2} [built] [] -> factory:0ms building:0ms = 0ms [76] ./~/ansi-regex/index.js 135 bytes {2} [built] [] -> factory:9235ms building:119ms = 9354ms [86] ./~/bluebird/js/browser/bluebird.js 178 kB {2} [built] [] -> factory:0ms building:507ms dependencies:27ms = 534ms [99] ./~/debug/debug.js 4.1 kB {2} [built] [] -> factory:19ms building:166ms = 185ms [100] ./~/events/events.js 8.33 kB {2} [built] [] -> factory:2290ms building:7060ms = 9350ms [111] ./~/ms/index.js 2.33 kB {2} [built] [] -> factory:48ms building:1ms = 49ms [112] ./~/punycode/punycode.js 14.7 kB {2} [built] [] -> factory:2207ms building:7055ms dependencies:1ms = 9263ms [113] ./~/querystring-es3/decode.js 2.51 kB {2} [built] [] -> factory:108ms building:46ms = 154ms [114] ./~/querystring-es3/encode.js 2.54 kB {2} [built] [] -> factory:108ms building:3406ms = 3514ms [115] ./~/querystring-es3/index.js 127 bytes {2} [built] [] -> factory:2207ms building:7046ms = 9253ms [116] ./~/querystringify/index.js 1.3 kB {2} [built] [] -> factory:218ms building:32ms = 250ms [117] ./~/regenerator-runtime/runtime.js 21.8 kB {2} [built] [] -> factory:73ms building:433ms dependencies:25ms = 531ms [118] ./~/requires-port/index.js 753 bytes {2} [built] [] -> factory:217ms building:26ms = 243ms [119] ./~/sockjs-client/lib/entry.js 244 bytes {2} [built] [] -> factory:2634ms building:7055ms dependencies:1ms = 9690ms [120] ./~/sockjs-client/lib/event/close.js 295 bytes {2} [built] [] -> factory:19ms building:356ms dependencies:21ms = 396ms [121] ./~/sockjs-client/lib/event/trans-message.js 292 bytes {2} [built] [] -> factory:19ms building:358ms dependencies:20ms = 397ms [122] ./~/sockjs-client/lib/facade.js 723 bytes {2} [built] [] -> factory:520ms building:87ms dependencies:0ms = 607ms [123] ./~/sockjs-client/lib/iframe-bootstrap.js 2.9 kB {2} [built] [] -> factory:15ms building:323ms dependencies:543ms = 881ms [124] ./~/sockjs-client/lib/info-iframe.js 1.52 kB {2} [built] [] -> factory:489ms building:178ms dependencies:1ms = 668ms [125] ./~/sockjs-client/lib/info-receiver.js 2.22 kB {2} [built] [] -> factory:12ms building:376ms dependencies:2ms = 390ms [126] ./~/sockjs-client/lib/main.js 11.9 kB {2} [built] [] -> factory:111ms building:3391ms dependencies:4ms = 3506ms [127] ./~/sockjs-client/lib/shims.js 18.2 kB {2} [built] [] -> factory:8ms building:386ms = 394ms [128] ./~/sockjs-client/lib/transport-list.js 613 bytes {2} [built] [] -> factory:110ms building:3405ms = 3515ms [129] ./~/sockjs-client/lib/transport/browser/websocket.js 172 bytes {2} [built] [] -> factory:83ms building:166ms dependencies:0ms = 249ms [130] ./~/sockjs-client/lib/transport/jsonp-polling.js 1.02 kB {2} [built] [] -> factory:268ms building:622ms dependencies:1ms = 891ms [131] ./~/sockjs-client/lib/transport/lib/buffered-sender.js 2.3 kB {2} [built] [] -> factory:26ms building:37ms dependencies:0ms = 63ms [132] ./~/sockjs-client/lib/transport/lib/polling.js 1.32 kB {2} [built] [] -> factory:26ms building:31ms dependencies:0ms = 57ms [133] ./~/sockjs-client/lib/transport/receiver/eventsource.js 1.58 kB {2} [built] [] -> factory:62ms building:161ms dependencies:15ms = 238ms [134] ./~/sockjs-client/lib/transport/receiver/htmlfile.js 2.2 kB {2} [built] [] -> factory:67ms building:155ms dependencies:22ms = 244ms [135] ./~/sockjs-client/lib/transport/receiver/jsonp.js 5.57 kB {2} [built] [] -> factory:37ms building:161ms dependencies:10ms = 208ms [136] ./~/sockjs-client/lib/transport/sender/jsonp.js 2.46 kB {2} [built] [] -> factory:37ms building:164ms dependencies:0ms = 201ms [137] ./~/sockjs-client/lib/transport/sender/xhr-fake.js 456 bytes {2} [built] [] -> factory:499ms building:184ms dependencies:16ms = 699ms [138] ./~/sockjs-client/lib/transport/websocket.js 2.71 kB {2} [built] [] -> factory:266ms building:584ms dependencies:21ms = 871ms [139] ./~/sockjs-client/lib/transport/xdr-polling.js 712 bytes {2} [built] [] -> factory:268ms building:597ms dependencies:56ms = 921ms [140] ./~/sockjs-client/lib/transport/xhr-streaming.js 1.25 kB {2} [built] [] -> factory:266ms building:591ms dependencies:63ms = 920ms [141] ./~/sockjs-client/lib/utils/browser-crypto.js 438 bytes {2} [built] [] -> factory:501ms building:81ms dependencies:0ms = 582ms [142] ./~/sockjs-client/lib/utils/escape.js 2.31 kB {2} [built] [] -> factory:16ms building:339ms dependencies:526ms = 881ms [143] ./~/sockjs-client/lib/utils/log.js 450 bytes {2} [built] [] -> factory:18ms building:355ms dependencies:1ms = 374ms [144] ./~/sockjs-client/lib/utils/transport.js 1.35 kB {2} [built] [] -> factory:17ms building:352ms dependencies:27ms = 396ms [145] ./~/strip-ansi/index.js 161 bytes {2} [built] [] -> factory:591ms building:414ms = 1005ms [146] ./~/url-parse/lolcation.js 1.58 kB {2} [built] [] -> factory:32ms building:181ms dependencies:1ms = 214ms [147] ./~/url/url.js 23.3 kB {2} [built] [] -> factory:525ms building:474ms = 999ms [148] ./~/url/util.js 314 bytes {2} [built] [] -> factory:2164ms building:6616ms = 8780ms [149] (webpack)-dev-server/client/socket.js 856 bytes {2} [built] [] -> factory:447ms building:113ms = 560ms [150] (webpack)/buildin/amd-options.js 43 bytes {2} [built] [] -> factory:27ms building:43ms = 70ms [151] (webpack)/hot/emitter.js 77 bytes {2} [built] [] -> factory:526ms building:380ms = 906ms [156] multi aurelia-bootstrap 100 bytes {2} [built] [16:59:16] factory:0ms building:0ms dependencies:151ms = 151ms [aurelia-bootstrapper-webpack] ./~/aurelia-bootstrapper-webpack/dist/commonjs/aurelia-bootstrapper-webpack.js 2.54 kB {2} [built] [] -> factory:188ms building:117ms dependencies:414ms = 719ms [aurelia-loader-webpack] ./~/aurelia-loader-webpack/dist/commonjs/aurelia-loader-webpack.js 7.04 kB {2} [built] [] -> factory:221ms building:278ms dependencies:219ms = 718ms [aurelia-pal-browser] ./~/aurelia-pal-browser/dist/commonjs/aurelia-pal-browser.js 15 kB {2} [built] [] -> factory:190ms building:84ms dependencies:269ms = 543ms [aurelia-polyfills] ./~/aurelia-polyfills/dist/commonjs/aurelia-polyfills.js 22.4 kB {2} [built] [] -> factory:189ms building:114ms dependencies:290ms = 593ms [bluebird] ./~/bluebird/js/browser/bluebird.js 172 bytes {2} [built] [] -> factory:702ms building:16ms dependencies:44ms = 762ms [16:59:16] Child html-webpack-plugin for "index.html": [16:59:16] chunk {0} index.html 529 kB [entry] [rendered] [0] ./~/lodash/lodash.js 527 kB {0} [built] [] -> factory:475ms building:436ms = 911ms [1] (webpack)/buildin/global.js 506 bytes {0} [built] [] -> factory:0ms building:17ms = 17ms [2] (webpack)/buildin/module.js 548 bytes {0} [built] [] -> factory:8ms building:2272ms = 2280ms [3] ./~/html-webpack-plugin/lib/loader.js!./index.html 1.16 kB {0} [built] [16:59:16] factory:49ms building:175ms = 224ms [16:59:16] Child extract-text-webpack-plugin: [16:59:16] chunk {0} extract-text-webpack-plugin-output-filename 4.67 kB [entry] [rendered] [0] ./~/css-loader/lib/css-base.js 1.51 kB {0} [built] [] -> factory:288ms building:553ms = 841ms [1] ./~/css-loader!./styles/styles.css 3.16 kB {0} [built] [16:59:16] factory:2ms building:3351ms = 3353ms [16:59:16] Child extract-text-webpack-plugin: [16:59:16] Asset Size Chunks Chunk Names [16:59:16] f4769f9bdb7466be65088239c12046d1.eot 20.1 kB [emitted] [16:59:16]<PHONE_NUMBER>7bd7575d6327160d64e760.svg 109 kB [emitted] [16:59:16] e18bbf611f2a2e43afc071aa2f4e1512.ttf 45.4 kB [emitted] [16:59:16] fa2772327f55d8198301fdb8bcfc8158.woff 23.4 kB [emitted] [16:59:16] 448c34a56d699c29117adc64c43affeb.woff2 18 kB [emitted] [16:59:16] chunk {0} extract-text-webpack-plugin-output-filename 156 kB [entry] [rendered] [0] ./~/bootstrap/dist/fonts/glyphicons-halflings-regular.eot 82 bytes {0} [built] [] -> factory:102ms building:171ms = 273ms [1] ./~/css-loader/lib/css-base.js 1.51 kB {0} [built] [] -> factory:0ms building:48ms = 48ms [2] ./~/bootstrap/dist/fonts/glyphicons-halflings-regular.svg 82 bytes {0} [built] [] -> factory:102ms building:171ms = 273ms [3] ./~/bootstrap/dist/fonts/glyphicons-halflings-regular.ttf 82 bytes {0} [built] [] -> factory:102ms building:171ms = 273ms [4] ./~/bootstrap/dist/fonts/glyphicons-halflings-regular.woff 83 bytes {0} [built] [] -> factory:197ms building:79ms = 276ms [5] ./~/bootstrap/dist/fonts/glyphicons-halflings-regular.woff2 84 bytes {0} [built] [] -> factory:197ms building:78ms = 275ms [6] ./~/css-loader!./~/bootstrap/dist/css/bootstrap.css 154 kB {0} [built] [16:59:16] factory:3297ms building:861ms = 4158ms [16:59:16] Child extract-text-webpack-plugin: [16:59:16] Asset Size Chunks Chunk Names [16:59:16] 25a32416abee198dd821b0b17a198a8f.eot 76.5 kB [emitted] [16:59:16] d7c639084f684d66a1bc66855d193ed8.svg 392 kB [emitted] [16:59:16] 1dc35d25e61d819a9c357074014867ab.ttf 153 kB [emitted] [16:59:16] c8ddf1e5e5bf3682bc7bebf30f394148.woff 90.4 kB [emitted] [16:59:16] e6cf7c6ec7c2d6f670ae9d762604cb0b.woff2 71.9 kB [emitted] [16:59:16] chunk {0} extract-text-webpack-plugin-output-filename 41.5 kB [entry] [rendered] [0] ./~/css-loader/lib/css-base.js 1.51 kB {0} [built] [] -> factory:66ms building:561ms = 627ms [1] ./~/font-awesome/fonts/fontawesome-webfont.eot 82 bytes {0} [built] [] -> factory:684ms building:174ms = 858ms [2] ./~/font-awesome/fonts/fontawesome-webfont.eot?v=4.6.3 82 bytes {0} [built] [] -> factory:679ms building:179ms = 858ms [3] ./~/font-awesome/fonts/fontawesome-webfont.svg?v=4.6.3 82 bytes {0} [built] [] -> factory:685ms building:205ms = 890ms [4] ./~/font-awesome/fonts/fontawesome-webfont.ttf?v=4.6.3 82 bytes {0} [built] [] -> factory:685ms building:173ms = 858ms [5] ./~/font-awesome/fonts/fontawesome-webfont.woff?v=4.6.3 83 bytes {0} [built] [] -> factory:780ms building:105ms = 885ms [6] ./~/font-awesome/fonts/fontawesome-webfont.woff2?v=4.6.3 84 bytes {0} [built] [] -> factory:687ms building:222ms = 909ms [7] ./~/css-loader!./~/font-awesome/css/font-awesome.css 39.5 kB {0} [built] [16:59:16] factory:3297ms building:277ms = 3574ms [16:59:16] webpack: bundle is now VALID. [default] Checking started in a separate process... [default] C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack\custom_typings\aurelia-protractor.d.ts:17:36 [16:59:18] Cannot find name 'Locator'. [default] C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack\node_modules\protractor\built\index.d.ts:6:9 [16:59:18] Export declaration conflicts with exported declaration of 'ProtractorBy' [default] C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack\node_modules\selenium-webdriver\node_modules\adm-zip\zipEntry.js:64:17 [16:59:18] Unreachable code detected. [default] C:\work\tests\Monterey-testing-09122016\skeleton-typescript-webpack\node_modules\selenium-webdriver\node_modules\adm-zip\zipEntry.js:110:21 [16:59:18] Unreachable code detected. [default] Checking finished with 4 errors https://github.com/aurelia/skeleton-navigation/issues/664 Seems like an issue with the skeleton itself
2025-04-01T04:34:48.218902
2024-08-18T23:10:37
2472188595
{ "authors": [ "milesj", "varshith257" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8796", "repo": "moonrepo/tools", "url": "https://github.com/moonrepo/tools/issues/9" }
gharchive/issue
Support asdf plugins through a proto WASM plugin There has been multiple requests now to support (piggy-back) off of asdf, since it has a ton of plugins and support for many tools. Unfortunately, building this into proto core simply isn't feasible without rewriting it from scratch, and at that point, it's really not worth it. The other issue is that asdf is Unix only, so that means no Windows support. However! I had a thought the other day, why not support every asdf plugin through a single proto WASM plugin, similar to how our TOML plugins work (they are WASM under the hood). For example, say you wanted to use the scala asdf plugin, it may look something like the following in proto: # .prototools scala = "latest" [plugins] scala = "https://path/to/asdf/plugin.wasm" Easy right? But what if the proto identifier doesn't match up with the asdf plugin name? Well, just override it with configuration. For example, using the nodejs asdf plugin. # .prototools node = "latest" [plugins] node = "https://path/to/asdf/plugin.wasm" [tools.node] asdf-plugin = "nodejs" We can also take this further by supporting the git repository directly. # .prototools [tools.node] asdf-repository = "https://github.com/asdf-vm/asdf-nodejs.git" Caveats With all this being said, I have no idea if this will actually work, or if the asdf plugin APIs will match up to the proto plugin APIs. It's definitely something that needs to be prototyped to ensure it's feasible. Requirements Plugin should be written in Rust and use our APIs/PDKs: https://moonrepo.dev/docs/guides/wasm-plugins https://moonrepo.dev/docs/proto/wasm-plugin= Does not need to support Windows Must work without asdf existing in the environment Should ideally support all asdf plugins Should detect versions from .tool-versions - https://asdf-vm.com/manage/configuration.html#tool-versions Should support ASDF_* env vars - https://asdf-vm.com/plugins/create.html#environment-variables-overview Implementation How this is implemented may require a lot of discussion, which can be had on Discord. With that said, we can outline a few things: How we download/install asdf plugins should probably mirror what asdf itself does. Stored at ~/.asdf - https://asdf-vm.com/manage/configuration.html#asdf-data-dir API reference - https://asdf-vm.com/plugins/create.html#required-scripts /bounty $500 @milesj Instead you have transferred the issue. Wit that . Almost done except installation changes. I am presuming my work of last done. With that /attempt #9
2025-04-01T04:34:48.226661
2018-08-22T22:38:30
353151769
{ "authors": [ "jiehuang001", "moonso" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8797", "repo": "moonso/genmod", "url": "https://github.com/moonso/genmod/issues/88" }
gharchive/issue
No "GeneticModels=" in the output Hi, there, Please see my test.vcf and test.ped file pasted at the bottom. Please use *sed 's/ /\t/g' in case the TAB are pasted into multiple spaces. After I run genmod models test.vcf --family_file test.ped, there is no GeneticModels= written to the output for some reason. Can you please kindly let me know what I missed or did incorrectly? Thank you & best regards, Jie ========== test.ped ========= #FamilyID SampleID Father Mother Sex Phenotype CEU ON180102131 0 0 1 1 CEU ON180102132 0 0 2 1 CEU ON180102133 ON180102131 ON180102132 1 2 =========== test.vcf ======== #fileformat=VCFv4.1 ##INFO=<ID=MQ,Number=1,Type=Float,Description="RMS Mapping Quality"> ##INFO=<ID=Annotation,Number=.,Type=String,Description="Annotates what feature(s) this variant belongs to."> ##INFO=<ID=Exonic,Number=0,Type=Flag,Description="Indicates if the variant is exonic."> ##contig=<ID=1,length=249250621,assembly=b37> ##reference=file:///humgen/gsa-hpprojects/GATK/bundle/current/b37/human_g1k_v37.fasta #CHROM POS ID REF ALT QUAL FILTER INFO FORMAT ON180102131 ON180102133 ON180102132 1 12908 . C A 67.13 PASS AC=1;AF=0.167;AN=6;BaseQRankSum=3.638;ClippingRankSum=0;DP=96;ExcessHet=3.0103;FS=0;MLEAC=1;MLEAF=0.167;MQ=26.68;MQRankSum=0.04;QD=2.24;ReadPosRankSum=-0.647;SOR=0.075;Annotation=DDX11L1,WASH7P GT:AD:DP:GQ:PL 0/0:33,0:33:99:0,99,995 0/0:33,0:33:67:0,67,955 0/1:24,6:30:98:98,0,651 1 13302 . C T 202.99 PASS AC=1;AF=0.167;AN=6;BaseQRankSum=4.727;ClippingRankSum=0;DP=116;ExcessHet=3.6798;FS=4.754;MLEAC=2;MLEAF=0.333;MQ=36.12;MQRankSum=-5.329;QD=4.61;ReadPosRankSum=-3.402;SOR=1.514;Annotation=DDX11L1,WASH7P GT:AD:DP:GQ:PL 0/1:31,13:44:99:229,0,589 0/0:42,0:42:2:0,2,855 0/0:30,0:30:0:0,0,587 Hi, I think now I found the reason for the error: I must put samples in the order of father-mother-proband in the VCF file; I could not have the "PL" field in the VCF file. Once I fixed the above two issues. Now it works fine. Please try this out and fix this bug. Thank you! Best regards, Jie Hi Jie, the reason why no models are annotated are that no models are followed in your example. The proband(affected child) has is ON180102133 according to the ped file. In your vcf it is the second column of the individuals. It has genotype 0/0 for both variants so no models are followed in this case.
2025-04-01T04:34:48.231931
2018-10-24T07:36:15
373344308
{ "authors": [ "moooofly" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8798", "repo": "moooofly/harbor-go-client", "url": "https://github.com/moooofly/harbor-go-client/issues/9" }
gharchive/issue
Rename CLI tool from harbor-go-client to harborctl As per https://github.com/goharbor/community/pull/6 , I will rename CLI tool from harbor-go-client to harborctl since v1.0.0 release. In order for multiple platforms using, harborctl is suffixed as follow: harborctl_darwin_amd64 harborctl_linux_amd64
2025-04-01T04:34:48.233753
2018-10-02T17:10:07
365994830
{ "authors": [ "adamdecaf", "ianbibby" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8799", "repo": "moov-io/ach", "url": "https://github.com/moov-io/ach/issues/280" }
gharchive/issue
Missing X-Total-Count headers on GET requests What were you trying to do? $ curl -I -XGET -H "cookie: ..." https://api.moov.io/v1/ach/files What did you expect to see? I expected to see a X-Total-Count header returned with the count of files available. What did you see? No such header was returned. Happy to pick this one up.
2025-04-01T04:34:48.235139
2022-09-14T04:13:20
1372302851
{ "authors": [ "adamdecaf", "matiasinsaurralde" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8800", "repo": "moov-io/fincen", "url": "https://github.com/moov-io/fincen/pull/15" }
gharchive/pull-request
Improve regexp usage by reusing compiled regular expression objects This is also thread-safe [0] [0] https://pkg.go.dev/regexp#Regexp Thank you @matiasinsaurralde!
2025-04-01T04:34:48.241242
2016-03-22T13:31:58
142650916
{ "authors": [ "coveralls", "snmgian" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8801", "repo": "moove-it/sidekiq-scheduler", "url": "https://github.com/moove-it/sidekiq-scheduler/pull/100" }
gharchive/pull-request
Add flag for pushing jobs on active queues only Issue: #99 Coverage decreased (-0.4%) to 93.986% when pulling ee28cabb87265f5f155269572d791a5e2ca58256 on snmgian:issues/99 into 0e0cc485ece127290618883fc06fa2bf4a160184 on moove-it:master. Coverage increased (+0.7%) to 95.129% when pulling 65e7ee4d9a7bc1bdfebd224d2a252a3b72819c8a on snmgian:issues/99 into 0e0cc485ece127290618883fc06fa2bf4a160184 on moove-it:master. Coverage increased (+0.7%) to 95.129% when pulling 8471401f7e5947e0608d2d882f2d3dc428c05092 on snmgian:issues/99 into 0e0cc485ece127290618883fc06fa2bf4a160184 on moove-it:master. Coverage increased (+0.6%) to 94.976% when pulling fc87e6358141b98bde51c74afe119abf441a740a on snmgian:issues/99 into 0e0cc485ece127290618883fc06fa2bf4a160184 on moove-it:master. Coverage increased (+0.6%) to 94.976% when pulling fc87e6358141b98bde51c74afe119abf441a740a on snmgian:issues/99 into 0e0cc485ece127290618883fc06fa2bf4a160184 on moove-it:master.
2025-04-01T04:34:48.310961
2015-10-10T00:55:43
110757854
{ "authors": [ "atifaziz", "sholland1" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8802", "repo": "morelinq/MoreLINQ", "url": "https://github.com/morelinq/MoreLINQ/pull/120" }
gharchive/pull-request
Implement AtLeast extension method #58 If this is good for this method, I can implement the Exactly method, too. I added the optimization for ICollections. Is there anything else I could do to wrap this up? You're right. This stack overflow answer corroborates this. @sholland1 Could you also remove the UTF-8 BOM marker from the following files? MoreLinq\AtLeast.cs MoreLinq.Test\AtLeastTest.cs I have finished implementing each of the improvements you suggested. I used the existing UnenumerableList class to improve the test. While I was looking through the test project, I found the InfiniteSequence class. I used it to refactor these tests a little bit. In the future, it'd probably be worth it to separate the tests and test helpers.
2025-04-01T04:34:48.341383
2022-03-24T20:18:08
1180010558
{ "authors": [ "lb-", "pankaj-tm" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8803", "repo": "moroshko/react-autosuggest", "url": "https://github.com/moroshko/react-autosuggest/issues/846" }
gharchive/issue
How to add autoComplete off Are you reporting a bug? Please create a Codepen that demonstrates your issue. You can start from the Basic example. Provide the steps to reproduce the issue, e.g.: Focus on the input field Type c, and wait for suggestions to appear Press Escape Observed behaviour: Suggestions stay open Expected behaviour: Suggestions should be closed Are you making a feature request? Please describe your use case from user journey point of view, e.g.: In my application, when user highlights suggestions (using the mouse or keyboard), I'd like to display additional information about the highlighted suggestion alongside the Autosuggest. If you have ideas how to extend the Autosuggest API to support your new feature, please share! If you know any examples online that already implement such functionality, please share a link. Pretty sure you can just pass autoComplete to the inputProps. <Autosuggest inputProps={{ autoComplete: 'donotsuggest' }} ...otherProps /> https://developer.mozilla.org/en-US/docs/Web/HTML/Attributes/autocomplete
2025-04-01T04:34:48.353160
2023-04-03T07:24:24
1651494739
{ "authors": [ "nargesr", "wangjiawen2013" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8804", "repo": "mortazavilab/PyWGCNA", "url": "https://github.com/mortazavilab/PyWGCNA/issues/25" }
gharchive/issue
high resolution PyWGCNA_overview.png Hi, Some letters on PyWGCNA_overview.png are too small to see clearly, would you mind providing higher resolution figures ? Hi, I upload the higher resolution here! Don't hesitate to reopen this issue if the quality is not good enough yet! Thanks for you reply! However, there are some flaws on the figure. such as 3) find modules, 4. the bottom of "identify gene moudles" was covered by the plot 4) study modules, 3. module GO term. The points are covered by the legend 5) compare pywgcnas, 1. comparison table. There are red wave lines under some letter.
2025-04-01T04:34:48.355998
2024-03-15T20:08:28
2189391830
{ "authors": [ "ShashankMosaicML", "samhavens", "vchiley" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8805", "repo": "mosaicml/llm-foundry", "url": "https://github.com/mosaicml/llm-foundry/pull/1039" }
gharchive/pull-request
[Draft] Get EOS token id from tokenizer Currently, we need to specify the eos token id in the dataset config for sequence id masking to work. It would be better to get it from the tokenizer (if the tokenizer has that info). This would prevent human errors where one changes the tokenizer but forgets to change the eos_token_id in the dataset config, or sets it incorrectly. Should we also pull out 'bos_token_id'? Should we also pull out 'bos_token_id'? ConcatenatedSequenceCollatorWrapper throws an error if we set both the eos_token_id and bos_token_id. That is why I only set eos_token_id here. I can add a flag that dictates which of these to automatically pick. But then, we will have to explicitly set one of those. @samhavens @codestar12 do you think eos_token_id or bos_token_id is more important and/or widely used? (probably eos but wanted to check) Llama and t5 models both expect both eos and bos, OPT I think is BOS only?
2025-04-01T04:34:48.358575
2022-05-20T15:22:01
1243303640
{ "authors": [ "seriousme" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8806", "repo": "moscajs/aedes-persistence-redis", "url": "https://github.com/moscajs/aedes-persistence-redis/issues/97" }
gharchive/issue
[suggestion] add stale issue handling to github actions I noticed that there quite some dated issues that are still open. My suggestion would be to create a workflow action that automatically closes dated issues after a certain number of days of inactivity. I use one myself based on githubs https://github.com/actions/stale action. E.g.: https://github.com/seriousme/fastify-openapi-glue/blob/master/.github/workflows/closeStale.yml An example of this flow in action can be seen in: https://github.com/seriousme/fastify-openapi-glue/issues/319 Hope this helps, Hans Ok
2025-04-01T04:34:48.361535
2019-12-20T17:19:49
541104024
{ "authors": [ "Digenis", "moshe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8807", "repo": "moshe/elasticsearch_loader", "url": "https://github.com/moshe/elasticsearch_loader/pull/80" }
gharchive/pull-request
Allow multiple values for es_hosts The elasticsearch client accepts not only strings but also dictionaries and lists for the hosts argument. I think elasticsearch_loader should also accept multiple es_hosts options and pass them to the client constructor as demonstrated in the patch. Are additional integration tests necessary? Since I added the multiple=True argument to click's option, even with the current tests, the client constructor receives "a list of a single string" instead of "a string". Hopefully, this will raise a TypeError if the API changes. Hi @Digenis, and thank you for your contribution🙌 There are end2end tests that running es instance and connecting to it. it should be enough Wow, I wasn't expecting such a fast merge. What about doc? click doesn't automatically document that an option can be specified multiple times. I see that released 0.6 but doc can still be added in a post-release 0.6.0.post1
2025-04-01T04:34:48.369117
2021-06-04T15:59:21
911647385
{ "authors": [ "m3ngi3", "moshfeu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8808", "repo": "moshfeu/vscode-compare-folders", "url": "https://github.com/moshfeu/vscode-compare-folders/issues/85" }
gharchive/issue
Visual Studio: One or both folders are not exist Must be something I am missing but just installed new computer and when comparing selected folders: As you see in screenshot, I have two folders selected but when selecting [Compare Folders] Compare selected folders in righ-mouse-click menu I only get notification: One or both folders are not exist Compare Folders v0.22.1 Visual Studio: 1.56.2 (user setup) 😱 Seems like it's a bug on Windows. Will deploy a fix asap! Oh men! I thought it could be newbie experience from my side 😅 Thank you and keep up the good work. Love this tool 💪🏼👍🏼 Thanks :) I'll let you know the status. I reverted the bad commit and published v0.22.2. Can you verify please? Oh yeah! Back to comparing folders again. Thank you so much 🙏 Released 👍
2025-04-01T04:34:48.374822
2021-07-14T15:52:21
944570661
{ "authors": [ "seeflood", "tianjipeng" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8809", "repo": "mosn/layotto", "url": "https://github.com/mosn/layotto/issues/138" }
gharchive/issue
Add unit tests for runtime/runtime.go and grpc/api.go What would you like to be added: Add unit tests for runtime/runtime.go and grpc/api.go. These two files are the core engine code of Layotto, responsible for component lifecycle management and grpc request processing respectively. chinese: 为runtime/runtime.go和grpc/api.go添加单元测试。这两个文件是layotto的核心引擎代码,各自负责组件生命周期管理(启动、注册、初始化)和grpc请求处理,完成此任务可以很好的了解Layotto架构 Why is this needed: Currently our test coverage is not high :( Assigned to @tianjipeng Thanks for your passion! OK @tianjipeng Hi, are you still working on it? I wiil commit a pr this weekend I wiil commit a pr this weekend Fine.If you find some code or module difficult to understand,just tell me and I will add comments and document for that module first.
2025-04-01T04:34:48.384274
2018-05-12T20:59:33
322542550
{ "authors": [ "itchyny", "orivej" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8810", "repo": "motemen/gore", "url": "https://github.com/motemen/gore/pull/96" }
gharchive/pull-request
Fix error format This fixes the error reported by Go 1.10 (that runs go vet with go test) on Travis: $ go test ./... # github.com/motemen/gore/gocode gocode/gocode_test.go:12: Error call has possible formatting directive %#v ok github.com/motemen/gore 5.837s FAIL github.com/motemen/gore/gocode [build failed] Could you merge this? Thank you.
2025-04-01T04:34:48.396296
2023-12-12T06:32:40
2037079589
{ "authors": [ "Pavankrishna98", "mounikakancharla1997" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8812", "repo": "mounikakancharla1997/gitdemoprojectmindq", "url": "https://github.com/mounikakancharla1997/gitdemoprojectmindq/pull/1" }
gharchive/pull-request
Pavan hi, i have added two classes robot and autoit and also updated pom.xml please verify and merge. kk tq
2025-04-01T04:34:48.454442
2024-07-16T14:30:57
2411322304
{ "authors": [ "0xmovses", "andygolay" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8813", "repo": "movementlabsxyz/movement", "url": "https://github.com/movementlabsxyz/movement/pull/209" }
gharchive/pull-request
Implement event logging traits for Aptos and Eth side Summary RFCs: Link to RFC, Link to RFC, or $\emptyset$. Categories: any of protocol-units, networks, scripts, util, cicd, or misc. Changelog Testing Outstanding issues @0xmovses this is not yet ready for review, right? @andygolay Correct. I've de-prioritised it for the time being while I get the contracts for movement -> eth bridge first, as we need that for the bridge service to function. Then will come back to this 👍
2025-04-01T04:34:48.456567
2023-10-08T20:41:55
1932060413
{ "authors": [ "FreemRL", "GabrielCTroia" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8814", "repo": "movesthatmatter/movex", "url": "https://github.com/movesthatmatter/movex/pull/63" }
gharchive/pull-request
removed applyAction() comment Fixes #61 Hey I removed the applyAction() comment in libs/movex/src/specs/MovexClientResourceObservable.spec.ts. Checks: [X] the commented code is removed [X] no other test is touched [X] all tests keep passing yarn test Thank you! Merging soon
2025-04-01T04:34:48.480309
2021-06-05T15:02:24
912301534
{ "authors": [ "acostalima", "satazor" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8815", "repo": "moxystudio/jest-config", "url": "https://github.com/moxystudio/jest-config/issues/39" }
gharchive/issue
Update to jest v27 Jest v27 was released a few days ago and we should upgrade it. I think we can't just bump the peer dependency because we need to upgrade babel jest etc. Let's wait a bit for the ecosystem to upgrade and then we upgrade too. //cc @acostalima @ruipneves @threequartersjohn ⚠️ About Jest v27 and React Testing Library: https://github.com/callstack/react-native-testing-library#custom-jest-preset
2025-04-01T04:34:48.681072
2020-11-10T14:26:55
739956139
{ "authors": [ "smarnach" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8816", "repo": "mozilla-services/remote-settings-lambdas", "url": "https://github.com/mozilla-services/remote-settings-lambdas/pull/1145" }
gharchive/pull-request
Use different Megaphone auth tokens for reading and updating. Megaphone auth tokens only give access to either read or write. It's impossible to do both with the same auth token. While we could probably change this on the Megaphone side if we wanted to, I decided it's easiest to change it in the lambda. r? @leplatrem I've addressed your comment, but I don't have the power to merge this PR. I don't even seem to have the power to trigger the Travis build. And we will need to release a new version as well, so I can deploy this. I'll leave this up to you as well.
2025-04-01T04:34:49.103212
2020-03-17T06:54:45
582795731
{ "authors": [ "Snuffleupagus", "a7vickey", "timvandermeij" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8817", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/11704" }
gharchive/issue
some annotations appear to have a rectangle entry that covers most of the page Attach (recommended) or Link to PDF file here: Configuration: Web browser and its version: Firefox 74 Operating system and its version: 8.1 PDF.js version:2.0 Is a browser extension: Steps to reproduce the problem: some annotations appear entire cover page. we cant able to access rest other annotation. Annotation pop layer overflow other annotation layer What is the expected behavior? Page 5 What went wrong? Page 6 we can access only one annotation. rest of other hide. sample.pdf Need solution if your already resolve this issue please provide source code. All annotation pop render wrong position. please check page 6. This is your one and only warning to stop spamming the bug tracking with duplicate issues! If you continue doing this, your account will most likely be blocked! Duplicate of #11286 your people not resolving the issue without resolving issue how is bug is closed Snuffleupagus Are You mad dont have seance @timvandermeij Can you please block that user, and also remove the latest round of spam that were posted to numerous issues; thank you! I have removed all comments and reported the user to GitHub.
2025-04-01T04:34:49.104484
2021-04-28T11:17:11
869831847
{ "authors": [ "rulet" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8818", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/13308" }
gharchive/issue
Remember previous position after close firefox How to make pdf.js remember previous position in pdf file when close and open firefox again? Maybe to edit some options in about:config? The thing is that when pdf.js is installed in chrome as extension and when it open pdf file it remember previouse position when las time that pdf file was opened. That function available by extension preferences in chrome browser. How to do that in firefox?
2025-04-01T04:34:49.109346
2023-07-11T22:14:11
1799830662
{ "authors": [ "Juanchoo055", "Snuffleupagus" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8819", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/16677" }
gharchive/issue
Annotations event listener Is there a way to add an event listener like "onchange" to the annotations when accessing them using let annotationData = await page.getAnnotations();? I have been attempting to do so, but haven't been successful. It's quite difficult to know exactly what you're asking about, but you might be able to achieve what you need by adding a onSetModified callback-function to the annotationStorage-instance since that's invoked every time form-data changes; see https://github.com/mozilla/pdf.js/blob/73d650af3e88037c03f89aecdadf4824ac99f2a7/src/display/annotation_storage.js#L35-L39 Unfortunately it's very difficult to help here since it seems that you've focused more on a particular solution, rather than describing your actual "problem" in sufficient detail. When opening an issue, please provide all of the information requested in https://github.com/mozilla/pdf.js/blob/master/.github/ISSUE_TEMPLATE.md including a clear and concise description of the problem. I am sorry yo didnt understand wath i am asking, i will say it different way, I am sorry for any confusion. Let me rephrase it for clarity. I have a PDF file that contains input fields. After some research, I discovered that I can access these input fields using the page.getAnnotations() function. This function provides an object that contains all the input fields, including their types and content. I have successfully implemented a feature that allows users to interact with the PDF by displaying the input annotations, enabling them to write on the fields, as you can see next code. `async setupAnnotationss(page, viewport) { let annotationData = await page.getAnnotations(); const pdf_canvas = document.getElementById('pdf-canvas'); // Canvas offset const canvas_offset = pdf_canvas.getBoundingClientRect(); // Canvas height const canvas_height = pdf_canvas.height; // Canvas width const canvas_width = pdf_canvas.width; // CSS for annotation layer const annotationLayer = document.getElementById('annotation-layer'); annotationLayer.style.left = canvas_offset.left + 'px'; annotationLayer.style.top = canvas_offset.top + 'px'; annotationLayer.style.height = canvas_height + 'px'; annotationLayer.style.width = canvas_width + 'px'; // //Renderizar las nuevas anotaciones pdfjsLib.AnnotationLayer.render({ viewport: viewport.clone({ dontFlip: true }), div: annotationLayer, annotations: annotationData, page: page, event: onclick, }); },` However, I'm currently working on adding an event listener to each annotation. The purpose of this event listener is to identify which values are being changed by the user. Furthermore, I aim to provide the users with additional options based on these changes, enhancing their experience with the PDF.
2025-04-01T04:34:49.130915
2024-08-05T10:30:15
2448203007
{ "authors": [ "Snuffleupagus", "calixteman", "paritosh-mahale", "rahulrajpal2911" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8820", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/18561" }
gharchive/issue
[Bug]: PDF.js not Performing Calculations on Interactive PDF Forms Attach (recommended) or Link to PDF file Lowes Measure sheet (2).pdf Web browser and its version Mozilla Operating system and its version Linux PDF.js version v4.5.136 Is the bug present in the latest PDF.js version? Yes Is a browser extension No Steps to reproduce the problem Open the attached PDF document in Adobe Reader. Enter values in the input fields. Observe that the total is automatically calculated using the hidden per-unit price field. Open the same PDF document using PDF.js. Enter values in the input fields. Notice that the total is not calculated. What is the expected behavior? PDF.js should perform the same calculations as Adobe Reader when values are entered into the input fields. What went wrong? PDF.js does not perform the calculations, and the total remains unchanged when values are entered into the input fields. Link to a viewer No response Additional context I am experiencing an issue with PDF.js when it comes to interactive PDF forms that involve calculations. I have a PDF document containing input fields for user data entry and hidden fields that perform calculations based on the input values. Specifically, when I open the document in Adobe Reader, entering values in the input fields automatically retrieves a per-unit price from a hidden field and calculates the total by multiplying the per-unit price by the quantity entered. However, this calculation is not happening when I open the same document using PDF.js. I am experiencing an issue with PDF.js when it comes to interactive PDF forms that involve calculations. I have a PDF document containing input fields for user data entry and hidden fields that perform calculations based on the input values. Specifically, when I open the document in Adobe Reader, entering values in the input fields automatically retrieves a per-unit price from a hidden field and calculates the total by multiplying the per-unit price by the quantity entered. However, this calculation is not happening when I open the same document using PDF.js. Expected result : actual result: Please place such content and screen-shots under the appropriate headings below, rather than dumping it at the top of the issue! I am experiencing an issue with PDF.js when it comes to interactive PDF forms that involve calculations. I have a PDF document containing input fields for user data entry and hidden fields that perform calculations based on the input values. Specifically, when I open the document in Adobe Reader, entering values in the input fields automatically retrieves a per-unit price from a hidden field and calculates the total by multiplying the per-unit price by the quantity entered. However, this calculation is not happening when I open the same document using PDF.js. Expected result : actual result: Please place such content and screen-shots under the appropriate headings below, rather than dumping it at the top of the issue! Hi @Snuffleupagus Updated as you mention, Thank you for highlighting this. If I enter a rate for Aluminum, then there's an error with either pdf.js or Acrobat, you should fix it and provide us a correct file. If possible you should provide us a basic test case instead of a big file (it makes the debugging more complex). there is not issue in Acrobat, we have given you a same file that we received and it's breaking in pdf.js only. in acrobat is working fine. Test case is : Whenever we add quantity, Rate get calculated with the help of hidden filed and whenever rate updates the total also get updated immediately. Please don't tell me there's no issue... https://github.com/user-attachments/assets/f954b476-c632-4826-816b-c6ac396e86a6 Please refer this recording, it's chrome browser and pdf.js Uploading screen-capture (1).webm… Please refer this recording, it's chrome browser and pdf.js screen-capture (1).webm It'd be nice to provide a clear STR instead of saying fuzzy things like "Enter values in the input fields.". Really providing us a basic file with only 2,3 fields in order to clearly show what the problem is would be a way more helpful than a big pdf and fuzzy instructions to reproduce the issue. See my answer in https://github.com/mozilla/pdf.js/issues/18561#issuecomment-2268840656, you answered me "there is not issue in Acrobat", I wrote something clear and if you test what I wrote in this comment in either Chrome, Acrobat or Firefox then it doesn't work correctly because of a wrong field name. So again, please make a good use of our time in filing clear bug reports. The embedded JS code used to calculate fields value is either using unexisting fields or has some invalid JS code !! so when we evaluate the C callback, there are some exceptions we don't catch when our competitors does. Hi @calixteman We understand that not all PDFs are causing trouble, but only this one. When we save it using any PDF viewer or editor, even some basic row-level multiplication is not happening. These PDFs are coming from a different source, and therefore, we are not allowed to edit them. We hope you understand our limitations. Also, with our limited expertise, we are not sure if this PDF is causing trouble. It isn't a problem to not have a basic test case if you aren't able to produce one, but the minimal thing you can do is to provide a clear STR, for example: on page 2, click in the field Qty for the line Aluminum and add 123 click on the field below Expected results: in the field Rate, we should have 12300 in the field Total (on the bottom of the page) we should have 12300 Actual results: In both fields, Rate and Total, we've 0. screenshots as you did. Hi @calixteman Please find below details: Steps to reproduce: Open PDF document and go to page no 2. In "customer labor" table ( middle table) add Qty as 10 for Aluminum and click on next row in Qty column. Expected behavior: Rate for Aluminum should get calculated automatically by multiplying Qty by 10, so rate should become 100. Also the Total ( In red, below table) should also be get calculated as 100 ( it's an total of rate) Screenshot of expected result (from Adobe Reader) Actual behavior: The calculation is not getting trigger to calculate the rate and total based on Qty provided. Screenshot of actual result (from pdf.js)
2025-04-01T04:34:49.134977
2017-05-09T19:22:14
227470627
{ "authors": [ "Snuffleupagus", "diracdeltas" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8821", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/8395" }
gharchive/issue
PDF renders as blank in Chrome extension viewer Link to PDF file (or attach file here): https://npgallery.nps.gov/pdfhost/docs/NHLS/Text/70000736.pdf Configuration: Web browser and its version: Chrome 58 Operating system and its version: Mac PDF.js version: 1.6.293 version of the chrome extension Is an extension: Yes Steps to reproduce the problem: Load the PDF It's blank What is the expected behavior? (add screenshot) Should not be blank; it displays properly in chrome's built-in PDF viewer What went wrong? (add screenshot) Link to a viewer (if hosted on a site other than mozilla.github.io/pdf.js or as Firefox/Chrome extension): Closing as duplicate of issue #2073.
2025-04-01T04:34:49.136524
2011-12-21T22:47:35
2630692
{ "authors": [ "notmasteryet", "saebekassebil" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8822", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/issues/968" }
gharchive/issue
Comment and Check Annotations - Re. PR#946 Updated, Linted, Better branched code, finally :) I've added two images, from SVG source. They're very simple, and I'm not sure that they are the one that we're going to use, but it's a start. Let me hear what you think. @pdfjsbot test Awesome patch. Thank you. (ref #935)
2025-04-01T04:34:49.138853
2018-11-06T11:40:33
377809247
{ "authors": [ "Snuffleupagus", "giridharkannan", "longFeiLi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8823", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/pull/10227" }
gharchive/pull-request
Update url-lib.js code optimization It is recommended to use es5 Or before adding get get toString() {} This is a completely unnecessary change, since that file in bundled into the built pdf.js/pdf.worker.js files and will thus be automatically translated into ES5 compatible code by Babel; see e.g. the output in the pdfjs-dist repo: https://github.com/mozilla/pdfjs-dist/blob/ab06013da9a92dd94ff06a7fadba0345554dec9b/build/pdf.js#L7687-L7689 Am using pdfjs-dist in Angular project and what you said is not happening there. If possible, could you change that one like to es5 compactable line?
2025-04-01T04:34:49.140730
2021-04-25T15:24:06
867038262
{ "authors": [ "Snuffleupagus", "timvandermeij" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8824", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/pull/13297" }
gharchive/pull-request
Add a note about minification to the webpack-example README (issue 13290) Since we really don't want to let a particular Webpack-mode dictate how we can/can't write code, let's add a note in the webpack-example README about minification instead. Fixes #13290 I'd also suggest that we amend the FAQ-entry, i.e. https://github.com/mozilla/pdf.js/wiki/Frequently-Asked-Questions#minified, with something like: If you are not using `gulp minified`, please note that you *must* configure the minifier to keep original class/function names intact; otherwise the build is not guaranteed to work correctly. Thanks! The FAQ is also updated now.
2025-04-01T04:34:49.142629
2024-02-19T12:59:56
2142321237
{ "authors": [ "Snuffleupagus" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8825", "repo": "mozilla/pdf.js", "url": "https://github.com/mozilla/pdf.js/pull/17694" }
gharchive/pull-request
Add better validation for the "PREFERENCE" kind AppOptions Given that the "PREFERENCE" kind is used e.g. to generate the preference-list for the Firefox PDF Viewer, those options need to be carefully validated. With this patch we'll now check this unconditionally in development mode, during testing, and when creating the preferences in the gulpfile. /botio-linux preview /botio test /botio unittest /botio-linux preview /botio test
2025-04-01T04:34:49.297867
2016-05-30T13:05:05
157499184
{ "authors": [ "JGDove99", "acabunoc", "aleimba", "blahah", "bmkramer", "brucellino", "nutsci" ], "license": "cc0-1.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8826", "repo": "mozillascience/global-sprint-2016", "url": "https://github.com/mozillascience/global-sprint-2016/issues/50" }
gharchive/issue
Tool to assess accessibility of a list of references [ Project Lead ] @JGDove99 [ GitHub Repo ] [ Track ] Looking to inspire someone to build such a tool [ Level ] Advanced [ Timezone ] Description Use-case: An author of a new scholarly article has assembled a list of references, possibly in Zotero or other bibliographic tool, possibly just in Word. This author is a supporter of Open Access and so wants to assess which of her/his referenced sources are going to be accessible to readers of her/his new article. Possible approach: I understand that Google Scholar does not have an API that would allow one to build a tool that uses Google Scholar to search for open versions of each of the articles in a list of references. And apparently Google Scholar will detect and then prevent attempts to use screen-scraping to do so. But perhaps one of the other search engines like Bing or Baidu or something I've not heard of could do this. Limiting the search to just articles with DOIs is not sufficient for this purpose, and CrossRef will usually just lead one to the "article of record" (often behind a paywall). For the purposes of this use-case, the author will want to know which of her/his referenced sources are accessible to an end-user free on the open web. Availability of any version of the article (so-called "green" OA) should be acceptible. Possible places the article might reside includes repositories (institutional or subject or government), departmental or personal websites, academia.edu, ResearchGate, etc. The author, armed with information about which referenced sources they are using are still behind paywalls may want to contact those authors to urge them to provide open versions of their articles so readers of the new article will have access to the references sources. Open Source code from the OA Button, or from http://dissem.in might be useful for some, but not all of this task. So might code from http://doai.io. Want to Contribute? Join us at the Global Sprint June 2-3. Leave a comment in this issue to let the project lead know you're interested in contributing during #mozsprint 2016! Note to the Project Lead Congrats, John! This is your official project listing for the Mozilla Science Global Sprint 2016. To confirm your submission, please complete the following: [x] Provide a GitHub repository for work and discussion on your project [x] Confirm in a comment that at least one person will be available to review and answer questions on this project from 9-5 in their timezone on both June 2 & 3. Please specify timezone. :clock1: Do you want to become a featured project? :tada: Here are some exercises that will help your project be more inviting to new contributors. We hope you'll try to complete some of these as you prepare for #mozsprint. If you complete all the exercises, your project will be eligible to be featured in our collection of open source science projects! Once you've finished this list, contact @acabunoc to submit your project for review. [x] README.md: [ ] LICENSE: [ ] CONTRIBUTING.md: [ ] Roadmap or an Issue Tracker (with tasks to complete): [ ] Data Reuse Plan (if applicable): [ ] Code of Conduct: [ ] Personas: Hey John! This sounds a lot like this project: https://github.com/mozillascience/global-sprint-2016/issues/13 Similar, but taking it to a much broader level. #13 assesses what percentages of a reference list represent articles available on PubMed Central. That's one very important repository but primarily focused on medical science subjects. #50 would work across all disciplines and will most likely involve use of a general search engine (probably not Google Scholar because GS has no API and prevents screen-scraping) to find out if an article is anywhere on the web (repositories, publisher cites, author's cites like academia.edu or researchgate, etc.) and needs to find "green" versions of the article if the article of record is not open. I like the idea. Maybe you could also approach OA publishers, a small OA icon in their reference lists (next to OA references) might be nice. But I don't know if that is feasible. first of all, general comment - I don't think a single thread issue is going to be the right medium for this discussion, which is tres interesting. Take that as you may :smiley_cat: Secondly, have you considered using the DataCite APIs to immplement this tool ? Resolving objects via DOI and then parsing the metadata (which should usually be entirely consumable) to extract cited objects via DOI should be a respectable and implementable baseline study. Andreas, I completely agree. LinkResolvers, which were the breakthru standard for ease-of-use access in the behind-the-fire-wall world of the 90s, have a feature which is an API which can pre-test whether or not clicking on the link is going to reveal full-text, on-line, access so the library-product can show a different icon if full-text is going to be available. [Saves enormous time when examining a reference list and you know you don't have time for any inter-library-loan requests.] What the user needs is a similar feature for links after a citation so that know which ones will resolve to free-full-text access even before clicking on it. I've just started having conversations with publishers about this, but first showing that it can be done would be a big help. -john On Tue, May 31, 2016 at 4:07 AM, Andreas Leimbach<EMAIL_ADDRESS>wrote: I like the idea. Maybe you could also approach OA publishers, a small OA icon in their reference lists (next to OA references) might be nice. But I don't know if that is feasible. — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/mozillascience/global-sprint-2016/issues/50#issuecomment-222619765, or mute the thread https://github.com/notifications/unsubscribe/ASsJWSdrjKbYdTqbpw5dk3FHWmwbN1K5ks5qG-xZgaJpZM4IpyBw . Not all citations have DOIs. Even those that do, DOIs are under the control of the publisher of the article-of-record, and so for most traditional journals they rarely provide a path to a shared version of the article. 79% of scholarly publishers acknowledge that the author may share a version of their article, but they are not [yet] making it easy for researchers to find those shared articles. This is a topic I've brought up with Cross Ref. -john On Tue, May 31, 2016 at 4:15 AM, Bruce Becker<EMAIL_ADDRESS>wrote: first of all, general comment - I don't think a single thread issue is going to be the right medium for this discussion, which is tres interesting. Take that as you may 😺 Secondly, have you considered using the DataCite APIs to immplement this tool ? Resolving objects via DOI and then parsing the metadata (which should usually be entirely consumable) to extract cited objects via DOI should be a respectable and implementable baseline study. — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/mozillascience/global-sprint-2016/issues/50#issuecomment-222621359, or mute the thread https://github.com/notifications/unsubscribe/ASsJWVO8uhHmNnpjAK9YJyvusOyXqFrWks5qG-4zgaJpZM4IpyBw . Very interesting project - I'm running into this at the moment in a side project I'm doing, wanting to check a list of doi's (in casu Sci-Hub downloads) for green OA availability. I won't be able to join this sprint project as I have my own to attend to :-) but will be following with interest. One additional source (in addition to the dissem.in api and the OA button already mentioned) might be the DOAI initiative built on BASE: http://doai.io/ You can scrape Google Scholar from a chrome extension- I do it for mine. Although if you load too frequently Google Scholar will still serve a captcha. https://chrome.google.com/webstore/detail/lazy-scholar/fpbdcofpbclblalghaepibbagkkgpkak It checks GScholar, PubMed, EuropePMC, and DOAI in search of free texts. Note: I've found that many of Google Scholar's "free texts" aren't always free, which is why I color the icon link yellow in the extension for lower confidence. I drop all the results in a database for each paper which could be queried by DOI, PMID, title, etc, but the number of papers is limited at this point. What the user needs is a similar feature for links after a citation so that know which ones will resolve to free-full-text access even before clicking on it.< This is what my extension does for PDFs, and you can be confident if full texts are free when linked on PubMed and EuropePMC based on html tags. bmkramer, DOAI is really neat. And does a lot of what's needed. I notice that it's run by CAPSH which also does DISSEM.IN, and is based on an academic search engine: [image: BASE Logo] https://www.base-search.net/ which I've not heard of before. Thanks for this suggestion. -john dove On Tue, May 31, 2016 at 8:05 AM, bmkramer<EMAIL_ADDRESS>wrote: Very interesting project - I'm running into this at the moment in a side project I'm doing, wanting to check a list of doi's (in casu Sci-Hub downloads) for green OA availability. I won't be able to join this sprint project as I have my own to attend to :-) but will be following with interest. One additional source (in addition to the dissem.in api and the OA button already mentioned) might be the DOAI initiative built on BASE: http://doai.io/ — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/mozillascience/global-sprint-2016/issues/50#issuecomment-222668967, or mute the thread https://github.com/notifications/unsubscribe/ASsJWcRaIqDxPgqY5WDXl5g6pZ_t8yStks5qHCQIgaJpZM4IpyBw . BASE (already mentioned) and CORE (don't think mentioned yet) will cover a lot of this ground. Richard, I think http://dissem.in uses CORE as part of its scoring of an author's published record. So that portion of testing accessibility could build on their code. For some use cases the search for "accessible" needs to throw a wide net. This means departmental websites and sites like ResearchGate and Academia.edu. -john Sent from my iPad. On Jun 1, 2016, at 4:47 PM, Richard Smith-Unna<EMAIL_ADDRESS>wrote: BASE (already mentioned) and CORE (don't think mentioned yet) will cover a lot of this ground. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or mute the thread. @JGDove99 do you have a chat room where we can talk about this project? I have a load of ideas to offload before turning in for the evening :) How about WhatsApp? I'm at +1-781-964-2325. -john Sent from my iPad. On Jun 2, 2016, at 1:34 PM, Richard Smith-Unna<EMAIL_ADDRESS>wrote: @JGDove99 do you have a chat room where we can talk about this project? I have a load of ideas to offload before turning in for the evening :) — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or mute the thread. @JGDove99 yes dissem.in does use CORE. I think what you're describing is very much within the scope of dissem.in actually, and also within the scope of science fair. I'll describe below how I would go about solving the problem... Briefly, I'd make a node.js module that, given some metadata about a scholarly document, tries to find all sources for that document and evaluate whether they are freely accessible. Also I would call it hypatia - because she was the last librarian at the library of Alexandria 📚. hypatia would be plugin based, so each possible source of information or fulltexts would be a plugin. For example, you could start with the following plugins: crossref doai.io The simplest case would work like this: First check the crossref API to get complete metadata about the document, if it's there. If the metadata is there, check if the article is open access according to the license field If it's open access, download the fulltext using the URL provided by crossref If it's not open access, now resolve the same identifier via the doai.io resolver. Compare the resolved URL via doai.io and the one given by crossref to see if they are the same. If not, you've got a free version via doai.io that wasn't in crossref. That will cover more papers than any other single approach I think. Then we could go on adding more plugins that would extend the reach of hypatia: arxiv figshare ssrn zenodo CORE BASE dissem.in microsoft academic graph free text web searches looking for author / departmental sites Then hypatia could be integrated into any node.js project, or any web service. All of this stuff fits well in the framework of tools we're developing at ContentMine, and at Science Fair, and what the OA button folks and Cottage Labs are working on. Does it sound like this is the sort of thing you were imagining? Richard, Yes, you're on the right track. I've already worked with the dissem.in team to take a citation list in marked up format to score them for accessibility. And they are helping SJ Klein and myself on a project at MIT. However, there are some philosophical objections which keeps them from handling accessibility for some of the use cases I have in mind. For example they do not consider my nephew's article to be open. He shares it on Academia.edu. Google Scholar finds that article perfectly well. I tried DOAI with one of his articles and it fails to find the Academia.edu version as well. I think we should arrange a Skype call in the next week or so, if you're able. It could be that there are ways ContentMine could help with this, or build on each other. For the Sprint, I am hoping there is someone who can explore using Bing, Baidu, or something else for the last of your bullet points. I won't be available tomorrow. Perhaps we can talk next week. Thanks for your suggestions already. -John Sent from my iPad. On Jun 2, 2016, at 1:55 PM, Richard Smith-Unna<EMAIL_ADDRESS>wrote: @JGDove99 yes dissem.in does use CORE. I think what you're describing is very much within the scope of dissem.in actually, and also within the scope of science fair. I'll describe below how I would go about solving the problem... Briefly, I'd make a node.js module that, given some metadata about a scholarly document, tries to find all sources for that document and evaluate whether they are freely accessible. Also I would call it hypatia - because she was the last librarian at the library of Alexandria 📚. hypatia would be plugin based, so each possible source of information or fulltexts would be a plugin. For example, you could start with the following plugins: crossref doai.io The simplest case would work like this: First check the crossref API to get complete metadata about the document, if it's there. If the metadata is there, check if the article is open access according to the license field If it's open access, download the fulltext using the URL provided by crossref If it's not open access, now resolve the same identifier via the doai.io resolver. Compare the resolved URL via doai.io and the one given by crossref to see if they are the same. If not, you've got a free version via doai.io that wasn't in crossref. That will cover more papers than any other single approach I think. Then we could go on adding more plugins that would extend the reach of hypatia: arxiv figshare ssrn zenodo CORE BASE dissem.in microsoft academic graph free text web searches looking for author / departmental sites Then hypatia could be integrated into any node.js project, or any web service. All of this stuff fits well in the framework of tools we're developing at ContentMine, and at Science Fair, and what the OA button folks and Cottage Labs are working on. Does it sound like this is the sort of thing you were imagining? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or mute the thread. ps I have to go and do parenting duty - will catch up again in the morning :) @JGDove99 I think I would agree with the dissem.in team that an article posted on Academia is not open unless it's under an open license. If it's just posted there without an open license, it technically possible to read it but it isn't open. I'm interested in the problem of providing pragmatic access where open access is not available, but it's the long tail of the problem and has much lower payoff... eventually it's much easier to just use SciHub (which I am carefully not endorsing). There are use cases where the world is a better place when a curious mind is not unnecessarily fettered from reading something that an author wants them to have free access to. Those use cases also present the opportunity to educate a scholar about better ways to share. My main scheme and a couple others easily imagined allows both. SciHub is only STEM. The article my son needs (a four page article published by Cambridge University Press in 1994 for which they want 35 Euros for [salary for humanities instructors at Matej Bel University in Slovakia where he teaches are less than 1,000 euros a month]) is not in SciHub. I think if we speak we'll find common ground, because I agree with you that either my nephew or academia.edu needs to change. -john PS: You might find this post interesting. It's about a lot more than SciHub. It has a good description of the use-case of an author assembling references for a paper to publish. And the Twitter trick using #icanhazPDF, is one that I've never heard of before. It successfully produced by email a copy of the Hungarian phonetics article for my son. <EMAIL_ADDRESS>Sent from my iPad. On Jun 2, 2016, at 2:29 PM, Richard Smith-Unna<EMAIL_ADDRESS>wrote: @JGDove99 I think I would agree with the dissem.in team that an article posted on Academia is not open unless it's under an open license. If it's just posted there without an open license, it technically possible to read it but it isn't open. I'm interested in the problem of providing pragmatic access where open access is not available, but it's the long tail of the problem and has much lower payoff... eventually it's much easier to just use SciHub (which I am carefully not endorsing). — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub, or mute the thread. @JGDove99 @blahah As a step towards my own use case (Sci-Hub data), but possibly also of interest here, I've now cobbled together a (crude) R-script that queries the Dissemin API for a list of DOIs and returns a table with information on both OA availability (in so far as determined by Dissemin) and publisher policies regarding sharing preprint, postprint and publisher versions (also from Dissemin, sourced from SherpaRomeo). Script and description here: [Dissemin_API_R] (https://github.com/bmkramer/scihub_netherlands/tree/Dissemin_API_R) And in reference to the discussion above, I have confirmed with the Dissemin/DOAI team that DOAI (not Disssemin) includes ResearchGate (but not Academia.net) in checking online availability: [Twitter thread] (https://twitter.com/MsPhelps/status/748831746297298946)
2025-04-01T04:34:49.354461
2014-12-18T16:40:08
52386851
{ "authors": [ "josePhoenix", "mperrin" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8827", "repo": "mperrin/poppy", "url": "https://github.com/mperrin/poppy/issues/30" }
gharchive/issue
Test user-specified weightings via source argument to calcPSF This would probably be an Instrument instance and an OpticalSystem instance set up to do the same thing and a check to make sure the PSFs are approximately equal. (I'm not sure if there is or should be a test of the OpticalSystem code path that combines the weighted PSFs...) Fixed via https://github.com/mperrin/poppy/pull/46
2025-04-01T04:34:49.357723
2016-11-16T04:09:21
189581354
{ "authors": [ "ERMS", "mpetroff" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8828", "repo": "mpetroff/pannellum", "url": "https://github.com/mpetroff/pannellum/issues/285" }
gharchive/issue
Canvas Bug? I have a panorama with hotspots and tour and it works perfectly. But when I added autoLoad and set it to true on firstScene. The output was this, Then I inspect element and look for its canvas, Its weird because the height and weight of the canvass is 0. But when I make some actions such as adjusting the panel, the values change. What could be the explanation for this? Thank you and sorry for my grammar. Are you using the latest version of Pannellum? Is there anything printed to the developer console? I am using the latest version sir. Only warning. I suspect there's something wrong with how you're initializing Pannellum, then. I can't reproduce this issue. Thank you :)
2025-04-01T04:34:49.365179
2017-12-19T17:27:04
283310304
{ "authors": [ "mpromonet", "roger-" ], "license": "unlicense", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8829", "repo": "mpromonet/v4l2rtspserver", "url": "https://github.com/mpromonet/v4l2rtspserver/issues/82" }
gharchive/issue
RTP over TCP? By default the SDP output seems to have m=video 0 RTP/AVP 96, which I believe means it uses UDP+TCP. Is there a way to enable TCP only (interleaved) output? Hi Roger, Reading https://tools.ietf.org/html/rfc2326, I guess RTP/AVP means UDP, to describe an RTP over TCP session it may be RTP/AVP/TCP. The default behaviour of live555 is to send RTP/AVP in DECRIBE answer and next the SETUP could ask for RTPoverTCP (or multicast session). This may be overidable but probably not easily, the message is formatted in OnDemandServerMediaSubsession::setSDPLinesFromRTPSink Maybe a simple way you could filter UDP with iptables Best Regards, Michel. You’re right, I remember now that the client has to initiate the TCP switch so everything probably fine.
2025-04-01T04:34:49.467063
2018-11-22T10:41:47
383476633
{ "authors": [ "hpvd", "voku", "xuhdev" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8830", "repo": "mrclay/minify", "url": "https://github.com/mrclay/minify/issues/643" }
gharchive/issue
HTML minimization: whitespace around <li> should not be trimmed The HTML minifier removes whitespace around <li> as a "blocked/undisplayed element": https://github.com/mrclay/minify/blob/258e495451c03adf57e1df81c2f0ef0c25b2f40d/lib/Minify/HTML.php#L136 However, <li> is commonly modified to be inline (e.g., "The Whitespace Issue" section here, tag clouds in WordPress themes Twenty Twelve and Twenty Seventeen). This often breaks the rendering of <li> elements. @xuhdev hi, can you please test the HtmlMin library for html compressing (https://github.com/voku/HtmlMin) or can you give me some example code (input + what you expected), thanks! I am currently thinking of a pull request, so that we can use "HtmlMin" directly via this library, but first I wanted to know if this has some benefits for the users. I am currently thinking of a pull request, so that we can use "HtmlMin" directly via this library, but first I wanted to know if this has some benefits for the users. +1
2025-04-01T04:34:49.506088
2013-08-04T01:38:33
17602788
{ "authors": [ "mreinstein" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8831", "repo": "mreinstein/wag", "url": "https://github.com/mreinstein/wag/issues/14" }
gharchive/issue
model but don't parse external resources it would be nice to have external stylesheets, scripts, etc modeled in the graph as external. It wouldn't be parsed, but it'd be linkable. Would need to add a new variable, maybe mode to indicate external or internal asset. supported via the re-write in 1.0
2025-04-01T04:34:49.529233
2016-03-31T19:24:44
144993923
{ "authors": [ "MarcoThePoro", "houseoftech", "mrhanlon" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8832", "repo": "mrhanlon/less-than-slash", "url": "https://github.com/mrhanlon/less-than-slash/issues/34" }
gharchive/issue
Wrong tag being closed after tags used When closing a tag after a php tag was used it adds a closing tag for the tag that was opened inside php tags. <div> <ul> <li></li> <?php if ($first == 1):?> <li></li> <?php endif;?> This would close an </li>. <div> <ul> <li></li> <?php echo "<a></a>";?> </a> Here it closed an </a>. This should be fixed by c8018e0, it just need to be published I published v0.13.0, which includes this fix.
2025-04-01T04:34:49.530284
2023-01-11T04:29:30
1528412288
{ "authors": [ "KillingSpark", "akhilles" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8833", "repo": "mrhooray/crc-rs", "url": "https://github.com/mrhooray/crc-rs/pull/76" }
gharchive/pull-request
Make Crc generic over implementation This allows for new implementations to be added without breaking API compatibility. @killingspark please take a look. Yep that works for me and it seems like a non-breaking change to me :+1:
2025-04-01T04:34:49.550738
2022-12-31T15:13:53
1515140376
{ "authors": [ "Fryuni", "mristin", "statuscake-machine" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8834", "repo": "mristin/opinionated-commit-message", "url": "https://github.com/mristin/opinionated-commit-message/issues/110" }
gharchive/issue
Merge commits are failing length check When a pull request is merged it will contain a subject message similar to the following: Merge pull request #117 from Organisaiton/name-of-the-branch-being-merged This is longer than the default 50 character limit. When the action runs on the main branch it fails. Hi @statuscake-machine Should we just ignore messages starting with Merge pull request #.* from .*? I think that makes sense as these are outside the control of the user. Alternatively perhaps there is a way to detect if the commit was a merge commit? I'm not sure exactly the range of values that come from github.context.eventName but maybe merge is one of them? Update on this. There is no way to detect for sure that a commit is a merge commit from just the event. It is possible by making some API calls. For each commit in the push event we'd need to: Make a request asking which PR introduced that commit Check if the commit in question is the same as the merge_commit_sha of the PR Check if the merge method was not a rebase merge, in which that value is just the last commit of the PR since there are no merge commits A more generic approach would be to have a parameter to ignore commits whose title matches the given patterns. The default patterns for such a filter could be for matching GitHub's default merge commit title. @Fryuni I think the pattern matching approach is far simpler and would generalize across Git repositories. It took me a while to get back to this, but PR is open now. Also, @mristin, I think you can send the command for the dependabot's PRs to re-evaluate and close themselves since everything was updated on a single PR
2025-04-01T04:34:49.864406
2016-05-18T16:07:07
155541301
{ "authors": [ "362228416", "mrniko" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8835", "repo": "mrniko/netty-socketio", "url": "https://github.com/mrniko/netty-socketio/issues/350" }
gharchive/issue
NamespaceClient too complex, help I started a Server, and then want to be in another Java program send the message to this Server, when I want to create a NamespaceClient object, beats me, write a demo? Hi! Here is the https://github.com/mrniko/netty-socketio-demo demo project. Try com.corundumstudio.socketio.demo.NamespaceChatLauncher
2025-04-01T04:34:49.874515
2024-03-19T19:03:05
2195764634
{ "authors": [ "frodriguez-hu", "mrousavy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8836", "repo": "mrousavy/react-native-vision-camera", "url": "https://github.com/mrousavy/react-native-vision-camera/issues/2668" }
gharchive/issue
Is it possible to use face-api? Question Hey!! I am also trying to use face-api with this library and this one: https://github.com/mrousavy/react-native-blob-jsi-helper In order to transform resized images buffers in blobs so In can send them to the face detection api from face api, is that possible? What I tried VisionCamera Version 3.9.1 Additional information [ ] I am using Expo [X] I have read the Troubleshooting Guide [X] I agree to follow this project's Code of Conduct [X] I searched for similar questions in the issues page as well as in the discussions page and found none. I realized that I do not need firebase cause MLKit is free. But my other question is remaining, is it possible to use face-api? Yes you can build a native Frame Processor plugin to use any native Obj-C/Swift/Java/Kotlin library.
2025-04-01T04:34:49.881956
2024-05-31T10:14:36
2327475916
{ "authors": [ "ibrahim-developes", "mrousavy", "nolife08021" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8837", "repo": "mrousavy/react-native-vision-camera", "url": "https://github.com/mrousavy/react-native-vision-camera/issues/2927" }
gharchive/issue
Build failed with face-detector plugin🔧 How were you trying to build the app? i tried to build an face detect app, but facing error after install "vision-camera-face-detector". I had done install react-native-vision-camera and react-native-worklets-core. try install "vision-camera-trustee-face-detector-v3" also facing same issue. Full build logs /Users/wlm/rn/face-app/node_modules/vision-camera-face-detector/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPlugin.java:29: error: package com.mrousavy.camera.frameprocessor does not exist import com.mrousavy.camera.frameprocessor.FrameProcessorPlugin; ^ /Users/wlm/rn/face-app/node_modules/vision-camera-face-detector/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPlugin.java:35: error: cannot find symbol public class VisionCameraFaceDetectorPlugin extends FrameProcessorPlugin { ^ symbol: class FrameProcessorPlugin /Users/wlm/rn/face-app/node_modules/vision-camera-face-detector/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPluginPackage.java:10: error: package com.mrousavy.camera.frameprocessor does not exist import com.mrousavy.camera.frameprocessor.FrameProcessorPlugin; ^ /Users/wlm/rn/face-app/node_modules/vision-camera-face-detector/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPlugin.java:134: error: method does not override or implement a method from a supertype @Override ^ /Users/wlm/rn/face-app/node_modules/vision-camera-face-detector/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPlugin.java:141: error: cannot access Detector Task<List<Face>> task = faceDetector.process(image); ^ class file for com.google.mlkit.vision.common.internal.Detector not found /Users/wlm/rn/face-app/node_modules/vision-camera-face-detector/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPluginPackage.java:19: error: cannot find symbol FrameProcessorPlugin.register(new VisionCameraFaceDetectorPlugin()); ^ symbol: variable FrameProcessorPlugin location: class VisionCameraFaceDetectorPluginPackage 6 errors Project dependencies "dependencies": { "react": "18.2.0", "react-native": "0.71.17", "react-native-vision-camera": "^4.0.5", "react-native-worklets-core": "^1.3.3", "vision-camera-face-detector": "^0.1.8" }, VisionCamera Version 4.0.5 Target platforms Android Operating system MacOS Can you build the VisionCamera Example app? I didn't try (⚠️ your issue might get ignored & closed if you don't try this) Additional information [ ] I am using Expo [X] I have enabled Frame Processors (react-native-worklets-core) [X] I have read the Troubleshooting Guide [X] I agree to follow this project's Code of Conduct [X] I searched for similar issues in this repository and found none. this is related to the plugin. the plugin needs to update for latest V4. it's just a small change. also, use titles for issues. without titles your issues just get closed. hi, sorry for the empty title, I tried downgraded vision-camera to v3.7.1 but still issue. /Users/wlm/rn/face-app/node_modules/vision-camera-trustee-face-detector-v3/android/src/main/java/com/visioncamerafacedetector/VisionCameraFaceDetectorPlugin.java:12: error: package com.mrousavy.camera.frameprocessors does not exist import com.mrousavy.camera.frameprocessors.Frame; ^ Do you have any documentation about how to use V4 to do for face detection/ recognition? @nolife08021 did it work? @ibrahim-developes nope, for android i use back react-native-camera, but having linking issue on ios.
2025-04-01T04:34:49.921652
2022-01-10T11:46:21
1097801217
{ "authors": [ "Angelk90", "NgocLe1001", "ShaheerArshad-nearpeer", "Stukz", "mikamations", "mrousavy", "nimran" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8838", "repo": "mrousavy/react-native-vision-camera", "url": "https://github.com/mrousavy/react-native-vision-camera/issues/727" }
gharchive/issue
🐛 The given Camera device could not be found for use-case binding! What were you trying to do? I clone the project: https://github.com/mrousavy/react-native-vision-camera I tried to run the example file, after many problems, I was able to run it on an android physical device. Reproduceable Code No response What happened instead? Main camera appears to be out of order, screen all black. The secondary camera (the one for video calls for example) seems to be working. When I try to take a photo with the secondary camera, I get this: You don't even see the HDR icon. Everything seems to go very slow, opening the app. The photo does not take it, but the video does it record. @mrousavy : What am I doing wrong? Relevant log output No response Device motorola one (xt1941-4) VisionCamera Version 2.11.2 Additional information [ ] I am using Expo [X] I have read the Troubleshooting Guide [X] I agree to follow this project's Code of Conduct [X] I searched for similar issues in this repository and found none. @mrousavy : Tablet Samsung Tab A(2016): It does not work of the two cameras. Motorola one fusion+: Main camera does not work. This is indeed very weird, possibly a CameraX bug. Can you share the adb logs? @mrousavy : It doesn't work with any of the physical android devices I've tried. 01-10 16:19:05.292 3628 3686 I ReactNativeJS: Re-rendering camera page with active camera. Device: "back (0)" (3840x2160 @ 60fps) 01-10 16:19:05.340 3628 3628 I CameraView: Configuring session... 01-10 16:19:05.341 3628 3628 I CameraView: Configuring session with Camera ID 0 and custom format... 01-10 16:19:05.350 3628 3628 I CameraView: Using custom format - photo: 3840x2160, video: 3840x2160 @ 60 FPS 01-10 16:19:05.351 3628 3628 I CameraView: Setting AE_TARGET_FPS_RANGE to 60-60, and SENSOR_FRAME_DURATION to 0 01-10 16:19:05.352 3628 3628 D Recorder: Video source has transitioned to state: INACTIVE 01-10 16:19:05.355 3628 3628 D DeferrableSurface: surface closed, useCount=1 closed=true androidx.camera.core.SurfaceRequest$2@51e30e1 01-10 16:19:05.355 3628 3628 D DeferrableSurface: surface closed, useCount=1 closed=true androidx.camera.core.SurfaceRequest$2@c1c8492 01-10 16:19:05.355 3628 3628 D DeferrableSurface: surface closed, useCount=1 closed=true androidx.camera.core.impl.ImmediateSurface@b913dbf 01-10 16:19:05.356 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Use cases [androidx.camera.core.Preview-b353fc59-4218-4dfb-8e33-97d4ddb52195142067502, androidx.camera.video.VideoCapture-ae066d4d-c72d-4c51-8e7c-6a4be45ec6d688343247, androidx.camera.core.ImageAnalysis-3cd47eb4-249d-483e-9c11-772505a2ce2232142428] now DETACHED for camera 01-10 16:19:05.356 3628 3628 I CameraView: Adding VideoCapture use-case... 01-10 16:19:05.356 3628 3756 D UseCaseAttachState: All use case: [] for camera: 1 01-10 16:19:05.357 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Resetting Capture Session 01-10 16:19:05.361 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Releasing session in state OPENED 01-10 16:19:05.361 3628 3756 D UseCaseAttachState: Active and attached use case: [] for camera: 1 01-10 16:19:05.363 3628 3628 I CameraView: Tried to add photo use-case (`photo={true}`) but the Camera device only supports a single use-case at a time. Falling back to Snapshot capture. 01-10 16:19:05.363 3628 3628 I CameraView: Adding ImageAnalysis use-case... 01-10 16:19:05.364 3628 3756 D UseCaseAttachState: Active and attached use case: [] for camera: 1 01-10 16:19:05.364 3628 3628 I CameraView: Attaching 2 use-cases... 01-10 16:19:05.365 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Closing camera. 01-10 16:19:05.366 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Transitioning camera internal state: OPENED --> CLOSING 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: Recalculating open cameras: 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: Camera State 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: ------------------------------------------------------------------- 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: Camera@2acd80c[id=1] CLOSING 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: Camera@39bde0[id=0] UNKNOWN 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: ------------------------------------------------------------------- 01-10 16:19:05.367 3628 3756 D CameraStateRegistry: Open count: 1 (Max allowed: 1) 01-10 16:19:05.368 3628 3756 D CameraStateMachine: New public camera state CameraState{type=CLOSING, error=null} from CLOSING and null 01-10 16:19:05.368 3628 3756 D CameraStateMachine: Publishing new public camera state CameraState{type=CLOSING, error=null} 01-10 16:19:05.369 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Resetting Capture Session 01-10 16:19:05.371 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Releasing session in state CLOSING 01-10 16:19:05.373 3628 3756 D CaptureSession: onSessionFinished() 01-10 16:19:05.374 3222 3222 I CameraService: Camera : Shutting down 3628 score 0 01-10 16:19:05.375 3222 3222 W libc : Unable to set property "camera.mot.is.client.foreground" to "1": error code: 0x18 01-10 16:19:05.376 3460 12630 I QCameraHalWatchdog: Starting Watchdog Thread... 01-10 16:19:05.378 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=8, fileFormat=2, videoCodec=2, videoBitRate=42000000, videoFrameRate=30, videoFrameWidth=3840, videoFrameHeight=2160, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.378 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=6, fileFormat=2, videoCodec=2, videoBitRate=17000000, videoFrameRate=30, videoFrameWidth=1920, videoFrameHeight=1080, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.379 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=5, fileFormat=2, videoCodec=2, videoBitRate=10000000, videoFrameRate=30, videoFrameWidth=1280, videoFrameHeight=720, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.379 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=4, fileFormat=2, videoCodec=2, videoBitRate=6000000, videoFrameRate=30, videoFrameWidth=720, videoFrameHeight=480, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.380 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=8, fileFormat=2, videoCodec=2, videoBitRate=42000000, videoFrameRate=30, videoFrameWidth=3840, videoFrameHeight=2160, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.381 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=6, fileFormat=2, videoCodec=2, videoBitRate=17000000, videoFrameRate=30, videoFrameWidth=1920, videoFrameHeight=1080, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.381 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=5, fileFormat=2, videoCodec=2, videoBitRate=10000000, videoFrameRate=30, videoFrameWidth=1280, videoFrameHeight=720, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.382 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=4, fileFormat=2, videoCodec=2, videoBitRate=6000000, videoFrameRate=30, videoFrameWidth=720, videoFrameHeight=480, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.382 3628 3628 D QualitySelector: supportedQualities = [ConstantQuality{value=8, name=UHD}, ConstantQuality{value=6, name=FHD}, ConstantQuality{value=5, name=HD}, ConstantQuality{value=4, name=SD}] 01-10 16:19:05.382 3628 3628 D QualitySelector: Select quality by fallbackStrategy = RuleStrategy{fallbackQuality=ConstantQuality{value=-1, name=NONE}, fallbackRule=0} 01-10 16:19:05.382 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30004 stream type=4 01-10 16:19:05.382 3628 3628 D VideoCapture: Found selectedQualities [ConstantQuality{value=8, name=UHD}] by QualitySelector{preferredQualities=[ConstantQuality{value=8, name=UHD}], fallbackStrategy=RuleStrategy{fallbackQuality=ConstantQuality{value=-1, name=NONE}, fallbackRule=0}} 01-10 16:19:05.382 3211 4039 I mm-camera: <CPP >< INFO> 4173: cpp_module_handle_streamoff_event: info: doing stream-off for identity 0x30004 01-10 16:19:05.383 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=8, fileFormat=2, videoCodec=2, videoBitRate=42000000, videoFrameRate=30, videoFrameWidth=3840, videoFrameHeight=2160, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.383 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=6, fileFormat=2, videoCodec=2, videoBitRate=17000000, videoFrameRate=30, videoFrameWidth=1920, videoFrameHeight=1080, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.384 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=5, fileFormat=2, videoCodec=2, videoBitRate=10000000, videoFrameRate=30, videoFrameWidth=1280, videoFrameHeight=720, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.384 3628 3628 D VideoCapabilities: profile = CamcorderProfileProxy{duration=30, quality=4, fileFormat=2, videoCodec=2, videoBitRate=6000000, videoFrameRate=30, videoFrameWidth=720, videoFrameHeight=480, audioCodec=3, audioBitRate=320000, audioSampleRate=48000, audioChannels=2} 01-10 16:19:05.384 3628 3628 D VideoCapture: Set supported resolutions = [3840x2160] 01-10 16:19:05.389 3628 3628 D CameraOrientationUtil: getRelativeImageRotation: destRotationDegrees=0, sourceRotationDegrees=90, isOppositeFacing=true, result=90 01-10 16:19:05.390 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 266670000 BW avg 361112256 BW inst 361112256 01-10 16:19:05.390 3211 4039 I mm-camera: <CPP >< INFO> 4358: cpp_module_handle_streamoff_event: info: stream-off done for identity 0x30004 01-10 16:19:05.391 3628 3628 D CameraOrientationUtil: getRelativeImageRotation: destRotationDegrees=0, sourceRotationDegrees=90, isOppositeFacing=true, result=90 01-10 16:19:05.393 3628 3628 D CameraOrientationUtil: getRelativeImageRotation: destRotationDegrees=0, sourceRotationDegrees=90, isOppositeFacing=true, result=90 01-10 16:19:05.394 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30004, stream type=4 01-10 16:19:05.394 3628 3628 D CameraOrientationUtil: getRelativeImageRotation: destRotationDegrees=0, sourceRotationDegrees=90, isOppositeFacing=true, result=90 01-10 16:19:05.395 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30002 stream type=1 01-10 16:19:05.395 3211 4039 I mm-camera: <CPP >< INFO> 4173: cpp_module_handle_streamoff_event: info: doing stream-off for identity 0x30002 01-10 16:19:05.395 3211 4039 I mm-camera: <CPP >< INFO> 4358: cpp_module_handle_streamoff_event: info: stream-off done for identity 0x30002 01-10 16:19:05.396 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30002, stream type=1 01-10 16:19:05.396 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 266670000 BW avg 298904256 BW inst 298904256 01-10 16:19:05.397 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30003 stream type=5 01-10 16:19:05.398 3211 4039 I mm-camera: <CPP >< INFO> 4173: cpp_module_handle_streamoff_event: info: doing stream-off for identity 0x30003 01-10 16:19:05.399 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 100000000 BW avg 0 BW inst 0 01-10 16:19:05.399 3211 4039 I mm-camera: <CPP >< INFO> 4358: cpp_module_handle_streamoff_event: info: stream-off done for identity 0x30003 01-10 16:19:05.403 3211 4039 I mm-camera: <MCT >< INFO> 223: stop_sof_check_thread: Stopping SOF timeout thread session =3 01-10 16:19:05.403 3211 4079 I mm-camera: <MCT >< INFO> 149: mct_bus_sof_thread_run: Closing SOF tracker thread 01-10 16:19:05.404 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30003, stream type=5 01-10 16:19:05.405 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30001 stream type=7 01-10 16:19:05.405 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30001, stream type=7 01-10 16:19:05.423 3628 3628 E CameraView: Failed to configure session: No supported surface combination is found for camera device - Id : 0 and Hardware level: 3. May be the specified resolution is too large and not supported. Existing surfaces: [] New configs: [androidx.camera.core.impl.PreviewConfig@2b14f04, androidx.camera.video.impl.VideoCaptureConfig@caa0ced, androidx.camera.core.impl.ImageAnalysisConfig@a5e722] 01-10 16:19:05.423 3628 3628 E CameraView: update() threw: [device/invalid-device] The given Camera device could not be found for use-case binding! 01-10 16:19:05.423 3628 3628 E CameraView: invokeOnError(...): 01-10 16:19:05.424 3628 3628 W System.err: com.mrousavy.camera.InvalidCameraDeviceError: [device/invalid-device] The given Camera device could not be found for use-case binding! 01-10 16:19:05.424 3628 3628 W System.err: at com.mrousavy.camera.CameraView.configureSession(CameraView.kt:517) 01-10 16:19:05.424 3628 3628 W System.err: at com.mrousavy.camera.CameraView.access$configureSession(CameraView.kt:68) 01-10 16:19:05.424 3628 3628 W System.err: at com.mrousavy.camera.CameraView$update$1$1.invokeSuspend(CameraView.kt:325) 01-10 16:19:05.424 3628 3628 W System.err: at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) 01-10 16:19:05.424 3628 3628 W System.err: at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) 01-10 16:19:05.424 3628 3628 W System.err: at android.os.Handler.handleCallback(Handler.java:883) 01-10 16:19:05.424 3628 3628 W System.err: at android.os.Handler.dispatchMessage(Handler.java:100) 01-10 16:19:05.424 3628 3628 W System.err: at android.os.Looper.loop(Looper.java:241) 01-10 16:19:05.424 3628 3628 W System.err: at android.app.ActivityThread.main(ActivityThread.java:7617) 01-10 16:19:05.424 3628 3628 W System.err: at java.lang.reflect.Method.invoke(Native Method) 01-10 16:19:05.424 3628 3628 W System.err: at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492) 01-10 16:19:05.424 3628 3628 W System.err: at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:941) 01-10 16:19:05.425 3628 3628 W System.err: Caused by: java.lang.IllegalArgumentException: No supported surface combination is found for camera device - Id : 0 and Hardware level: 3. May be the specified resolution is too large and not supported. Existing surfaces: [] New configs: [androidx.camera.core.impl.PreviewConfig@2b14f04, androidx.camera.video.impl.VideoCaptureConfig@caa0ced, androidx.camera.core.impl.ImageAnalysisConfig@a5e722] 01-10 16:19:05.425 3628 3628 W System.err: at androidx.camera.lifecycle.LifecycleCameraRepository.bindToLifecycleCamera(LifecycleCameraRepository.java:280) 01-10 16:19:05.425 3628 3628 W System.err: at androidx.camera.lifecycle.ProcessCameraProvider.bindToLifecycle(ProcessCameraProvider.java:532) 01-10 16:19:05.425 3628 3628 W System.err: at androidx.camera.lifecycle.ProcessCameraProvider.bindToLifecycle(ProcessCameraProvider.java:360) 01-10 16:19:05.425 3628 3628 W System.err: at com.mrousavy.camera.CameraView.configureSession(CameraView.kt:500) 01-10 16:19:05.425 3628 3628 W System.err: ... 11 more 01-10 16:19:05.430 3628 3686 E ReactNativeJS: { [device/invalid-device: [device/invalid-device] The given Camera device could not be found for use-case binding!] 01-10 16:19:05.430 3628 3686 E ReactNativeJS: name: 'device/invalid-device', 01-10 16:19:05.430 3628 3686 E ReactNativeJS: _code: 'device/invalid-device', 01-10 16:19:05.430 3628 3686 E ReactNativeJS: _message: '[device/invalid-device] The given Camera device could not be found for use-case binding!', 01-10 16:19:05.430 3628 3686 E ReactNativeJS: _cause: 01-10 16:19:05.430 3628 3686 E ReactNativeJS: { stacktrace: 'java.lang.IllegalArgumentException: No supported surface combination is found for camera device - Id : 0 and Hardware level: 3. May be the specified resolution is too large and not supported. Existing surfaces: [] New configs: [androidx.camera.core.impl.PreviewConfig@2b14f04, androidx.camera.video.impl.VideoCaptureConfig@caa0ced, androidx.camera.core.impl.ImageAnalysisConfig@a5e722]\n\tat androidx.camera.lifecycle.LifecycleCameraRepository.bindToLifecycleCamera(LifecycleCameraRepository.java:280)\n\tat androidx.camera.lifecycle.ProcessCameraProvider.bindToLifecycle(ProcessCameraProvider.java:532)\n\tat androidx.camera.lifecycle.ProcessCameraProvider.bindToLifecycle(ProcessCameraProvider.java:360)\n\tat com.mrousavy.camera.CameraView.configureSession(CameraView.kt:500)\n\tat com.mrousavy.camera.CameraView.access$configureSession(CameraView.kt:68)\n\tat com.mrousavy.camera.CameraView$update$1$1.invokeSuspend(CameraView.kt:325)\n\tat kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33)\n\tat kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106)\n\tat android.os.Handler.handleCallback(Handler.java:883)\n\tat android.os.Handler.dispatchMessage(Handler.java:100)\n\tat android.os.Looper.loop(Looper.java:241)\n\tat android.app.ActivityThread.main(ActivityThread.java:7617)\n\tat java.lang.reflect.Method.invoke(Native Method)\n\tat com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492)\n\tat com.android.internal.os.ZygoteInit.main(ZygoteInit.java:941)\n', 01-10 16:19:05.430 3628 3686 E ReactNativeJS: message: 'No supported surface combination is found for camera device - Id : 0 and Hardware level: 3. May be the specified resolution is too large and not supported. Existing surfaces: [] New configs: [androidx.camera.core.impl.PreviewConfig@2b14f04, androidx.camera.video.impl.VideoCaptureConfig@caa0ced, androidx.camera.core.impl.ImageAnalysisConfig@a5e722]' } } 01-10 16:19:05.433 3628 3628 D StreamStateObserver: Update Preview stream state to IDLE 01-10 16:19:05.466 3628 3686 I ReactNativeJS: Unable to symbolicate stack trace: Network request failed 01-10 16:19:05.467 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-ON on stream 0x30001 stream type=7 01-10 16:19:05.467 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-ON success for stream 0x30001, stream type=7 01-10 16:19:05.469 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-ON on stream 0x30004 stream type=4 01-10 16:19:05.469 3211 4039 I mm-camera: <SENSOR>< INFO> 4263: sensor_get_resolution_info: sensor info: name: s5k4h7, res: 1, max_fps: 120.000000, w: 640, h: 480 op pix clk: 280000000, FLL: 632, LLPCK: 3688, mode: 2, PDAF support: 0 01-10 16:19:05.469 3211 4039 I mm-3a-core: moto_led_calibration_init:502 Enter: cam_id:1 status:0 01-10 16:19:05.469 3211 4039 I mm-3a-core: moto_led_calibration_get_mod_serial_num: sn=0 01-10 16:19:05.469 3211 4039 E mm-3a-core: moto_led_calibration_read_write_led_cal:359 open /persist/camera/ledcal/front failed (No such file or directory) 01-10 16:19:05.469 3211 4039 E mm-3a-core: moto_led_calibration_apply No valid front chromatix pointer,LED calibration not applied, camera id 1 01-10 16:19:05.469 3211 4039 E mm-camera: <STATS_AF ><ERROR> 3393: af_port_handle_sensor_update: Depth_Err: actuator sensitivity: 0.000000 is invalid 01-10 16:19:05.470 3211 4039 I mm-camera: <CPP >< INFO> 1849: cpp_module_handle_stream_cfg_event: frame_offset=3, input_fps=60.00, identity=0x3000f 01-10 16:19:05.470 3211 4039 I mm-camera: <CPP >< INFO> 1849: cpp_module_handle_stream_cfg_event: frame_offset=3, input_fps=60.00, identity=0x30002 01-10 16:19:05.470 3211 4039 I mm-camera: <CPP >< INFO> 1849: cpp_module_handle_stream_cfg_event: frame_offset=3, input_fps=60.00, identity=0x30003 01-10 16:19:05.470 3211 4039 I mm-camera: <CPP >< INFO> 1849: cpp_module_handle_stream_cfg_event: frame_offset=3, input_fps=60.00, identity=0x30004 01-10 16:19:05.472 3211 4035 I mm-camera: <STATS_AF >< INFO> 1277: af_haf_update_tuning_data: Setting default init pos to far end 01-10 16:19:05.480 3211 4039 I mm-camera: <ISP >< INFO> 3697: mesh_rolloff40_normalize_table: full 640 480 out 640 480 offset 0 0 scale 1 01-10 16:19:05.484 3211 4036 I mm-camera: <ISP >< INFO> 90: rs_stats44_stats_config_update: warning: max_support_h_rgn invalid 8 01-10 16:19:05.484 3211 4039 I mm-camera: <CPP >< INFO> 4082: cpp_module_handle_streamon_event: identity=0x30004, stream-on done 01-10 16:19:05.484 3211 4039 E mm-camera: <SENSOR><ERROR> 1888: sensor_set_aec_init_settings: no valid entries in aec_get 01-10 16:19:05.484 3211 4039 E mm-camera: <SENSOR><ERROR> 4200: module_sensor_set_start_stream_on: Sensor AEC INIT failed!!! - non fatal 01-10 16:19:05.484 3211 4039 I mm-camera: <MCT >< INFO> 192: start_sof_check_thread: Starting SOF timeout thread session id =3 01-10 16:19:05.485 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 100000000 BW avg 65836800 BW inst 65836800 01-10 16:19:05.485 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-ON success for stream 0x30004, stream type=4 01-10 16:19:05.486 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-ON on stream 0x30002 stream type=1 01-10 16:19:05.487 3211 4036 I mm-camera: <ISP >< INFO> 90: rs_stats44_stats_config_update: warning: max_support_h_rgn invalid 8 01-10 16:19:05.487 3211 4039 I mm-camera: <CPP >< INFO> 4082: cpp_module_handle_streamon_event: identity=0x30002, stream-on done 01-10 16:19:05.487 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-ON success for stream 0x30002, stream type=1 01-10 16:19:05.488 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 100000000 BW avg 128044800 BW inst 128044800 01-10 16:19:05.488 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-ON on stream 0x30003 stream type=5 01-10 16:19:05.491 3211 4036 I mm-camera: <ISP >< INFO> 90: rs_stats44_stats_config_update: warning: max_support_h_rgn invalid 8 01-10 16:19:05.491 3211 4039 I mm-camera: <CPP >< INFO> 4082: cpp_module_handle_streamon_event: identity=0x30003, stream-on done 01-10 16:19:05.493 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 320000000 BW avg 426949056 BW inst 426949056 01-10 16:19:05.493 3211 4039 I mm-camera: <ISP >< INFO> 451: be_stats44_trigger_update: warning: Denomitor is zero hnum 0 vnum 0 01-10 16:19:05.497 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-ON success for stream 0x30003, stream type=5 01-10 16:19:05.498 3460 12630 I QCameraHalWatchdog: Stopped Watchdog Thread (121ms)[flush] 01-10 16:19:05.499 3222 3222 I Camera3-Device: disconnectImpl: E 01-10 16:19:05.499 3222 3222 I CameraLatencyHistogram: ProcessCaptureRequest latency histogram (572) samples: 01-10 16:19:05.499 3222 3222 I CameraLatencyHistogram: 40 80 120 160 200 240 280 320 360 inf (max ms) 01-10 16:19:05.499 3222 3222 I CameraLatencyHistogram: 99.83 0.17 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 (%) 01-10 16:19:05.499 3460 12630 I QCameraHalWatchdog: Starting Watchdog Thread (timeout 3000 msec)... 01-10 16:19:05.499 3460 12630 I QCamera : <HAL><INFO> close_camera_device: 12673: [KPI Perf]: E camera id 1 01-10 16:19:05.501 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30004 stream type=4 01-10 16:19:05.501 3211 4039 I mm-camera: <CPP >< INFO> 4173: cpp_module_handle_streamoff_event: info: doing stream-off for identity 0x30004 01-10 16:19:05.501 3211 4039 I mm-camera: <CPP >< INFO> 4358: cpp_module_handle_streamoff_event: info: stream-off done for identity 0x30004 01-10 16:19:05.503 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 266670000 BW avg 361112256 BW inst 361112256 01-10 16:19:05.504 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30004, stream type=4 01-10 16:19:05.505 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30002 stream type=1 01-10 16:19:05.505 3211 4039 I mm-camera: <CPP >< INFO> 4173: cpp_module_handle_streamoff_event: info: doing stream-off for identity 0x30002 01-10 16:19:05.505 3211 4039 I mm-camera: <CPP >< INFO> 4358: cpp_module_handle_streamoff_event: info: stream-off done for identity 0x30002 01-10 16:19:05.505 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30002, stream type=1 01-10 16:19:05.505 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30003 stream type=5 01-10 16:19:05.506 3211 4039 I mm-camera: <CPP >< INFO> 4173: cpp_module_handle_streamoff_event: info: doing stream-off for identity 0x30003 01-10 16:19:05.506 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 266670000 BW avg 298904256 BW inst 298904256 01-10 16:19:05.507 3211 4031 I mm-camera: <CPP >< INFO> 359: cpp_hardware_set_clock: Set clock 100000000 BW avg 0 BW inst 0 01-10 16:19:05.507 3211 4039 I mm-camera: <CPP >< INFO> 4358: cpp_module_handle_streamoff_event: info: stream-off done for identity 0x30003 01-10 16:19:05.510 3211 4039 I mm-camera: <MCT >< INFO> 223: stop_sof_check_thread: Stopping SOF timeout thread session =3 01-10 16:19:05.510 3211 4179 I mm-camera: <MCT >< INFO> 149: mct_bus_sof_thread_run: Closing SOF tracker thread 01-10 16:19:05.510 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30003, stream type=5 01-10 16:19:05.511 3211 4039 I mm-camera: <MCT >< INFO> 5748: mct_pipeline_process_set: STREAM-OFF on stream 0x30001 stream type=7 01-10 16:19:05.511 3211 4039 I mm-camera: <MCT >< INFO> 5763: mct_pipeline_process_set: STREAM-OFF success for stream 0x30001, stream type=7 01-10 16:19:05.514 3211 4039 I mm-camera: <MCT >< INFO> 5828: mct_pipeline_process_set: Issuing DEL_STREAM on stream 0x30002 and stream type=1 01-10 16:19:05.514 3211 4039 I mm-camera: <ISP >< INFO> 961: isp_port_check_caps_unreserve: ide 30002 port 0xf1bc7680 01-10 16:19:05.515 3211 4039 I mm-camera: <MCT >< INFO> 5839: mct_pipeline_process_set: Stream 0x30002 and stream type=1, successfully deleted 01-10 16:19:05.515 3211 4039 I mm-camera: <MCT >< INFO> 5828: mct_pipeline_process_set: Issuing DEL_STREAM on stream 0x30003 and stream type=5 01-10 16:19:05.515 3211 4039 I mm-camera: <ISP >< INFO> 961: isp_port_check_caps_unreserve: ide 30003 port 0xf1bc7680 01-10 16:19:05.516 3211 4039 I mm-camera: <MCT >< INFO> 5839: mct_pipeline_process_set: Stream 0x30003 and stream type=5, successfully deleted 01-10 16:19:05.517 3211 4039 I mm-camera: <MCT >< INFO> 5828: mct_pipeline_process_set: Issuing DEL_STREAM on stream 0x30004 and stream type=4 01-10 16:19:05.517 3211 4039 I mm-camera: <ISP >< INFO> 961: isp_port_check_caps_unreserve: ide 30004 port 0xf1bc7680 01-10 16:19:05.517 3211 4039 I mm-camera: <MCT >< INFO> 5839: mct_pipeline_process_set: Stream 0x30004 and stream type=4, successfully deleted 01-10 16:19:05.517 3211 4039 I mm-camera: <MCT >< INFO> 5828: mct_pipeline_process_set: Issuing DEL_STREAM on stream 0x30001 and stream type=7 01-10 16:19:05.517 3211 4039 I mm-camera: <MCT >< INFO> 5839: mct_pipeline_process_set: Stream 0x30001 and stream type=7, successfully deleted 01-10 16:19:05.518 3211 3362 E mm-camera: <IMGLIB><ERROR> 303: AllocateBuffers: Invalid dimensions 0x0 01-10 16:19:05.518 3211 3362 E mm-camera: <IMGLIB><ERROR> 203: img_algo_preload: Preload: Failed to allocate buffer, rc -4 01-10 16:19:05.518 3211 4039 I mm-camera: <ISP >< INFO> 6721: isp_util_print_meta_stream_info: Deallocate resources 01-10 16:19:05.518 3211 3362 E mm-camera: <IMGLIB><ERROR> 119: module_imgbase_client_preload_exec: IMG_CORE_PRELOAD failed -4 01-10 16:19:05.521 3211 4039 I mm-camera: <ISP >< INFO> 2534: isp_resource_allocate: INFO: ISP resource acquired: session 3 num_isp 1 01-10 16:19:05.521 3211 4039 I mm-camera: <ISP >< INFO> 2536: isp_resource_allocate: INFO: ISP resource acquired: hw_id 1 01-10 16:19:05.521 3211 4039 I mm-camera: <ISP >< INFO> 6571: isp_util_decide_stream_mapping: stream_port_map num streams 0 01-10 16:19:05.522 3211 4039 E mm-camera: <MCT ><ERROR> 1468: mct_pipeline_decide_hw_wakeup: Couldn't find meta stream 01-10 16:19:05.523 3460 12630 I QCamera : <HAL><INFO> closeCamera: 1071: [KPI Perf]: E PROFILE_CLOSE_CAMERA camera id 1 01-10 16:19:05.523 3211 3211 I mm-camera: <MCT >< INFO> 225: mct_controller_destroy: Initiating destroy sequence for session = 3 01-10 16:19:05.524 3211 3211 I mm-camera: <MCT >< INFO> 243: mct_controller_destroy: serv_thread closed 01-10 16:19:05.524 3211 3211 I mm-camera: <MCT >< INFO> 258: mct_controller_destroy: bus_handler thread closed 01-10 16:19:05.524 3211 3211 I mm-camera: <MCT >< INFO> 6569: mct_pipeline_stop_session: Initiating stop_session on session 3 01-10 16:19:05.524 3211 3211 I mm-camera: <MCT >< INFO> 5828: mct_pipeline_process_set: Issuing DEL_STREAM on stream 0x3000f and stream type=10 01-10 16:19:05.524 3211 3211 I mm-camera: <ISP >< INFO> 961: isp_port_check_caps_unreserve: ide 3000f port 0xf1bc7680 01-10 16:19:05.525 3211 3211 D MotISP_VSTAB: destroyed 01-10 16:19:05.527 3211 3211 I mm-camera: <MCT >< INFO> 5839: mct_pipeline_process_set: Stream 0x3000f and stream type=10, successfully deleted 01-10 16:19:05.527 3211 4183 I mm-camera: <MCT >< INFO> 6303: mct_pipeline_stop_session_thread: Stop module name: iface - E 01-10 16:19:05.527 3211 4182 I mm-camera: <MCT >< INFO> 6303: mct_pipeline_stop_session_thread: Stop module name: sensor - E 01-10 16:19:05.527 3211 4182 I mm-camera: <MCT >< INFO> 6311: mct_pipeline_stop_session_thread: Stop module name: sensor - X 01-10 16:19:05.528 3211 4184 I mm-camera: <MCT >< INFO> 6303: mct_pipeline_stop_session_thread: Stop module name: isp - E 01-10 16:19:05.528 3211 4184 I mm-camera: <ISP >< INFO> 491: isp_module_stop_session: session id 3 01-10 16:19:05.528 3211 4184 I mm-camera: <ISP >< INFO> 532: isp_module_stop_session: warning: ISP resource is not freed on time! 01-10 16:19:05.528 3211 4183 I mm-camera: <MCT >< INFO> 6311: mct_pipeline_stop_session_thread: Stop module name: iface - X 01-10 16:19:05.528 3211 4186 I mm-camera: <MCT >< INFO> 6303: mct_pipeline_stop_session_thread: Stop module name: pproc - E 01-10 16:19:05.529 3211 4185 I mm-camera: <MCT >< INFO> 6303: mct_pipeline_stop_session_thread: Stop module name: stats - E 01-10 16:19:05.529 3211 3211 I mm-camera: <MCT >< INFO> 6591: mct_pipeline_stop_session: Modules left: 5 01-10 16:19:05.529 3211 3211 I mm-camera: <MCT >< INFO> 6591: mct_pipeline_stop_session: Modules left: 4 01-10 16:19:05.529 3211 4187 I mm-camera: <MCT >< INFO> 6303: mct_pipeline_stop_session_thread: Stop module name: imglib - E 01-10 16:19:05.531 3211 4186 E quadracfa_dummy: inside dummy remosaic_deinit 1 01-10 16:19:05.531 3211 4186 I mm-camera: <MCT >< INFO> 6311: mct_pipeline_stop_session_thread: Stop module name: pproc - X 01-10 16:19:05.531 3211 3211 I mm-camera: <MCT >< INFO> 6591: mct_pipeline_stop_session: Modules left: 3 01-10 16:19:05.531 3211 4187 I mm-camera: <MCT >< INFO> 6311: mct_pipeline_stop_session_thread: Stop module name: imglib - X 01-10 16:19:05.531 3211 3211 I mm-camera: <MCT >< INFO> 6591: mct_pipeline_stop_session: Modules left: 2 01-10 16:19:05.532 3211 4185 I mm-camera: <MCT >< INFO> 6311: mct_pipeline_stop_session_thread: Stop module name: stats - X 01-10 16:19:05.532 3211 3211 I mm-camera: <MCT >< INFO> 6591: mct_pipeline_stop_session: Modules left: 1 01-10 16:19:05.534 3211 4184 I mm-camera: <MCT >< INFO> 6311: mct_pipeline_stop_session_thread: Stop module name: isp - X 01-10 16:19:05.534 3211 3211 I mm-camera: <MCT >< INFO> 6591: mct_pipeline_stop_session: Modules left: 0 01-10 16:19:05.534 3211 3211 I mm-camera: <MCT >< INFO> 6601: mct_pipeline_stop_session: Stopped session 3 successfully 01-10 16:19:05.534 3211 3211 I mm-camera: <MCT >< INFO> 6886: mct_pipeline_destroy: E: session:3 01-10 16:19:05.534 3211 3211 I mm-camera: <MCT >< INFO> 6964: mct_pipeline_destroy: X: Pipeline destroyed successfully session =3 01-10 16:19:05.534 3211 3211 I mm-camera: <MCT >< INFO> 292: mct_controller_destroy: X Successfully closed mct_controller session 3 01-10 16:19:05.535 3141 3169 I SDM : ResourceImpl::SetMaxBandwidthMode: new bandwidth mode=0 01-10 16:19:05.535 3222 3222 I CameraProviderManager: Camera device<EMAIL_ADDRESS>torch status is now AVAILABLE_OFF 01-10 16:19:05.535 3222 3222 I CameraService: onTorchStatusChangedLocked: Torch status changed for cameraId=1, newStatus=1 01-10 16:19:05.535 3460 12630 I QCamera : <HAL><INFO> closeCamera: 1117: [KPI Perf]: X PROFILE_CLOSE_CAMERA camera id 1, rc: 0 01-10 16:19:05.536 3460 12630 I QCamera : <HAL><INFO> close_camera_device: 12675: [KPI Perf]: X 01-10 16:19:05.536 3460 12630 I QCameraHalWatchdog: Stopped Watchdog Thread (37ms)[close_camera_device] 01-10 16:19:05.538 3222 3222 I CameraLatencyHistogram: Stream 0 dequeueBuffer latency histogram (572) samples: 01-10 16:19:05.538 3222 3222 I CameraLatencyHistogram: 5 10 15 20 25 30 35 40 45 inf (max ms) 01-10 16:19:05.538 3222 3222 I CameraLatencyHistogram: 100.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 (%) 01-10 16:19:05.541 3222 3222 I CameraLatencyHistogram: Stream 1 dequeueBuffer latency histogram (572) samples: 01-10 16:19:05.541 3222 3222 I CameraLatencyHistogram: 5 10 15 20 25 30 35 40 45 inf (max ms) 01-10 16:19:05.541 3222 3222 I CameraLatencyHistogram: 99.30 0.70 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 (%) 01-10 16:19:05.541 3331 3993 W GraphicBufferSource: released unpopulated slots: [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63] 01-10 16:19:05.542 3222 3222 I Camera3-Device: disconnectImpl: X 01-10 16:19:05.543 3222 3222 D CameraService: disconnect: unlinkToDeath 01-10 16:19:05.546 3222 3222 I CameraService: disconnect: Disconnected client for camera 1 for PID 3628 01-10 16:19:05.548 3628 3756 D UseCaseAttachState: Active and attached use case: [] for camera: 1 01-10 16:19:05.549 3628 3756 D DeferrableSurface: use count-1, useCount=0 closed=true androidx.camera.core.SurfaceRequest$2@51e30e1 01-10 16:19:05.549 3628 3756 D DeferrableSurface: Surface no longer in use[total_surfaces=3, used_surfaces=2](androidx.camera.core.SurfaceRequest$2@51e30e1} 01-10 16:19:05.549 3628 3756 D DeferrableSurface: Surface terminated[total_surfaces=2, used_surfaces=2](androidx.camera.core.SurfaceRequest$2@51e30e1} 01-10 16:19:05.550 3628 3756 D DeferrableSurface: use count-1, useCount=0 closed=true androidx.camera.core.impl.ImmediateSurface@b913dbf 01-10 16:19:05.550 3628 3756 D DeferrableSurface: Surface no longer in use[total_surfaces=2, used_surfaces=1](androidx.camera.core.impl.ImmediateSurface@b913dbf} 01-10 16:19:05.550 3628 3756 D DeferrableSurface: Surface terminated[total_surfaces=1, used_surfaces=1](androidx.camera.core.impl.ImmediateSurface@b913dbf} 01-10 16:19:05.550 3628 3756 D DeferrableSurface: use count-1, useCount=0 closed=true androidx.camera.core.SurfaceRequest$2@c1c8492 01-10 16:19:05.550 3628 3756 D DeferrableSurface: Surface no longer in use[total_surfaces=1, used_surfaces=0](androidx.camera.core.SurfaceRequest$2@c1c8492} 01-10 16:19:05.550 3628 3756 D DeferrableSurface: Surface terminated[total_surfaces=0, used_surfaces=0](androidx.camera.core.SurfaceRequest$2@c1c8492} 01-10 16:19:05.551 3628 3628 D SurfaceViewImpl: Safe to release surface. 01-10 16:19:05.551 3628 4012 D Recorder: Surface closed: 18985970, Current surface: 18985970 01-10 16:19:05.551 3628 4012 D Recorder: Transitioning Recorder internal state: IDLING --> INITIALIZING 01-10 16:19:05.551 3628 4012 D Recorder: Transitioning audio state: INITIALIZING --> INITIALIZING 01-10 16:19:05.552 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} CameraDevice.onClosed() 01-10 16:19:05.553 3628 3756 D Camera2CameraImpl: {Camera@2acd80c[id=1]} Transitioning camera internal state: CLOSING --> INITIALIZED 01-10 16:19:05.553 3628 4012 I MediaCodec: (0x72059b5000) release 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: Recalculating open cameras: 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: Camera State 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: ------------------------------------------------------------------- 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: Camera@2acd80c[id=1] CLOSED 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: Camera@39bde0[id=0] UNKNOWN 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: ------------------------------------------------------------------- 01-10 16:19:05.556 3628 3756 D CameraStateRegistry: Open count: 0 (Max allowed: 1) 01-10 16:19:05.557 3628 3756 D CameraStateMachine: New public camera state CameraState{type=CLOSED, error=null} from CLOSED and null 01-10 16:19:05.557 3628 3756 D CameraStateMachine: Publishing new public camera state CameraState{type=CLOSED, error=null} 01-10 16:19:05.561 3331 3993 I OMX-VENC: Component Deinit 01-10 16:19:05.562 3331 3993 I OMX-VENC: Video encode perflock released 01-10 16:19:05.568 3628 4012 D VideoEncoder: Transitioning encoder internal state: CONFIGURED --> RELEASED 01-10 16:19:05.977 3308 4189 E ResolverController: No valid NAT64 prefix (100, <unspecified>/0) 01-10 16:19:08.002 3308 4191 E ResolverController: No valid NAT64 prefix (100, <unspecified>/0) Maybe try disabling video or photo or frameProcessor? How should I do? What do you mean? Disable video or photo or frameProcessor. Just don't pass them to the <Camera> component Does the flash not fire when taking a photo or when recording (or both)? HDR is a bit weird - it works when the device supports it, but when the device doesn't support it it is somehow still being reported as "supported". @mrousavy : It also happens that with the main camera the videos are not recorded. If I record a video with the secondary camera they are recorded. You don't even see the HDR icon. It's an icon that's missing. In an example app. I have even fixed it right now. On android there seems to be a lot of problems. Thanks, I actually enabled forking on this repository so you can send me PRs fixing those problems anytime :) What android device have you tested on? Huawei P10 It also happens that with the main camera the videos are not recorded. What is the main camera, and what is the second camera? @mrousavy : Main camera the one you use to take pictures normally (those behind the device) Secondary room for selfies (those on the device screen). @mrousavy : I pull on the new changes. Icon Hdr on see The goBack is not working @mrousavy : I tried on the Tablet Samsung Tab A (2016), to make the camera work I have to disable the video. Can any one help me what i am doing wrong ? Im using react native vision camera const devices = useCameraDevices() const device = devices.back @ShaheerArshad-nearpeer : Did you create a new project and then install the library? Are you testing on an Android emulator or Android physical device? You should ask @mrousavy. @mrousavy : Are there any news? @mrousavy : Are there any news? @mrousavy : Are there any news? What were you trying to do? clone the project: https://github.com/mrousavy/react-native-vision-camera try to run the example code on emulator (Pixel 2 XL API 29) & got the same error on the back camera What happened instead? also I have the same error on my physical device Motorola Moto G20 on both cameras front & back @mrousavy : Are there any news? @Stukz : Does the flash work on the g20? @Stukz : Does the flash work on the g20? @Angelk90 it doesn't work @Stukz : What other devices have you tried on? Both cameras don't work? @Stukz : What other devices have you tried on? Both cameras don't work? I only have my g20 as a physical device. yes, both cameras have the same issue. it's happening with video={true} @Stukz : I have tested on several Android devices it seems to not work correctly, if you say it does not work on the emulator either it will probably @mrousavy have to be able to test it against a physical device. Ps I saw that you are a sponsor what are the advantages? got it, so it's not a specific device, sad to hear that. Ps I saw that you are a sponsor what are the advantages? you can read them on his page. the fact that he's trying to maintain a critical feature like a Camera library on RN it's something to appreciate. I know it's not that much but I hope in the near future to increase my sponsors over the RN OSS maintainers @Stukz : We hope that this problem on Android will be solved in the future. From what I understand the problem appears to be camerax, but I don't understand why. Hey! So the problem here is the core difference between iOS Cameras (AVFoundation) and Android Cameras (CameraX). On iOS, the concept is similar to how I expose it in VisionCamera, you have a list of devices, each has a list of formats. With CameraX, all of this is a bit abstracted away and you only have a list of devices, but the concept of "formats" does not really exist. On Android, some Camera devices do not support photo and video at the same time. I am not sure why this is the case, but on iOS this always works. Also, apparently on CameraX there is no way to determine whether a camera supports video and photo at the same time before mounting the Camera, which is just dumb. I can only use guesses to find out if a device in the devices list supports video and photo. I have a contact to Google and will ask him a few questions, maybe I can come up with an alternative API for VisionCamera that solves this issue. Sorry for the complications, but this is a tough one! And thank you for the sponsorship @Stukz, means a lot to me! 🙏 Hey Marc thank you for the detailed explanation of what's going on in Android, let me know if you have any progress on it so I can give you some feedback about it hey @mrousavy hope you are doing well, I'm curious if you have any update about this issue? Hey Nicolas, I didn't investigate this any further for now since I didn't have the time to do so. I am thinking about a new API that could potentially solve this issue! I have same issue. Any one has solution please :( I removed all related frameRateIncluded then it works. It's fix temporarily @ozgurbayram, @NgocLe1001 : On five android devices I tested on, it didn't work properly. Seems like some problem with cameraX, when it says @mrousavy. CameraX is really not as good as they make it sound... :/ @mrousavy : So for the moment on android it doesn't work properly, right? @mrousavy unable to capture image in Android. I dont see any logs either, it works in iOS when i click,it captures. This is still an issue in 2.15.2. Most Android devices don't work unless video is disabled. Anyone find any better workarounds? Hey all! CameraX is just too limiting. It's not based on Camera Device Formats as AVFoundation is, so I need to do a little rewriting for the Android part to make everything work as flexible and powerful as on iOS. For this part specifically, I need to: Rewrite the native Android part from CameraX to Camera2 Create a new declarative API ontop of the native iOS and Android parts to make this whole Camera Device + Format selection way easier (describe your target FPS, HDR, resolution, etc and it will pick everything automatically) However since this is a really big project, I can't really work on this in my free time (I have a company to run). I have created a discussion for this, and want to put it all into VisionCamera V3, if you or your company is interested in this new version and wants to help support/fund the development of that, please reach out to me via Twitter or Email. Here's the technical discussion: https://github.com/mrousavy/react-native-vision-camera/issues/1376
2025-04-01T04:34:49.967020
2024-11-29T17:08:11
2705701797
{ "authors": [ "karol-haltenberger", "mrvisscher" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8840", "repo": "mrvisscher/1pt5deg-lifestyles-visualizations", "url": "https://github.com/mrvisscher/1pt5deg-lifestyles-visualizations/issues/1" }
gharchive/issue
EMBED issue So the website - being a content management system - is set up to embed external content from various providers serving it using oembed as media entities. I did find a module that allows inserting "raw" HTML in a similar way, but in order to have these reference local files (main.js, css, csvs and such) I would need to manually place them somewhere exactly and then reference them exactly ... a bit hacky. Even then I did not manage to open this HTML and have it work (d3 missing, error messages with imports etc.) If you could turn this into a single piece of HTML (without using <head> as I'll need to copy the body) with inline js, css and other resources ( minified where possible ) and only referencing external libraries (d3), it might just work. Hi @karol-haltenberger, The reason why you're not getting it to work is because the files hosted here are the source files that still need to be build using npm. I've attached a zip that contains the final .html .js and .css files. It is common practice to host the .js and .css files statically somewhere, and then reference their exact locations from the html file. I would argue that that's less hacky then putting everything inline as that makes the codebase harder to understand and to update at a later time. However, we can still do this if you prefer. Let me know! Kind regards, Marin dist-v2.zip @mrvisscher "...still need to be build..." Got it, my bad. It may be common practice for more "raw" websites, but - again - this is a higher level content manager. Adding the files would require direct access to the filesystem, knowledge of its structure and precision which is not currently available to / practiced by the average "user". If this is a one-off then I guess I can "hack it in" - although it kinda breaks the "concept" of the site - but if the project plans on creating and embedding more of these in the future then these extra hacks would have to be done each time. So yes, ideally I would like you to "compile" the whole thing into a single HTML snippet and if you (or someone else) do that for any other such diagram in the future, the corresponding media asset can just be created through the regular interface. Note, that this is might not work at all. NVM....it does not work...I'll have to come up with something else. It doesn't work on my end...the solution within the CMS I came up with...I'll get back to you later when I figure it out and if I need some changes in this code. Thanks.
2025-04-01T04:34:49.983846
2022-01-04T09:15:17
1093153588
{ "authors": [ "AlexGoemanDigipolis", "msawczyn", "vvboborykin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8841", "repo": "msawczyn/EFDesigner2022", "url": "https://github.com/msawczyn/EFDesigner2022/issues/6" }
gharchive/issue
Visual Designer <IP_ADDRESS> in VS2022 17.0.4 Problem with diagrams In a newly installed VS2022 I installed Visual Designer <IP_ADDRESS> I tried a new project from scratch, and then I cannot create a seconds diagram. In the entity model explorer when I right click on the root node the popup menu contains almost no actions besides "Copy" and "Properties". The default diagram works When opening an existing model (created in VS2019) we get a COM error trying to open one of the extra diagrams can someone confirm this is a problem ? I got the same problem (can't create second diagram) And neither can I :-( It's a Microsoft issue with the Modeling SDK. I'm working on getting them the necessary data to track down and fix the problem. That's two errors so far ... the conversion to asynchronous-only extensions and 64-bit hasn't been kind to this SDK.
2025-04-01T04:34:49.990776
2014-08-20T03:45:42
40660880
{ "authors": [ "dmamaril", "midhun1993" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8842", "repo": "mscdex/mmmagic", "url": "https://github.com/mscdex/mmmagic/issues/28" }
gharchive/issue
Error on Azure Deployment remote: npm ERR! Failed at the<EMAIL_ADDRESS>install script. remote: npm ERR! This is most likely a problem with the mmmagic package, remote: npm ERR! not with npm itself. remote: npm ERR! Tell the author that this fails on your system: remote: npm ERR! node-gyp rebuild remote: npm ERR! You can get their info via: remote: npm ERR! npm owner ls mmmagic remote: npm ERR! There is likely additional logging output above. remote: remote: npm ERR! System Windows_NT 6.2.9200 remote: npm ERR! command "d:\Program Files (x86)\nodejs\0.10.29\node.exe" "D:\Program Files (x86)\npm\1.4.10\node_modules\npm\bin\npm-cli.js" "install" "--production" remote: npm ERR! cwd d:\home\site\wwwroot remote: npm ERR! node -v v0.10.29 remote: npm ERR! npm -v 1.4.10 remote: npm ERR! code ELIFECYCLE remote: npm remote: remote: Error - Changes committed to remote repository but deployment to website failed. @dmamaril Hello, I'm experiencing the same issue, How we can fix it ?
2025-04-01T04:34:50.005834
2024-08-29T17:35:31
2495152299
{ "authors": [ "doggy8088", "jochenkirstaetter" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8847", "repo": "mscraftsman/generative-ai", "url": "https://github.com/mscraftsman/generative-ai/issues/30" }
gharchive/issue
gemini release 3 new models gemini-1.5-pro-exp-0827 gemini-1.5-flash-8b-exp-0827 gemini-1.5-flash-exp-0827 Ref 1: https://x.com/OfficialLoganK/status/1828480081574142227 Ref 2: https://www.unite.ai/google-releases-three-new-experimental-gemini-models/ I see that these three Models have already been added. Hello @doggy8088 Thank you so much for the references to new models. Really appreciate your support. Yes, I added them with Release 1.6.0 among others. Furthermore, the model class is not an enumeration but provides constants to the model names (I'm terrible at memorizing). Any method accepting a model is based on string data type. In case that a model is not present as a constant one can use the value as string directly. Thanks again, JoKi
2025-04-01T04:34:50.012527
2024-05-09T16:33:18
2288051877
{ "authors": [ "Its4Nik", "moonrailgun" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8848", "repo": "msgbyte/tianji", "url": "https://github.com/msgbyte/tianji/issues/66" }
gharchive/issue
improve monitor page like this: I was thinking about opening an issue with exactly the same idea and screenshot, lmao. I am sadly not really good with ts but I could try :D Hi, @Its4Nik I am sadly not really good with ts but I could try :D cool, its easy, because we prepare everything, just need a config which compose they together. if you have the intention to realize this requirement. I will be happy to assign it to you Hi, @Its4Nik I am sadly not really good with ts but I could try :D cool, its easy, because we prepare everything, just need a config which compose they together. if you have the intention to realize this requirement. I will be happy to assign it to you I would love to help the project yeah, but then i think i would need a deeper dive on discord or so, because i am still not finding eberyfile rhat i would like to have