added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:39:30.789168
2017-01-12T15:23:44
200395361
{ "authors": [ "marcosicp", "mateuszmarkowski" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8110", "repo": "mateuszmarkowski/jQuery-Seat-Charts", "url": "https://github.com/mateuszmarkowski/jQuery-Seat-Charts/issues/48" }
gharchive/issue
Need the 'id' also in the white spaces ('_') Hi, I need to have id also in underscores white spaces ('') because I need to work with that when the map complete renders. Do you know where I need to change @mateuszmarkowski ? I think is in: $row.append(character != '' ? //if the character is not an underscore (empty space) (function (naming) { //so users don't have to specify empty objects settings.seats[character] = character in settings.seats ? settings.seats[character] : {}; var id = overrideId ? overrideId : naming.getId(character, naming.rows[row], naming.columns[column]); seats[id] = new seat({ id: id, label: overrideLabel ? overrideLabel : naming.getLabel(character, naming.rows[row], naming.columns[column]), row: row, column: column, character: character }); seatIds.push(id); return seats[id].node(); })(settings.naming) : //this is just an empty space (_) $('<div></div>').addClass('seatCharts-cell seatCharts-space') //// Here??? ); Thanks! Yes, I think you found the right place, sorry for a late response. I did it thanks @mateuszmarkowski
2025-04-01T06:39:30.810658
2016-05-01T10:03:18
152191843
{ "authors": [ "KingScooty", "Tatsh", "jeffbyrnes", "mathiasbynens" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8111", "repo": "mathiasbynens/dotfiles", "url": "https://github.com/mathiasbynens/dotfiles/issues/656" }
gharchive/issue
What is the ~/bin/bash symlink used for? I'm just sieving through this project to make my own dotfiles, and i can't figure out what the bash symlink in ~/bin is used for. It's getting added to PATH in .bash_profile, but what does that enable? Is there a reason why it needs to be added to the path? After installing bash v4 with homebrew, it still needs to be added to /etc/shell manually, so i'm not sure what value this symlink is adding. Everything seems to work just fine without it?? Really appreciate it! If I run bash or bash --help or bash -c $cmd I want to use the brew-installed version of bash, not the old Bash v3 that OS X ships with. I thought as much, but bash --version reports 4.3.42(1)-release even without the bash symlink in ~/bin, which made me question whether it's even needed? echo $BASH_VERSION will always report the old Bash v3 that ships with OS X, regardless of the symlink being present or not. It's only until you add /usr/local/bin/bash to /etc/shell and run chsh -s /usr/local/bin/bash that the $BASH_VERSION changes to 4.3.42(1)-release. I automated that whole bit like so in my brew.sh: if ! grep '/usr/local/bin/bash' /etc/shells; then echo '/usr/local/bin/bash' | sudo tee -a /etc/shells; chsh -s /usr/local/bin/bash; fi; @jeffbyrnes Use fgrep -q or grep -F -q, so output does not occur on the if statement. sudo echo '/usr/local/bin/bash' >> /etc/shells, not sure why you would use tee for that. Ah, nice! @jeffbyrnes you thought about opening a PR for this? Ah, nice! @jeffbyrnes you thought about opening a PR for this? Also, does that mean that ~/bin/bash isn't needed anymore? @KingScooty I haven‘t, but I certainly can. Gonna be on the train in about an hour, so I’ll try to remember to open one once I settle in. @Tatsh sudo echo '/usr/local/bin/bash' >> /etc/shells — not sure why you would use tee for that. Because the sudo echo … version doesn’t work. You’re running echo as super user but it’s the >> that needs it.
2025-04-01T06:39:30.813917
2015-11-26T09:26:13
119003858
{ "authors": [ "Mottie", "mathiasbynens" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8112", "repo": "mathiasbynens/esrever", "url": "https://github.com/mathiasbynens/esrever/issues/8" }
gharchive/issue
Support Unicode 9 prepended concatenation marks It’s still a proposal for now, but hey. http://blog.unicode.org/2015/11/new-character-property-for-prepended.html Prepended_Concatenation_Mark Just curious... it looks like Prepended_Concatenation_Marks were added in Unicode 8 (ref)? I do see mention of these marks in Unicode 9, but I can't tell if any new ones were added. And on that note... the package.json for this library is still using Unicode 8, shouldn't it be updated for v9? http://unicode.org/reports/tr29/#Table_Combining_Char_Sequences_and_Grapheme_Clusters
2025-04-01T06:39:30.821014
2021-09-18T16:39:31
1000078533
{ "authors": [ "mathincubator" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8113", "repo": "mathincubator/mathlib", "url": "https://github.com/mathincubator/mathlib/issues/15" }
gharchive/issue
Mode tutorial of mode Mode tutorial of mode with example questions and math contest questions. http://ec2-34-200-244-39.compute-1.amazonaws.com/boards/5/topics/28/ Better to add a math contest example
2025-04-01T06:39:30.840423
2024-09-28T08:40:06
2554112798
{ "authors": [ "dpvc", "physikerwelt" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8114", "repo": "mathjax/MathJax", "url": "https://github.com/mathjax/MathJax/issues/3292" }
gharchive/issue
\mathcal uses incorrect font in Client side MathJax modes Issue Summary As described in https://phabricator.wikimedia.org/T375932 LaTeX input like $\mathcal A$ uses different fonts than it used with MathJax-node With a preprocessing script the LaTeX input is converted into MathML of this form <mi>&#x1D49C;</mi> To support MathJax we also added the classes <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">𝒜</mi> which might have been contra productive? Which seems to be recommended. The MathJax rendering looks more like $\mathscr A$ Steps to Reproduce: Login to https://en.wikipedia.beta.wmflabs.org/ Select Client Side Math rendering in the preferences https://en.wikipedia.beta.wmflabs.org/wiki/Special:Preferences#mw-prefsection-rendering-math Goto https://en.wikipedia.beta.wmflabs.org/wiki/T375932 See that it works with FF but fails with Chrome. Technical details: MathJax Version: 3.2.2 Client OS: Mac OS X 10.8.4 Browser: Firefox 130.0.1 I am using the following MathJax configuration: https://github.com/wikimedia/mediawiki-extensions-Math/blob/db81c960623d346ca8f43fb91ca3da92f868ea29/modules/ext.math.mathjax.js#L7 ( function () { 'use strict'; const extensionAssetsPath = mw.config.get( 'wgExtensionAssetsPath' ); window.MathJax = { loader: { // see https://docs.mathjax.org/en/latest/input/mathml.html load: [ '[mml]/mml3' ], // see https://docs.mathjax.org/en/latest/options/startup/loader.html paths: { mathjax: extensionAssetsPath + '/Math/modules/mathjax/es5' } } }; }() ); and loading MathJax via <script src="/w/extensions/Math/modules/mathjax/es5/input/mml.js" charset="UTF-8"></script> <script src="/w/extensions/Math/modules/mathjax/es5/input/mml/extensions/mml3.js" charset="UTF-8"></script> Supporting information: Please supply a link to a (live) minimal example page, when possible: https://en.wikipedia.beta.wmflabs.org/wiki/T375932 If your issue is with the display of the mathematics produced by MathJax, include a screen snapshot that illustrates the problem, when possible. See also https://phabricator.wikimedia.org/T375932 for more screenshots Check your browser console window for any error messages, and include them here. None Include the MathJax configuration you are using, and the script tag that loads MathJax itself. load: [ '[mml]/mml3' ], The underlying problem is that the Unicode Math Alphanumerics block includes only one script variant, and the choice of the variant to use is up to the fonts. Most seem to use a Roundhand style (as in \mathscr), so the Chancery form (\mathcal) is absent. (The STIX2 Math fonts have the Chancery version.) Since the MathML mathvariant has only one script value, that makes it difficult to implement both \mathscr and \mathcal. MathJax uses data-mjx-variant="-tex-calligraphic" to overcome this difficulty, but since MathML renderers don't know about this attribute, MathJax adds mathvariant="script" to at least get a script variant (whichever one the font implements). Note, however, that MathJax uses <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">A</mi> not <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">&#x1D49C;</mi>. This is because the MathML specification indicates that the Math Alphanumeric values are not to be influenced by CSS styles or mathvariant settings (see end of the third paragraph of the description of mathvariant in the specification): Note that the appearance of a mathematical alphanumeric symbol character should not be altered by surrounding mathvariant or other style declarations. That is, if you do <mi mathvariant="bold">&#x1D504;</mi> you get the non-bold Fraktur capital A U+1D504, not the bold one at U+1D56C. I suspect the main reason for that is when the mathvariant is on an mstyle element, lik <mstyle mathvariant="bold"> <mi>x<mi> <mo>+</mo> <mi> &#x1D504;</mi> </mstyle> The U+1D504 will not bin inadvertently changed to bold, while the other characters will. In any case, since you are specifying the character U+1D49C explicitly (not just A), the mathvariant and data-mjx-variant won't change the result. So you get the font's U+1D49C character, which for the MathJax fonts is the Roundhand version, not the Chancery one. Note that MathML-Core has removed the mathvariant attribute from its specification, except for the mi element, where only mathvariant="normal" is allowed (to avoid the automatic italicization that is done for single-character content). See the mi section of the specification. Unfortunately, that means you can't use <mi data-mjx-variant="-tex-calligraphic" mathvariant="script">A</mi> to get this to work in both MathML-Core and MathJax. The MathML-Core documentation notes that the issue exists (in the second note of this section of the specification). It suggests two different potential solutions, both being font dependent. The first is to use variant selectors U+FE00 and U+FE01 to specify Chancery and Roundhand respectively. I don't know whether browsers actually implement this or not; testing in Safari, Chrome, and Firefox on MacOS suggests they do not. Also, for <mi>&#x1D49C.</mi>, I get the Chancery version (from STIX2) in Firefox and Chrome, but the Roundhand version in Safari. As I said, it is font dependent. In any case, MathJax doesn't support the variant selectors, either. The second suggestion is to use font features and the font-variant-alternatives CSS property, but again, that is font dependent, and MathJax doesn't support that, either. I haven't checked what the browser support for that is. So there seems to be no direct way to reliably get both \mathscr and \mathcal to work in MathML-Core. A cross-browser solution that works with both MathML-Core and MathJax seems out of reach at the moment. In any case, MathJax is producing the correct output for the MathML you are using. Thank you for enlightening me! In general, this makes a lot of sense. However, this brings one backdoor to my mind. So there seems to be no direct way to reliably get both \mathscr and \mathcal to work in MathML-Core. A cross-browser solution that works with both MathML-Core and MathJax seems out of reach at the moment. We don't have \mathscr as an option. It's not safelisted in the texvc grammar. Is there a way (maybe via an input filter) to make MathJax display \mathcal like \mathsrc? Is there a way (maybe via an input filter) to make MathJax display \mathcal like \mathsrc? Well, you could replace the Roundhand script characters by the Chancery ones in MathJax's font definitions. That can be done in the MathJax startup ready() function as follows: window.MathJax = { remapChars(v1, v2, base, map, font) { const c1 = v1.chars; const c2 = v2.chars; for (let i = 0; i < 26; i++) { const data1 = c1[map[i] || base + i] || []; const data2 = c2[0x41 + i]; if (data1.length === 0) c1[base + i] = data1; [0, 1, 2].forEach(j => data1[j] = data2[j]); data1[3] = {...data2[3], f: font, c: String.fromCharCode(0x41 + i)}; } }, startup: { ready() { MathJax.startup.defaultReady(); const variant = MathJax.startup.document.outputJax.font.variant; const map = {1: 0x212C, 4: 0x2130, 5: 0x2131, 7: 0x210B, 8: 0x2110, 11: 0x2112, 12: 0x2133, 17: 0x211B}; MathJax.config.remapChars(variant.normal, variant['-tex-calligraphic'], 0x1D49C, map, 'C'); MathJax.config.remapChars(variant.normal, variant['-tex-bold-calligraphic'], 0x1D4D0, {}, 'CB'); } } }; This replaces the original script (Roundhand) data with the calligraphic (Chancery) character definitions, so that the Unicode Math Alphanumeric script characters will be displayed as Chancery characters. That means the Roundhand variant is no longer available. I think that is what you were asking for. This exactly matches what I understand was desired. Thank you so much.
2025-04-01T06:39:30.862686
2015-05-08T14:07:54
74374779
{ "authors": [ "matigo" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8115", "repo": "matigo/nice_social", "url": "https://github.com/matigo/nice_social/issues/17" }
gharchive/issue
Firefox Writes Global Timeline "Backwards" On Load Firefox, and only Firefox, seems to disregard the .before() and .after() functions in jQuery, publishing posts from the timeline at the top of the column regardless of their publication order. A better way of filling the timeline is needed here. This has been completely resolved. Yay!
2025-04-01T06:39:30.918775
2017-09-01T23:29:40
254772177
{ "authors": [ "matortheeternal", "megaloblast" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8116", "repo": "matortheeternal/zedit", "url": "https://github.com/matortheeternal/zedit/issues/6" }
gharchive/issue
Compatibility with MO Not necessarily an issue with zEdit, but when running zEdit through MO v<IP_ADDRESS>, I am unable to save the esp due to failure to create directory. failure to create directory. Can you expand? There should be no directory created when plugins are saved by zEdit. It just saves plugins into the data directory. MO should detect the saved plugins and move them to overwrite. http://i.imgur.com/CMDu7Am.png the issue happens even when I try outside of MO as well. I'm unable to save any changes. Sounds like an issue with file system permissions. The only directory that gets created when files are saved is the directory for backups. Resolved with Megaloblast on Discord. Fix will be included in the next build. If you're experienced this issue and need a a build before then, grab this (this build only includes a fix for saving plugins, it's otherwise identical to v0.0.1).
2025-04-01T06:39:30.972689
2023-01-09T10:23:18
1525300786
{ "authors": [ "erikjohnston" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8117", "repo": "matrix-org/complement", "url": "https://github.com/matrix-org/complement/pull/579" }
gharchive/pull-request
Fix panics due to closed channel in some tests A number of tests start goroutines that write into channels when completed, however the tests would close the channel if it exited early (e.g. due to a timeout). Instead, make the goroutine that writes into the channel the one that closes it. This does mean that the goroutines will continue past the end of the test (?), but they did that before too. I am wondering if there is a better way of doing these things... I suppose the right thing to do is to have the goroutine check a "signal channel" after it completes long-running operations (such as /sync) to make sure that it should continue execution before writing to channels. I suppose that might make sense for the one goroutine that is multiple statements, but I don't see the benefit for the others (like the one in the example)? (merging now to unblock fixing flakey tests)
2025-04-01T06:39:30.975158
2022-02-05T08:22:09
1124869539
{ "authors": [ "kittykat", "reivilibre", "squahtx" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8118", "repo": "matrix-org/conference-bot", "url": "https://github.com/matrix-org/conference-bot/issues/119" }
gharchive/issue
Allow talks with no Q&A to disable it so the bot doesn't shout @kittykat For WEDF, do any of the talks have Q&A? Currently all JSON-schedule talks will have Q&A disabled; I'd have to add in support for Q&A if it's desired. @reivilibre Sander confirmed no Q&A rooms see https://github.com/matrix-org/conference-bot/pull/135 for this year: Q&A availability will be determined by Room ID — D.* and M.test will have Q&A
2025-04-01T06:39:30.983310
2023-09-22T12:24:04
1908810655
{ "authors": [ "jacotec", "sandhose" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8119", "repo": "matrix-org/matrix-authentication-service", "url": "https://github.com/matrix-org/matrix-authentication-service/issues/1835" }
gharchive/issue
ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107) I've configured the MAS exactly following the documentation, found the "metrics" issue here and changed "none" to "prometheus", so MAS starts. But always shortly after the start, I'm getting repeated error messages: 2023-09-22T12:18:06.012090Z INFO cli.run.init: mas_cli::commands::server: Starting task worker worker_name="7lvKv8JSRD" 2023-09-22T12:18:06.076241Z INFO cli.run.init: mas_cli::commands::server: Listening on http://[::]:8082 with resources [Discovery, Human, OAuth, Compat, GraphQL { playground: true }, Assets { path: "./share/assets/" }] 2023-09-22T12:18:06.079178Z INFO cli.run.init: mas_cli::commands::server: Listening on http://[::]:8083 with resources [Health] 2023-09-22T12:18:15.781569Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107) 2023-09-22T12:18:25.785975Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107) 2023-09-22T12:18:35.787174Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107) 2023-09-22T12:18:45.788918Z ERROR mas_listener::server: error shutting down connection: Socket not connected (os error 107) The ports 8082 and 8083 (8080 and 8081 are already used by bridges) are free, and after starting MAS I see that 8082 and 8083 are happily used as listeners from MAS. So I have no clue what these error messages want to tell me. Also, when opening Element-Web I'm always getting: Es gab ein Problem bei der Kommunikation mit dem Heim-Server. Bitte versuche es später erneut.(M_UNRECOGNIZED) But I guess as long as the errors above are showing up, MAS is not working properly. Either there is an issue, or something is still undocumented ... This should probably be clearer in the documentation, but in the section about setting up a reverse proxy in front of it, it mentions that a few Matrix C-S API should be handled by MAS See the last paragraph here: https://matrix-org.github.io/matrix-authentication-service/setup/reverse-proxy.html#example-nginx-configuration The errors spitting out are I think unrelated, but something I need to investigate @sandhose I have a proxy in front, my HAProxy. But no relation to this error, if MAS does not run there's nothing to proxy ;-) My point is, in your setup, /_matrix/client/v3/login (and a few other endpoints) should not be handled by Synapse, they should be handled by MAS. This is a requirement for "legacy" (non-OIDC-native clients including Element Web) to work with MAS @sandhose Indeed the passing of the endpoints to MAS was missing, MAS is now called and the registration / authentication flow looked good up to a point where I grant the permissions for the Element-Web client. Then ... #1864 I think I captured most of the issues you had in other issues? #1940 #1941 And I think the logs you had were mostly resolved by #1814 and #1775
2025-04-01T06:39:30.996581
2020-07-28T09:46:17
666934671
{ "authors": [ "michaelkaye", "t3chguy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8120", "repo": "matrix-org/matrix-static", "url": "https://github.com/matrix-org/matrix-static/issues/97" }
gharchive/issue
Docker image does not contain register-guest helper I think it only copies the main matrix-static binary into the image. I'm not up to date on how such helpers are normally handled in dockerland, would it be loaded into the same image even though the only way to run it would be to mangle the entrypoint at runtime?
2025-04-01T06:39:30.997758
2023-01-20T09:46:39
1550568804
{ "authors": [ "andremenrath", "axelsimon" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8121", "repo": "matrix-org/moodle-mod_matrix", "url": "https://github.com/matrix-org/moodle-mod_matrix/issues/15" }
gharchive/issue
Undefined function profile_load_custom_fields in src/Plugin/Infrastructure/MoodleFunctionBasedMatrixUserIdLoader.php is not found. Probably it needs to be escaped and the /user/profile/lib.php needs to be required. Hi @andremenrath, sorry for not getting to this earlier. Would you want to make a PR for this?
2025-04-01T06:39:31.052120
2016-02-29T10:08:36
137208059
{ "authors": [ "ibutra", "iosdeveloper" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8122", "repo": "matryer/bitbar", "url": "https://github.com/matryer/bitbar/pull/261" }
gharchive/pull-request
Add templateImage parameter Added a templateImage parameter as discussed in Slack. This allows Mac OS X to change the styling in the StatusBar to fit the different themes better. 👍 We should add a tutorial plugin to show how to use images too. Alright, I try to think of one and make a pull request Pull request: https://github.com/matryer/bitbar-plugins/pull/290
2025-04-01T06:39:31.063037
2020-11-03T03:09:23
734980637
{ "authors": [ "Vbbab" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8123", "repo": "mattberkowitz/autoclose-html", "url": "https://github.com/mattberkowitz/autoclose-html/issues/219" }
gharchive/issue
Autoclose closes "never-close" tags [Issue when loading configs] Certain tags, like <input />, do not have a closing tag, but autoclose-html closes them anyway (despite them being configured as never-close tags!), in the improper format: Instead of closing like <input /> or even just leaving the tag be, it closes the tag like <input></input> Which browsers can generally autocorrect, but is probably not best practice... And yes, <input> is configured as a never-close tag. Determined this to be a config loading issue, with atom.config.observe. Gonna try to fix. Seems to be an Atom bug? Submitted Issue #21642 on Atom.
2025-04-01T06:39:31.090600
2019-05-14T21:28:37
444133666
{ "authors": [ "deanwhillier", "mattermod" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8124", "repo": "mattermost/desktop", "url": "https://github.com/mattermost/desktop/pull/976" }
gharchive/pull-request
[MM-15389] Prevent installing 32bit on 64bit and vise versa This PR configures the Desktop app MSI installers to prevent installation on non-matching-bit systems, i.e. the 32-bit installer won’t install on 64-bit systems. This issue has been automatically labelled "stale" because it hasn't had recent activity. A core team member will check in on the status of the PR to help with questions. Thank you for your contribution! /cc @jasonblais @hanzei Closing in favour of a single branch based off of release-test-msi for all MSI related updates to make further updates more manageable. New PR pending.
2025-04-01T06:39:31.092485
2021-09-21T13:38:39
1002505235
{ "authors": [ "cwarnermm" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8125", "repo": "mattermost/docs", "url": "https://github.com/mattermost/docs/pull/5060" }
gharchive/pull-request
Added sphinx-tabs extension This PR adds a maintained public Sphinx extension called sphinx-tabs to enable the ability to develop product documentation content in a tabbed format. We've locked this extension down to the current 3.2.0 version. When this extension is merged into master, I will close the initial POC of this functionality implemented via https://github.com/mattermost/docs/pull/5042
2025-04-01T06:39:31.098250
2021-03-10T16:24:18
827996024
{ "authors": [ "enahum", "josephbaylon", "mattermod" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8126", "repo": "mattermost/mattermost-mobile", "url": "https://github.com/mattermost/mattermost-mobile/pull/5219" }
gharchive/pull-request
MM-31341 Use first metadata info for inline md images Summary When an inline image url contains some query strings, the metadata for the image doesn't match the link thus the image was not showing, with this PR if the link does not match the one in the metadata, we will then use the first image available in the metadata to be displayed. PR also takes care of black images on top of a black background. Ticket Link https://mattermost.atlassian.net/browse/MM-31341 https://mattermost.atlassian.net/browse/MM-25811 Building app in separate branch. Successfully building: https://circleci.com/gh/mattermost/mattermost-mobile/14347 https://circleci.com/gh/mattermost/mattermost-mobile/14346 Artifact links: https://14347-70265724-gh.circle-artifacts.com/0/tmp/artifacts/Mattermost_Beta.apk https://14346-70265724-gh.circle-artifacts.com/0/tmp/artifacts/Mattermost_Beta.ipa Hi @enahum The image is now being displayed but compared to webapp, the background and sizing look different. Not sure if this is expected. Mobile Webapp @josephbaylon that is actually expected, we have always needed to render the inline images into its own line
2025-04-01T06:39:31.099641
2023-02-08T14:45:56
1576251932
{ "authors": [ "enahum", "mattermost-build" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8127", "repo": "mattermost/mattermost-mobile", "url": "https://github.com/mattermost/mattermost-mobile/pull/7099" }
gharchive/pull-request
Android fix Summary Fixes android 13 notification permission request and unsigned builds NONE /cherry-pick release-2.0 Cherry pick is scheduled.
2025-04-01T06:39:31.162935
2023-03-20T11:57:54
1631958629
{ "authors": [ "mattermost-build", "pvev" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8128", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/12374" }
gharchive/pull-request
MM-49862 - remove no longer needed webapp calls to action and pulsating dots Summary This PR removes two no longer needed calls to action: Message Actions pulsating dot and New badge on the forward button. Also adds a condition for the custom status pulsating dot and update status post button where these two are just going to be shown once the user creation time has passed 7 days. Ticket Link https://mattermost.atlassian.net/browse/MM-49862 Related Pull Requests n/a Screenshots Before: After: The only change for this one is that won't show during the first seven days after user creation Release Note NONE E2E tests not automatically triggered, because PR has no approval yet. Please ask a developer to review and then try again to attach the QA label. Migrated to https://github.com/mattermost/mattermost-server/pull/22699
2025-04-01T06:39:31.168247
2018-12-10T16:38:30
389385812
{ "authors": [ "grundleborg", "hmhealey" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8129", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/2166" }
gharchive/pull-request
MM-13342: Fix dividers showing when menu section is empty. Summary Fix dividers showing when menu section is empty. This fix is a little ugly - it checks the children-of-children are not null to work around connected components themselves not being null but having the actual component within being null. Ticket Link https://mattermost.atlassian.net/browse/MM-13342 Checklist [x] Ran make check-style to check for style errors (required for all pull requests) [x] Ran make test to ensure unit and component tests passed [x] Added or updated unit tests (required for all new features) If you're wanting to avoid adding that extra introspection, perhaps we could actually fix this using CSS? If you look here, we already use .divider + .divider to hide the dividers for empty sections other than the first, so we can add .divider:firstChild to hide that first one as well. I don't think that will work, as we're not successfully hiding any of the dividers currently with the CSS. This fix targets all the unnecessary ones, not just the first one. Then again my CSS-foo isn't great so maybe I'm misunderstanding it? Sigh. Looks like this was a duplicate ticket anyway and Asaad has fixed it here https://github.com/mattermost/mattermost-webapp/pull/2152 I've merged and cherry picked that PR, so we could test again now to see if it's still an issue Looks fixed to me now with Asaad's changes. Closing this PR.
2025-04-01T06:39:31.170869
2019-05-24T07:21:29
448016710
{ "authors": [ "composednitin", "saturninoabril", "thekiiingbob" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8130", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/2845" }
gharchive/pull-request
M14126 Message Draft Pencil Icon - No text, only file attachment Summary Added e2e test to validate message draft with attachment after switching channel. Ticket Link Fixes https://github.com/mattermost/mattermost-server/issues/10274 @saturninoabril Could you please review @composednitin Sure, will do that, thanks @thekiiingbob could you please review again @saturninoabril could you please review it again @composednitin And just sync with master please. :)
2025-04-01T06:39:31.174585
2019-08-05T16:39:50
476955622
{ "authors": [ "lindalumitchell", "mgdelacroix" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8131", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/3366" }
gharchive/pull-request
[MM-17200] Add render methods for new guest messages Ticket Link https://mattermost.atlassian.net/browse/MM-17200 Related Pull Requests Has server changes Has redux changes The ticket only refers to the system post that appears when a guest is added to a channel by another user, but these PRs add a new message to the system post that appears when a guest joins a channel (for example, through an invitation). @wiersgallak can you please confirm this second change? Checked that the test failure is caused by mattermost-redux not pointing to the right commit, should be fixed when the redux PR is merged. Thanks @wiersgallak! Updated Re-assigned QA review to Rohitesh, as he and Ogi are QA point people on guest accounts. (Sorry; closed with a bad tab/Enter combo and re-opened. :) )
2025-04-01T06:39:31.177992
2021-03-05T13:51:35
823111439
{ "authors": [ "ethervoid", "marianunez", "mattermod", "pablovelezvidal" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8132", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/7645" }
gharchive/pull-request
MM32642 - increase invite others visibility Summary MM32642 - increase invite others visibility - These PR adds enhancements to the channels intros to improve the user experience and add more visibility to the call to action for inviting new members to workspaces. Ticket Link https://mattermost.atlassian.net/browse/MM-32642 Related Pull Requests MM-Redux: https://github.com/mattermost/mattermost-redux/pull/1395 Screenshots @pablovelezvidal @ethervoid I believe this was not included in the cloud branch either. Could you confirm ? If so, this needs to be cherry picked to be included in the March 24th release. /cherry-pick cloud Cherry pick is scheduled.
2025-04-01T06:39:31.180860
2021-07-22T19:45:58
950993482
{ "authors": [ "mattermod", "neallred" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8133", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/8451" }
gharchive/pull-request
fix: Subscription title in system console billing is singular Summary Fixes a typo. Subscription title should be singular in the System Console billing page for cloud accounts. Ticket Link https://mattermost.atlassian.net/browse/MM-30538 Screenshots Title now reads "Subscription" Release Note Fixed System Console billing subscription title to be singular Hello @neallred, Thanks for your pull request! A Core Committer will review your pull request soon. For code contributions, you can learn more about the review process here.
2025-04-01T06:39:31.182953
2022-02-08T16:03:50
1127452252
{ "authors": [ "michelengelen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8134", "repo": "mattermost/mattermost-webapp", "url": "https://github.com/mattermost/mattermost-webapp/pull/9777" }
gharchive/pull-request
hide accordion items with status 'OK' Summary title says it all. Items that have status 'ok' or 'none' get hidden from the AccordionCard. They still get counted in the overall score. Ticket Link n/a Related Pull Requests n/a Screenshots n/a Release Note NONE @BenCookie95 gentle ping on this one.
2025-04-01T06:39:31.258264
2018-01-22T18:10:31
290562749
{ "authors": [ "benron", "slizeray" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8135", "repo": "matthewpalmer/Locksmith", "url": "https://github.com/matthewpalmer/Locksmith/issues/186" }
gharchive/issue
Is there any way to get access to the error that's swallowed in readFromSecureStore? https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L495 https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L498 https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L484 https://github.com/matthewpalmer/Locksmith/blob/49b08bf8fa16ff8d7ce41609a2adbc6b5237f65f/Source/Locksmith.swift#L481
2025-04-01T06:39:31.310224
2022-10-05T19:36:44
1398278270
{ "authors": [ "Obelix74Git", "alekslyse", "laithsaid", "madradrakie", "matthijsberk", "matthijsvisser" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8136", "repo": "matthijsvisser/kamstrup-402-mqtt", "url": "https://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7" }
gharchive/issue
Cannot extract meter value - Kamstrup Multical 401 I hope you vcan help me. I have a Kamstrup Multical 401 and trying to use your script. I have tried to place de IR head at every possible way at the meter but the following warning keeps appearing. In MQTT i see the topic but no values. [2022-10-05 21:27:18 daemon.py init: 41 - INFO - initializing daemon] [2022-10-05 21:27:18 mqtt_handler.py connect: 43 - INFO - Connected to MQTT at: <IP_ADDRESS>:1883] [2022-10-05 21:27:18 mqtt_handler.py connect: 45 - INFO - with username pi, QoS level = 1 and retain = False] [2022-10-05 21:27:18 kamstrup_meter.py close: 126 - DEBUG - Closed serial port] [2022-10-05 21:27:18 kamstrup_meter.py open: 118 - DEBUG - Opened serial port] [2022-10-05 21:27:20 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:20 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:22 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:22 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:24 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:24 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:26 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:26 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:26 kamstrup_meter.py close: 126 - DEBUG - Closed serial port] [2022-10-05 21:27:26 mqtt_handler.py publish: 55 - INFO - Publishing 'kamstrup/values' '{}' to <IP_ADDRESS>:1883] [2022-10-05 21:27:26 daemon.py run: 84 - INFO - Waiting 1 minute(s) for the next meter readout] This is my configuration file: mqtt: host: <IP_ADDRESS> port: 1883 client: kamstrup402 topic: kamstrup qos: 1 retain: False authentication: True username: pi password: [password] serial_device: com_port: /dev/serial/by-id/usb-Silicon_Labs_CP2102_USB_to_UART_Bridge_Controller_0001-if00-port0 kamstrup: parameters: - energy - volume - temp1 - temp2 poll_interval: 1 i hope anyone has an idea what i am doing wrong. Thanks in advance. Hi, Are you sure that your meter is "awake"? You can do so by pressing the button. It will remain active for 30 minutes and remains active if a valid IR signal is received. From: madradrakie @.> Sent: Wednesday, October 5, 2022 9:36:54 PM To: matthijsvisser/kamstrup-402-mqtt @.> Cc: Subscribed @.***> Subject: [matthijsvisser/kamstrup-402-mqtt] Cannot extract meter value - Kamstrup Multical 401 (Issue #7) I hope you vcan help me. I have a Kamstrup Multical 401 and trying to use your script. I have tried to place de IR head at every possible way at the meter but the following warning keeps appearing. In MQTT i see the topic but no values. [2022-10-05 21:27:18 daemon.py init: 41 - INFO - initializing daemon] [2022-10-05 21:27:18 mqtt_handler.py connect: 43 - INFO - Connected to MQTT at: <IP_ADDRESS>:1883] [2022-10-05 21:27:18 mqtt_handler.py connect: 45 - INFO - with username pi, QoS level = 1 and retain = False] [2022-10-05 21:27:18 kamstrup_meter.py close: 126 - DEBUG - Closed serial port] [2022-10-05 21:27:18 kamstrup_meter.py open: 118 - DEBUG - Opened serial port] [2022-10-05 21:27:20 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:20 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:22 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:22 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:24 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:24 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:26 kamstrup_meter.py rd: 131 - DEBUG - Rx timeout] [2022-10-05 21:27:26 kamstrup_meter.py readparameter: 198 - WARNING - No response from meter] [2022-10-05 21:27:26 kamstrup_meter.py close: 126 - DEBUG - Closed serial port] [2022-10-05 21:27:26 mqtt_handler.py publish: 55 - INFO - Publishing 'kamstrup/values' '{}' to <IP_ADDRESS>:1883] [2022-10-05 21:27:26 daemon.py run: 84 - INFO - Waiting 1 minute(s) for the next meter readout] This is my configuration file: mqtt: host: <IP_ADDRESS> port: 1883 client: kamstrup402 topic: kamstrup qos: 1 retain: False authentication: True username: pi password: a1connect! serial_device: com_port: /dev/serial/by-id/usb-Silicon_Labs_CP2102_USB_to_UART_Bridge_Controller_0001-if00-port0 kamstrup: parameters: energy volume temp1 temp2 poll_interval: 1 i hope anyone has an idea what i am doing wrong. Thanks in advance. — Reply to this email directly, view it on GitHubhttps://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AK4J3SUCARZUZNJO7PEGTJLWBXKFNANCNFSM6AAAAAAQ53QGB4. You are receiving this because you are subscribed to this thread.Message ID: @.***> hi matthijs. before i activate the script i press the button on the meter. i assume the 401 is working the same as the 402. i have tried for an hour to find the sweet spot but still no luck. it is also very hard to determine if the IR leds of the head are alligned perfectly You can try the official software from kamstrup to verify if the meter will respond to those commands See https://www.kamstrup.com/nl-nl/warmteoplossingen/slimme-warmtemeters/meter-software You can try the official software from kamstrup to verify if the meter will respons to those commands See https://www.kamstrup.com/nl-nl/warmteoplossingen/slimme-warmtemeters/meter-software Op wo 5 okt. 2022 21:45 schreef madradrakie @.***>: hi matthijs. before i activate the script i press the button on the meter. i assume the 401 is working the same as the 402. i have tried for an hour to find the sweet spot but still no luck. it is also very hard to determine if the IR leds of the head are alligned perfectly — Reply to this email directly, view it on GitHub https://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7#issuecomment-1268885487, or unsubscribe https://github.com/notifications/unsubscribe-auth/AK4J3SV2ORAOBV5QSAMRRMTWBXLGLANCNFSM6AAAAAAQ53QGB4 . You are receiving this because you commented.Message ID: @.***> i have tried the official software but my meter, the kamstrup 401 is not listed. It starts with the 402. I'am really curious if someone has managed to get it working because thereis a sensor eye present on the 401. I did have the same problem, but after reading the manual I found out that the communication is different between a M402 and M401. At the moment I can read the values from the meter (m401), but I'm still looking into the possibility to send it via Mqtt (no great coder myself). Maybe you can adapt (USB port) and try this bit of code if your IR is placed correctly and is working?: ` import serial from time import sleep mc401 = serial.Serial(port='/dev/ttyUSB1', bytesize=serial.SEVENBITS, parity=serial.PARITY_EVEN, stopbits=serial.STOPBITS_ONE, timeout=2) mc401.baudrate = 300 mc401.write(bytes("/#1", 'UTF-8')) mc401.flush() sleep(1) mc401.baudrate = 1200 mc401.flushInput() data = mc401.read(87).split() try: print(data[0], data[1], data[3], data[4], data[5], data[6], data[7]) except IndexError: pass ` The main problem that I encountered is that you must send a signal at 300 baudrate, but you can only receive at 1200 baudrate. I have used the signal /#1 as this gives back energy, volume, temp1, temp2, tempdiff. hello matthijs. Nico to read that you got it working. Can you tell me what you mean by adapt and at which place i should place the code? The IR reader is connected with a raspberry pi. I am still learning everyday :-) I have same issue with MutliCAL 602. It should be using same protocol "KMP" but it is just echoing the same values I'm sending to it. Any idea? Hi, I don't have the 602, so unfortunately I cant help you with that. But here are my thoughts: If you receive the exact same message as the code sends, than its possible that you having issues with reflections. You can try to put some paper between the IR head and the meter to diffuse the light a little. Or try some more positions with the meter head. When I was testing the code, I experienced that the official tooling was able to handles more 'noise' than my code. Thanks for the info. I actually noticed that it is echoing the data even if it is not sitting on the kamstrup meter? Should the script wait little before reading the data or maybe flush before reading 🤔Laith…On 7 Dec 2022, at 21.46, Matthijs @.***> wrote: Hi, I don't have the 602, so unfortunately I cant help you with that. But here are my thoughts: If you receive the exact same message as the code sends, than its possible that you having issues with reflections. You can try to put some paper between the IR head and the meter to diffuse the light a little. Or try some more positions with the meter head. When I was testing the code, I experienced that the official tooling was able to handles more 'noise' than my code. —Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you commented.Message ID: @.***> I have a similar problem. We have a Kamstrup MultiCAL 66C. According to the specifications, it works with: 300 baud req /300 baud reply, 1 start bit, 7 data bits, equal parity, 2 stop bits. How can I change this because if I change this in kamstrup_meter.py to baud rate = 300, parity = serial.PARITY_EVEN, bytesize = serial.SEVENBITS, I get this error within 4 seconds : Traceback (most recent call last): File "/tmp/test/kamstrup-402-mqtt/daemon.py", line 92, in main() File "/tmp/test/kamstrup-402-mqtt/daemon.py", line 89, in main daemon.run() File "/tmp/test/kamstrup-402-mqtt/daemon.py", line 81, in run values ​​= self.heat_meter.run() File "/tmp/test/kamstrup-402-mqtt/kamstrup_meter.py", line 114, in run value = self.readparameter(int(str(kamstrup_402_params[parameter]), 0)) File "/tmp/test/kamstrup-402-mqtt/kamstrup_meter.py", line 205, in read parameter elif ((receivedMessage[0] != 0x3f) or IndexError: byte array index out of range What do I need to change to make the communication work properly? It looks like you don't receive any data from the meter since the received Message array is empty. Sorry verkeerde error. This is the correct error. Traceback (most recent call last): File "/opt/test/kamstrup-402-mqtt/daemon.py", line 92, in main() File "/opt/test/kamstrup-402-mqtt/daemon.py", line 88, in main daemon = KamstrupDaemon() File "/opt/test/kamstrup-402-mqtt/daemon.py", line 69, in init self.heat_meter = kamstrup(serial_cfg["com_port"], kamstrup_cfg["parameters"]) File "/opt/test/kamstrup-402-mqtt/kamstrup_meter.py", line 97, in init bytesize = serial.SEVENTBITS, AttributeError: module 'serial' has no attribute 'SEVENTBITS' So if no data comes in, the application gives an error? You are using an invalid parameter name. See https://pyserial.readthedocs.io/en/latest/pyserial_api.html It should be serial.sevenbits. From: Obelix74Git @.> Sent: Friday, December 16, 2022 10:13:47 PM To: matthijsvisser/kamstrup-402-mqtt @.> Cc: Matthijs @.>; Comment @.> Subject: Re: [matthijsvisser/kamstrup-402-mqtt] Cannot extract meter value - Kamstrup Multical 401 (Issue #7) Sorry verkeerde error. This is the correct error. Traceback (most recent call last): File "/opt/test/kamstrup-402-mqtt/daemon.py", line 92, in main() File "/opt/test/kamstrup-402-mqtt/daemon.py", line 88, in main daemon = KamstrupDaemon() File "/opt/test/kamstrup-402-mqtt/daemon.py", line 69, in init self.heat_meter = kamstrup(serial_cfg["com_port"], kamstrup_cfg["parameters"]) File "/opt/test/kamstrup-402-mqtt/kamstrup_meter.py", line 97, in init bytesize = serial.SEVENTBITS, AttributeError: module 'serial' has no attribute 'SEVENTBITS' — Reply to this email directly, view it on GitHubhttps://github.com/matthijsvisser/kamstrup-402-mqtt/issues/7#issuecomment-1355617360, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AK4J3SUNSEYUU66QVUD3IFDWNTLQXANCNFSM6AAAAAAQ53QGB4. I have a 401, and put the meter over the reader at the bottom right. It seems to be seated properly, and tried all kinds of positions, but no reading. I also clicked the enter/play button on the reader. Did you guys get this working? I have a 401, and put the meter over the reader at the bottom right. It seems to be seated properly, and tried all kinds of positions, but no reading. I also clicked the enter/play button on the reader. For info I am using this reader, as it was suggested by someone. It should be working I think? https://www.aliexpress.com/item/1005003509520122.html Hello Alekslyse. I got it working with for the 401 meter. I used the following code: https://github.com/matthijsberk/M401-irreader See for more information: https://github.com/matthijsberk/M401-irreader/issues/2
2025-04-01T06:39:31.323660
2015-07-21T15:29:25
96345994
{ "authors": [ "L1ghtn1ng", "PowerShellMafia", "enigma0x3", "secabstraction" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8137", "repo": "mattifestation/PowerSploit", "url": "https://github.com/mattifestation/PowerSploit/issues/70" }
gharchive/issue
x64 Detection using Get-ProcAddress Within Invoke--Shellcode, things often fail on x86 systems. After some digging, it appears that $64BitCPU returns true; even on x86 systems. Since $64BitCPU appears to always return true, I followed it down to see how $64BitCPU is being determined. To test this, I copied Get-Win32Functions, Get-ProcAddress and Get-DelegatedType into a separate script. I then added the logic within Invoke--Shellcode "$IsWow64ProcessAddr = Get-ProcAddress kernel32.dll IsWow64Process". When running this script, it always returns a value on x86, instead of null. Here is the script: function local:Get-Win32Functions { $Win32Functions = New-Object System.Object $OpenProcessAddr = Get-ProcAddress kernel32.dll OpenProcess $OpenProcessDelegate = Get-DelegateType @([UInt32], [Bool], [UInt32]) ([IntPtr]) $OpenProcess = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($OpenProcessAddr, $OpenProcessDelegate) $Win32Functions | Add-Member NoteProperty -Name OpenProcess -Value $OpenProcess $VirtualAllocExAddr = Get-ProcAddress kernel32.dll VirtualAllocEx $VirtualAllocExDelegate = Get-DelegateType @([IntPtr], [IntPtr], [Uint32], [UInt32], [UInt32]) ([IntPtr]) $VirtualAllocEx = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($VirtualAllocExAddr, $VirtualAllocExDelegate) $Win32Functions | Add-Member NoteProperty -Name VirtualAllocEx -Value $VirtualAllocEx $WriteProcessMemoryAddr = Get-ProcAddress kernel32.dll WriteProcessMemory $WriteProcessMemoryDelegate = Get-DelegateType @([IntPtr], [IntPtr], [Byte[]], [UInt32], [UInt32].MakeByRefType()) ([Bool]) $WriteProcessMemory = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($WriteProcessMemoryAddr, $WriteProcessMemoryDelegate) $Win32Functions | Add-Member NoteProperty -Name WriteProcessMemory -Value $WriteProcessMemory $CreateRemoteThreadAddr = Get-ProcAddress kernel32.dll CreateRemoteThread $CreateRemoteThreadDelegate = Get-DelegateType @([IntPtr], [IntPtr], [UInt32], [IntPtr], [IntPtr], [UInt32], [IntPtr]) ([IntPtr]) $CreateRemoteThread = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($CreateRemoteThreadAddr, $CreateRemoteThreadDelegate) $Win32Functions | Add-Member NoteProperty -Name CreateRemoteThread -Value $CreateRemoteThread $WaitForSingleObjectAddr = Get-ProcAddress kernel32.dll WaitForSingleObject $WaitForSingleObjectDelegate = Get-DelegateType @([IntPtr], [UInt32]) $WaitForSingleObject = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($WaitForSingleObjectAddr, $WaitForSingleObjectDelegate) $Win32Functions | Add-Member NoteProperty -Name WaitForSingleObject -Value $WaitForSingleObject $CloseHandleAddr = Get-ProcAddress kernel32.dll CloseHandle $CloseHandleDelegate = Get-DelegateType @([IntPtr]) ([Bool]) $CloseHandle = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($CloseHandleAddr, $CloseHandleDelegate) $Win32Functions | Add-Member NoteProperty -Name CloseHandle -Value $CloseHandle $GetLastErrorAddr = Get-ProcAddress kernel32.dll GetLastError $GetLastErrorDelegate = Get-DelegateType @() ([Uint32]) $GetLastError = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($GetLastErrorAddr, $GetLastErrorDelegate) $Win32Functions | Add-Member NoteProperty -Name GetLastError -Value $GetLastError $NtCreateThreadExAddr = Get-ProcAddress NtDll.dll NtCreateThreadEx $NtCreateThreadExDelegate = Get-DelegateType @([IntPtr].MakeByRefType(), [UInt32], [IntPtr], [IntPtr], [IntPtr], [IntPtr], [Bool], [UInt32], [UInt32], [UInt32], [IntPtr]) ([UInt32]) $NtCreateThreadEx = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($NtCreateThreadExAddr, $NtCreateThreadExDelegate) $Win32Functions | Add-Member -MemberType NoteProperty -Name NtCreateThreadEx -Value $NtCreateThreadEx # A valid pointer to IsWow64Process will be returned if CPU is 64-bit $IsWow64ProcessAddr = Get-ProcAddress kernel32.dll IsWow64Process if ($IsWow64ProcessAddr) { $IsWow64ProcessDelegate = Get-DelegateType @([IntPtr], [Bool].MakeByRefType()) ([Bool]) $IsWow64Process = [System.Runtime.InteropServices.Marshal]::GetDelegateForFunctionPointer($IsWow64ProcessAddr, $IsWow64ProcessDelegate) $Win32Functions | Add-Member NoteProperty -Name IsWow64Process -Value $IsWow64Process } return $Win32Functions } function Local:Get-ProcAddress { Param ( [OutputType([IntPtr])] [Parameter( Position = 0, Mandatory = $True )] [String] $Module, [Parameter( Position = 1, Mandatory = $True )] [String] $Procedure ) # Get a reference to System.dll in the GAC $SystemAssembly = [AppDomain]::CurrentDomain.GetAssemblies() | Where-Object { $_.GlobalAssemblyCache -And $_.Location.Split('\\')[-1].Equals('System.dll') } $UnsafeNativeMethods = $SystemAssembly.GetType('Microsoft.Win32.UnsafeNativeMethods') # Get a reference to the GetModuleHandle and GetProcAddress methods $GetModuleHandle = $UnsafeNativeMethods.GetMethod('GetModuleHandle') $GetProcAddress = $UnsafeNativeMethods.GetMethod('GetProcAddress') # Get a handle to the module specified $Kern32Handle = $GetModuleHandle.Invoke($null, @($Module)) $tmpPtr = New-Object IntPtr $HandleRef = New-Object System.Runtime.InteropServices.HandleRef($tmpPtr, $Kern32Handle) # Return the address of the function Write-Output $GetProcAddress.Invoke($null, @([System.Runtime.InteropServices.HandleRef]$HandleRef, $Procedure)) } function Local:Get-DelegateType { Param ( [OutputType([Type])] [Parameter( Position = 0)] [Type[]] $Parameters = (New-Object Type[](0)), [Parameter( Position = 1 )] [Type] $ReturnType = [Void] ) $Domain = [AppDomain]::CurrentDomain $DynAssembly = New-Object System.Reflection.AssemblyName('ReflectedDelegate') $AssemblyBuilder = $Domain.DefineDynamicAssembly($DynAssembly, [System.Reflection.Emit.AssemblyBuilderAccess]::Run) $ModuleBuilder = $AssemblyBuilder.DefineDynamicModule('InMemoryModule', $false) $TypeBuilder = $ModuleBuilder.DefineType('MyDelegateType', 'Class, Public, Sealed, AnsiClass, AutoClass', [System.MulticastDelegate]) $ConstructorBuilder = $TypeBuilder.DefineConstructor('RTSpecialName, HideBySig, Public', [System.Reflection.CallingConventions]::Standard, $Parameters) $ConstructorBuilder.SetImplementationFlags('Runtime, Managed') $MethodBuilder = $TypeBuilder.DefineMethod('Invoke', 'Public, HideBySig, NewSlot, Virtual', $ReturnType, $Parameters) $MethodBuilder.SetImplementationFlags('Runtime, Managed') Write-Output $TypeBuilder.CreateType() } $var = Get-ProcAddress kernel32.dll IsWow64Process $var Here is the output: From my understanding, this should return $null on x86 systems. I'm not sure if this has something to do with the fact that it is a VM. My test environment is a full patched win7 x86 box running on VMware fusion with a Mac host. Let me know if clarification is needed. Thanks! Matt N. sounds like a pull-request for Get-SystemInfo or { $processor = gwmi win32_processor; $processor.architecture } My system info powershell script has some logic you could use for this which could be found https://github.com/L1ghtn1ng/powershell-scripts/blob/master/systemInformation.ps1 @L1ghtn1ng because your logic is based on $env variables, the output will depend on which version of powershell you're running using. e.g. x86 powershell returns "32-bit" on a x64 machine. Fair enough am still new to powershell and IIRC set has some logic and so does metasploit Thanks for all the input, folks! I fixed the bug using the AddressWidth field of the Win32_Processor class while being mindful of multiple instances being returned as mentioned in issue #75. Pester tests passed when tested in Win 10 (64-bit) and Windows XP (32-bit).
2025-04-01T06:39:31.327501
2015-04-13T05:42:00
68012773
{ "authors": [ "mathDR", "mattjj" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8138", "repo": "mattjj/pyhsmm", "url": "https://github.com/mattjj/pyhsmm/issues/44" }
gharchive/issue
Multimodal emission probabilites? Did you ever extend this package to include HDP priors for Gaussian Mixtures for the emission probabilities? I am trying to locate the acceptable probabilities for emission distributions, but can only find Gaussian. Thanks The distributions are mostly in pybasicbayes's distributions.py. To make a mixture that can be plugged into an HMM, you'd also want to use the MixtureDistribution class. Something like this: from pybasicbayes.models import MixtureDistribution from pybasicbayes.distributions import Gaussian obs_distns = \ [MixtureDistribution( alpha_0=5., components=[Gaussian(**obs_hypparams) for _ in xrange(5)]) for state in range(Nmax)] By default those MixtureDistributions act like weak limit approximations to Dirichlet Process mixture models. If you want to make an HDP model (including multiple mixture models) you'll have to write the hierarchical prat yourself. You can make mixtures of any of the distributions in pybasicbayes (and it's pretty easy to add new ones).
2025-04-01T06:39:31.333573
2016-06-24T23:08:52
162246706
{ "authors": [ "coveralls", "mattkrick" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8139", "repo": "mattkrick/cashay", "url": "https://github.com/mattkrick/cashay/pull/73" }
gharchive/pull-request
SSR fixes #53 fixes #64 Coverage remained the same at 87.179% when pulling 27cdadab7d22c8078ff2e80daf7d8a31800006f2 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master. Coverage remained the same at 87.179% when pulling 27cdadab7d22c8078ff2e80daf7d8a31800006f2 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master. Coverage remained the same at 87.179% when pulling 3994f331702497e952e6197853ac678bce1bad88 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master. Coverage remained the same at 87.179% when pulling 85e1a8f8a13d49b08479144c3169f3a27d1ef085 on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master. Coverage remained the same at 87.179% when pulling eecfe566ea81510585e15fad91e85eb4c56d627c on ssr into 75c2202fac7d15fcfc2b86d134386f6b7b3e030f on master.
2025-04-01T06:39:31.356293
2020-02-03T21:36:42
559349472
{ "authors": [ "lukinma", "rittneje" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8140", "repo": "mattn/go-sqlite3", "url": "https://github.com/mattn/go-sqlite3/issues/780" }
gharchive/issue
Select returns wrong value when SQL type is DATETIME I use the following code: In this code I insert into a database date 2 Jan 2020 and insert works correct as shows SQLite DB Browser. After that I read data from database using SELECT and expect to get 2 Jan 2020. But I get 1 Jan 0001. package main import ( "database/sql" "fmt" _ "github.com/mattn/go-sqlite3" "os" "time" ) func InitDummyDB(path string) *sql.DB { db, err := sql.Open("sqlite3", path) if err != nil { panic(err) } if db == nil { panic("db is nil") } return db } func CreateDummyTable(db *sql.DB) { sqlTable := ` CREATE TABLE IF NOT EXISTS items ( ID INTEGER PRIMARY KEY, TimeT DATETIME); ` res, err := db.Exec(sqlTable) if err != nil { panic(err) } fmt.Println(res) } func InsertDummyDate(db *sql.DB) { t1 := time.Date(2020, 1, 2, 3, 0, 0, 0, time.UTC) request, err := db.Prepare(`INSERT INTO items (TimeT) VALUES (?)`) if err != nil { panic(err) } _, rerr := request.Exec(t1.String()) if rerr != nil { panic(rerr) } } func PrintDummyData(db *sql.DB) { rows, err := db.Query(`SELECT ID, TimeT FROM items ORDER BY ID`) if err != nil { panic(err) } defer rows.Close() for rows.Next() { var id int var readTime time.Time // 0001-01-01 00:00:00 +0000 UTC will be returned //var readTime string // string does not help: we get 0001-01-01T00:00:00Z err = rows.Scan(&id, &readTime) if err != nil { panic(err) } fmt.Println(id, readTime) } } func main() { remErr := os.Remove("dummy.db") if remErr != nil { fmt.Println(remErr) } db := InitDummyDB("dummy.db") CreateDummyTable(db) InsertDummyDate(db) PrintDummyData(db) } See https://www.sqlite.org/lang_datefunc.html When you call time.String(), the output is "2020-01-02 03:00:00 +0000 UTC", which is not one of the supported formats. But if you change it to just request.Exec(t1), this library takes care of it so the resulting read works as expected. (The problem here is that SQLite itself doesn't really support timestamps as a proper datatype. This library does a best effort to deal with them, but as you can see, it isn't perfect.) It works now, thank you!
2025-04-01T06:39:31.361268
2020-01-07T11:38:39
546236363
{ "authors": [ "AshfordN", "coveralls", "rittneje" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8141", "repo": "mattn/go-sqlite3", "url": "https://github.com/mattn/go-sqlite3/pull/772" }
gharchive/pull-request
added missing mode functionality the mode option in the dsn did not work as expected, particularly when trying to open a database in read-only mode @AshfordN Do you have an example of it not working correctly? The mode parameter is handled by SQLite itself and is basically applied on top of SQLITE_OPEN_READWRITE/SQLITE_OPEN_READONLY/SQLITE_OPEN_CREATE. Therefore, this change is not needed. Well the database isn't actually read-only when you specify mode=ro Coverage increased (+0.2%) to 51.639% when pulling 76ce3bd1fcd6db5f980a2403bdfd5be7755bcf43 on AshfordN:patch-1 into 53cff3fcebd1e177d04129dc131523635bc45d3d on mattn:master. Please provide an example code segment where it does not work as expected. The following code successfully inserts and updates a record in the database, using go1.13.1 on linux/amd64: package main import ( "database/sql" "fmt" _ "github.com/mattn/go-sqlite3" ) func main() { //open database db, err := sql.Open("sqlite3", "main.db?mode=ro") if err != nil { panic(err) } //attempt insert if _, err := db.Exec("INSERT INTO test(col1, col2) VALUES(1, 2)"); err != nil { fmt.Printf("Insert failed: %w\n", err) } //attempt update if _, err := db.Exec("UPDATE test SET col2 = 4 WHERE col1 = 1"); err != nil { fmt.Printf("Update failed: %w\n", err) } } See https://www.sqlite.org/uri.html. It should be sql.Open("sqlite3", "file:main.db?mode=ro"). Ok, this works, thanks.
2025-04-01T06:39:31.369946
2021-04-14T15:15:28
858000635
{ "authors": [ "aeneasr", "codecov-io", "mattn", "neilalexander" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8142", "repo": "mattn/go-sqlite3", "url": "https://github.com/mattn/go-sqlite3/pull/940" }
gharchive/pull-request
Update amalgamation code Fixes #926, #927 Codecov Report Merging #940 (d33341f) into master (ab91e93) will decrease coverage by 0.13%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #940 +/- ## ========================================== - Coverage 46.80% 46.67% -0.14% ========================================== Files 11 11 Lines 1457 1457 ========================================== - Hits 682 680 -2 - Misses 640 641 +1 - Partials 135 136 +1 Impacted Files Coverage Δ sqlite3.go 52.84% <0.00%> (-0.23%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update ab91e93...d33341f. Read the comment docs. Test fail on Windows since installing go-acc failed. But I tested this PR on Windows. So merging. Thank you so much for the fast turnaround, really appreciate it! 🥳 @mattn thank you so much for the quick turnaround - author of go-acc here also (by chance). I can try and look into the Windows issue - specifically by adding a windows VM to the CI :)
2025-04-01T06:39:31.377887
2022-01-23T20:36:17
1111979087
{ "authors": [ "Cringely", "mattsemar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8143", "repo": "mattsemar/dsp-personal-logistics", "url": "https://github.com/mattsemar/dsp-personal-logistics/issues/20" }
gharchive/issue
Status text placement Describe the bug Status text is off screen. This seems to have started immediately after following the most recent major game update. To Reproduce Steps to reproduce the behavior: Install mod with mod manager Run game Expected behavior Readable status text Screenshots Mods installed jaYzZWjA0a Extra Unable to upload save and mod save data after trying to ZIP and upload. Out of curiosity, what resolution are you using? Also is the "UI layout resolution reference height" (in Settings->Gameplay) different than the resolution's height? This is 2560x1440p with a reference height of 900. This worked just fine prior to update, however with that being said I did also forget my reference height was so low (I enjoy the larger UI elements). For now I've set the reference height to 1080 and everything is correctly visible again. I'd still appreciate using 900 but also don't mind just using 1080 as a workaround for this. Thanks for your time. Feel free to close this, due to this reference heigh workaround :) I'm definitely going to fix it, I'm just compiling a list of resolutions and ref heights to test before I release versions. I think this should be fixed with the release from yesterday Confirming this appears to be resolved. Thanks for your support!
2025-04-01T06:39:31.393630
2024-04-16T21:33:35
2246917989
{ "authors": [ "MARK2580", "Mayki07", "matyalatte" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8144", "repo": "matyalatte/UE4-DDS-Tools", "url": "https://github.com/matyalatte/UE4-DDS-Tools/issues/30" }
gharchive/issue
Virtual textures are not supported Hello. Here are some textures from the game The day before. Engine version UE5 <IP_ADDRESS> I was never able to get anything adequate from these files, even through rawtex / texconv thedaybefore_assets.zip Hi, thanks for the report. Do you have more infomation about your assets? T_drum_kit_01_d.uasset says it's a 2d texture. and T_drum_kit_01_d.uexp says it's a 1024x1024 texture that have DXT1 format. Its DDS data should be 683KB at most but T_drum_kit_01_d.ubulk is 830KB. (Usually, the texture ubulk has only pixel data for DDS.) Also, I can see PF_DXT1 twice in T_drum_kit_01_d.uexp. I've never seen like this before. I assume that the day before uses a custom version of UE or a less well-known option. no, unfortunately these files, the version of the game and what kind of game this is is the only thing I can give you, I’m very far from all this stuff with coding) if you need more examples... by the way, the game also contains ordinary textures that umodel was able to immediately convert to a readable format. I confirmed that they are streaming virtual textures. I got similar structures when I enabled the Virtual Texture Streaming option in Unreal Engine. Presumably, their pixel data is subdivided into small pieces and stored in specific data structures. But unfortunately, it's beyond my understanding. Idk how to convert them to the standard image formats. So, I added an error message for virtual textures at 3d6ac1c2708fd9680d809740c628409dd904c717, and gave up to support your assets. It’s a pity, I’m increasingly encountering these virtual textures, even though they were originally just an ordinary picture =( I apologize in advance for a question that doesn't quite belong here, but could you give me some advice on how to properly set up the virtual texture generation in the unreal editor? I've exported the texture in png via Umodel, edited it in Photoshop and would like to create a uasset back from it directly in the editor. But I can't seem to set it up properly. I always end up with a black texture instead of a picture.
2025-04-01T06:39:31.403221
2023-01-09T15:22:39
1525807536
{ "authors": [ "Get-SomeCofee", "chrisecklar", "jacqtscpl" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8146", "repo": "maurice-daly/DriverAutomationTool", "url": "https://github.com/maurice-daly/DriverAutomationTool/issues/444" }
gharchive/issue
PowerShell Error. You cannot call a method on a null-valued expression. At line:17154 char:5 Trying to get started using the tool. I'm getting this error when running DriverAutomationTool.exe Any idea how I can get past this? I could not find the PS1 file to look at the PowerShell code to troubleshoot this further. Also Microsoft is checked and grayed out and can't be unchecked. [Start Up] - Initialising Driver Automation Tool - Loading required PowerShell modules - Checking for preinstalled PS modules - Importing module BitsTransfer [Launching GUI] [Azure Registy Settings Detected] - Obtaining all Azure values from registry path HKLM:\SOFTWARE\MSEndpointMgr\DriverAutomationTool - Setting values in the UI You cannot call a method on a null-valued expression. At line:17154 char:5 + ... $OSBuild = ($WindowsBuildHashTable.Item("$OSBuild")).Spli ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : InvalidOperation: (:) [], RuntimeException + FullyQualifiedErrorId : InvokeMethodOnNull Try clicking the Reset Tool button at the bottom left and setting only the settings needed. The reset button worked for me. Thanks for the tip OK i have done the reset... now i am getting... Start-BitsTransfer : The server name or address could not be resolved At line:1093 char:8 ... Start-BitsTransfer -Source $DellXMLCabinetSource -Destina ... ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ CategoryInfo : InvalidOperation: (:) [Start-BitsTransfer], Exception FullyQualifiedErrorId : StartBitsTransferCOMException,Microsoft.BackgroundIntelligentTransfer.Management.NewBitsTransferCommand
2025-04-01T06:39:31.409251
2015-02-25T22:58:35
58984513
{ "authors": [ "TimPigden", "mauricio", "webgress" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8147", "repo": "mauricio/postgresql-async", "url": "https://github.com/mauricio/postgresql-async/issues/134" }
gharchive/issue
jsonb support to get read/write jsonb columns to/from play json Starting a new project and jsonb looks very attractive. Any pointers on how to deal with jsonb columns ? Right now there is no support for jsonb as including support would require building a jsonb serializer and de-serializer, but we would definitely accept a PR that introduces it :) I'd be happy to, even if I have to hire somebody to do it :) So two questions: 1 - Is reading/writing jsonb currently possible at all ? 2 - How would I approach "building a jsonb serializer" ? I am not sure how good my scala is, but postgresql with jsonb and non-blocking play is what I want to use for this project. Right now it will be read as string, not sure how PostgreSQL will produce a string out of it. To include it you need to build an encoder and a decoder and register it [here](https://github.com/mauricio/postgresql-async/blob/master/postgresql-async/src/main/scala/com/github/mauricio/async/db/postgresql/column/PostgreSQLColumnEncoderRegistry.scala] and here. Great, already something ! So at least I should be able to use play json support to parse/format the strings. Thank you for a quick reply. Hi @webgress - did you do anything about this? I too want jsonb ok - So if you convert your json into a string before passing it across to posgres and if you parse it when it comes back there's nothing to do (which mauicio more or less said above). It works already Thinking about it, there are many different json ASTs in the scala world. Any decision to write a decoder as part of this package would probably be a mistake - as you'd get in the way of the people with a different Json representation. So I'd suggest this doesn't need fixing. Yeah, which is kind of what I'm thinking as well, we could include a way for you to provide your own JSON parser if you'd like to if people think it's necessary, but I'd rather not force people to use something, specially when there are many different solutions out there already and they aren't even remotely compatible between them.
2025-04-01T06:39:31.446234
2019-03-06T21:10:10
418010229
{ "authors": [ "himat", "petrleocompel", "tankers746" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8148", "repo": "mauron85/background-geolocation-android", "url": "https://github.com/mauron85/background-geolocation-android/issues/29" }
gharchive/issue
Unable to use library with commons-io My project has dependency on commons-io wich conflicts with your placed classes in -> src/main/java/org/apache/commons/io/ You probably should use dependency commons-io:commons-io:2.6. Add to VERSIONS.gradle after line 132 commonsio: [group: 'commons-io', name:'commons-io', version: '2.6'], To build.gradle in array oreoDependencies [configuration: "implementation", dependency: libs.commonsio], And to array preoreoDependencies [configuration: "compile", dependency: libs.commonsio], Have made a PR for this: https://github.com/mauron85/background-geolocation-android/pull/31 @mauron85 would it be possible to get this merged? Its interfering with expo's unimodules. Almost year - still waiting... Next problem is going to be release of older version of mauron85/react-native-background-geolocation Using version 0.5 @petrleocompel Just wanted to say thank you for posting this fix. I already spent so many hours trying to get the library to compile in release mode, and your PR finally helped me get it to work. Thank you for being here for this library since @mauron85 isn't even looking at it anymore. There is "maintained" fork which implemented this. -> https://github.com/darron1217/background-geolocation-android/ And in react-native part https://github.com/darron1217/react-native-background-geolocation I am closing this because it is 3 years later and no change.
2025-04-01T06:39:31.448908
2016-12-19T15:30:25
196442009
{ "authors": [ "culland", "mauron85" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8149", "repo": "mauron85/cordova-plugin-background-geolocation", "url": "https://github.com/mauron85/cordova-plugin-background-geolocation/issues/229" }
gharchive/issue
accuracy issues with points Not necessarily an issue, but wanted to see if other folks are experiencing this kind of thing. May app uses the standard geolocation tools with Phonegap as well as this plugin for the background geolocations. Everything is working reasonably well on android/iphone, except I have these random points with terrible accuracy, 20km+ out. Is this normal? Is there something I can do with the settings to limit this? My initial thought it to use location.accuracy to discard any bad points, but wanted to see if this is an issue I have with my setup overall. Thanks! Your Environment Plugin version: most recent Platform: iOS or Android OS version: varies Phonegap Build 5.2 Notice: this issue has been closed because it has been reported for branch 2.x or later. It could be already fixed in v3 or is not relevant anymore. You may reopen this issue if it has been closed in error.
2025-04-01T06:39:31.470222
2017-12-20T10:12:07
283511684
{ "authors": [ "mauron85", "pcrombach" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8150", "repo": "mauron85/cordova-plugin-background-geolocation", "url": "https://github.com/mauron85/cordova-plugin-background-geolocation/issues/360" }
gharchive/issue
Partial reconfiguration Your Environment Plugin version: 2.3.2 Platform: iOS and Android OS version: macOS HighSierra Device manufacturer and model: Samsung S7 Cordova version (cordova -v): 7.1.0 Cordova platform version (cordova platform ls): Plugin configuration options: all Link to your project: na Context IMHO partial reconfiguration doesn't work as described. If I try debug: true object as a parameter debug is not started. example BackgroundGeolocation.configure({ debug: true }); Expected Behavior Partial reconfigure should work Actual Behavior IMHO partial reconfiguration doesn't work as described. If I try debug: true object as a parameter debug is not started. example BackgroundGeolocation.configure({ debug: true }); Possible Fix I hope there is a fix or I am doing it wrong. Steps to Reproduce Context Debug logs Just tested in on android and works as advertised. What are your expectations about toggle debug? Plugin must be started to see difference Just realized your problem is plugin version 2.3.2. This feature is 3.0.0 only. You are right. My version of the plugin is 2.3.3 Where is version 3.0 If I reinstall the plugin it is still 2.2.2 Op 20 dec. 2017, om 18:46 heeft Marián Hello<EMAIL_ADDRESS>het volgende geschreven: Closed #360 https://github.com/mauron85/cordova-plugin-background-geolocation/issues/360. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/mauron85/cordova-plugin-background-geolocation/issues/360#event-1396577998, or mute the thread https://github.com/notifications/unsubscribe-auth/AAZlp2x1HUmUUCe_Dylvj70z7FaCUwkzks5tCUgMgaJpZM4RIK2d. Hi, I found the 3.0.0 version of the plugin: 3.0.0-alpha.12 right? If I remove the 2.3.2 version and install 3.0.0 my app doesn’t build anymore. Here the log of build proces: (node:64674) UnhandledPromiseRejectionWarning: Unhandled promise rejection (rejection id: 1): Error: /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/gradlew: Command failed with exit code 1 Error output: Note: Some input files use or override a deprecated API. Note: Recompile with -Xlint:deprecation for details. Note: Some input files use or override a deprecated API. Note: Recompile with -Xlint:deprecation for details. /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:6: error: package org.apache.http does not exist import org.apache.http.HttpResponse; ^ /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:7: error: package org.apache.http.client.methods does not exist import org.apache.http.client.methods.HttpPost; ^ /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:8: error: package org.apache.http.entity does not exist import org.apache.http.entity.StringEntity; ^ /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:9: error: package org.apache.http.impl.client does not exist import org.apache.http.impl.client.DefaultHttpClient; ^ /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:662: error: cannot find symbol DefaultHttpClient httpClient = new DefaultHttpClient(); ^ symbol: class DefaultHttpClient location: class LocationUpdateService /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:662: error: cannot find symbol DefaultHttpClient httpClient = new DefaultHttpClient(); ^ symbol: class DefaultHttpClient location: class LocationUpdateService /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:663: error: cannot find symbol HttpPost request = new HttpPost(url); ^ symbol: class HttpPost location: class LocationUpdateService /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:663: error: cannot find symbol HttpPost request = new HttpPost(url); ^ symbol: class HttpPost location: class LocationUpdateService /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:677: error: cannot find symbol StringEntity se = new StringEntity(params.toString()); ^ symbol: class StringEntity location: class LocationUpdateService /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:677: error: cannot find symbol StringEntity se = new StringEntity(params.toString()); ^ symbol: class StringEntity location: class LocationUpdateService /Users/pietcrombach/Development/Projecten/trinl/trinlfirebase/platforms/android/src/com/zencity/cordova/bgloc/LocationUpdateService.java:691: error: cannot find symbol HttpResponse response = httpClient.execute(request); ^ symbol: class HttpResponse location: class LocationUpdateService Note: Some input files use or override a deprecated API. Note: Recompile with -Xlint:deprecation for details. Note: Some input files use unchecked or unsafe operations. Note: Recompile with -Xlint:unchecked for details. 11 errors After reinstalling 2.00 version of the plugin these errors stay now. So what to do now????? Op 20 dec. 2017, om 18:46 heeft Marián Hello<EMAIL_ADDRESS>het volgende geschreven: Closed #360 https://github.com/mauron85/cordova-plugin-background-geolocation/issues/360. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/mauron85/cordova-plugin-background-geolocation/issues/360#event-1396577998, or mute the thread https://github.com/notifications/unsubscribe-auth/AAZlp2x1HUmUUCe_Dylvj70z7FaCUwkzks5tCUgMgaJpZM4RIK2d. You've installed wrong plugin. Probably this one https://www.npmjs.com/package/cordova-plugin-background-geolocation I can tell because com/zencity/cordova/bg is not package name I'm using, but it exactly matches this one https://github.com/ZenCity/cordova-plugin-background-geolocation/blob/master/src/android/LocationUpdateService.java#L1
2025-04-01T06:39:31.495446
2019-09-09T15:53:28
491181833
{ "authors": [ "Anuj-Raghuvanshi", "Anuj-logiciel", "busaku", "diegojoel98", "grabkipower", "itcurves", "jackie-d", "stgraham2000", "tfelici", "tkondej" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8151", "repo": "mauron85/cordova-plugin-background-geolocation", "url": "https://github.com/mauron85/cordova-plugin-background-geolocation/issues/619" }
gharchive/issue
iOS 13.1 Beta - Geolocation not reported when app is in background Seems like the geolocation is not reported when the app runs in background despite the "Always" permission is set Your Environment Plugin version: 3.0.6 Platform: iOS 13.1 beta OS version: Device manufacturer and model: IPhone 6s Cordova version (cordova -v): 9.0.0 (cordova-lib 9.0.1) Cordova platform version (cordova platform ls): ios 5.0.1 Plugin configuration options: Link to your project: Context On iOS 13.1 when the user has checked in the settings of the app the "always" permission for the iOS, the location is not beeing recorded after one leaves the app (switches the app to background i.e.). I can see that on the navtive Cordova Background Geolocation It this issue has been already raised and solved: transistorsoft/cordova-background-geolocation-lt#952 Expected Behavior The Geolocation should be properly sent even if the app is in the background Steps to Reproduce Turn on the "Always" geolocation permission (turned on from the settings page, not from the prompt - as it is no longer available in the iOS 13). Start tracking the geolocation Go to the desktop After approx. 30 seconds the location arrow on the top bar will disappear and the geolocation requests will cease beeing sent to the server. Going back to the application reenables sending of the geolocation requests @mauron85 I am sorry for being pushy, but do you have a clue about this one? iOS 13 releases tomorrow, there might be a great impact if the background geoloc report would not work in the background. For now, the issue still persists. It seems to be a general issue with iOS 13. You dont get (not every time, but sometimes, cant say explicit when) the blue tracking icon If the new authority confirm box (Only in foreground, Once, Never) you got BackgroundGeolocation.AUTHORIZED, if the user select "only in foreground". -- So the trackings stops in background without any response or callback etc. If you resume the app, you got the question from ios if you allowed every time in background and the tracking jumps from the last point in foreground to the current point But i dont have any idea atm to solve the issue :/ @busaku Mhm, thank you for the reply, well, at least I am not the only one fighting this bug. Is seems the issue happens only with ACTIVITY_PROVIDER, when using RAW_PROVIDER (and app has proper permissions) geolocation in background works fine. Hello @tkondej how to change that setting? @itcurves simply by specifying locationProvider, check the example https://github.com/mauron85/cordova-plugin-background-geolocation#example @tkondej, I'm still having issues with background data sending to my server using RAW_PROVIDER - I currently use DISTANCE_FILTER_PROVIDER but I switched to RAW_PROVIDER for testing. Both fail. I was wondering if you could explain what you mean by "and app has proper permissions". My app used to work prior to IOS13 so I was wondering if there is a new permission required for IOS13? @stgraham2000 on iOS 13 you have to enable access to use "geolocation always" in application settings or by accepting additional prompt which popups when after locking and unlocking the phone, check this article for more details https://medium.com/@transistorsoft/ios-13-and-android-q-support-beb7595d2c24 @tkondej, interesting, I had the "always on" under permissions for most of the day and it appeared to be broken, but it looks like another thing to look out for is you need to go for a longer distance test run. In earlier tests today I was probably only walking maybe 50m (around my office). I just tried again by doing a short drive and the points were collected as I expected. Before when I was testing on ios12 I could get points with a much shorter distance travelled. This will be an interesting usability issue to address as the article you referenced also indicates. It looks like this should still be possible to accomplish location updates while the permissions are set to "while in use": https://developer.apple.com/documentation/corelocation/getting_the_user_s_location/handling_location_events_in_the_background https://developer.apple.com/documentation/corelocation/cllocationmanager/1620568-allowsbackgroundlocationupdates I will experiment with this in the morning and report back Any news? :) Unfortunately I took horrible notes on this investigation but I have fixed the problem and here are the changes I in the commit that fixed this: Unfortunately, I took horrible notes on this investigation but I have fixed the problem and here are the changes I have in the commit that fixed this... Went from version 3.0.0-alpha.50 to 3.1.0 Following is a snipped of the config changed: startForeground: true, # Unset prior pauseLocationUpdates: false # Unset prior stationaryRadius: 10 # Used to be 50 Was using this.backgroundGeolocation.finish() at the end of the location handler. I removed these. I hope one of the above changes works for you. Sorry I can't be of more help. Hello, please look at my posts #684 In your config.xml be sure to add the following configuration under the iOs platform xml node: <config-file parent="UIBackgroundModes" target="*-Info.plist"> <array> <string>fetch</string> </array> <array> <string>location</string> </array> <array> <string>processing</string> </array> </config-file> <edit-config file="*-Info.plist" mode="merge" target="NSLocationWhenInUseUsageDescription"> <string>need location access to find things nearby</string> </edit-config> that's equivalent to the following setup from Xcode @jackie-d - Give a try to this plugin- https://www.npmjs.com/package/cordova-plugin-bg-geolocation https://github.com/Anuj-Raghuvanshi/cordova-plugin-bg-geolocation This works, Thanks! I have finally got this to work on a real device! the key change: locationProvider: BackgroundGeolocation.RAW_PROVIDER i,e, do NOT use ACTIVITY_PROVIDER as shown in the geolocation example in the docs as this doe NOT work in the background on IOS devices as explained here https://github.com/mauron85/cordova-plugin-background-geolocation/blob/HEAD/PROVIDERS.md I am not sure these are also required in the config.xml file under the 'ios platform' section as suggested above. I suspect the default already set by the plugin (i.e. 'location') might be sufficient. @tfelici - Can you please share reference of your plugin if you have forked and made changes? I also facing same issue. I have not done any fork. I am using the vanilla 3.1.0 version Kind regards Tom On 11/09/2020 12:05, Anuj Singh wrote: @tfelici https://github.com/tfelici - Can you please share reference of your plugin if you have forked and made changes? I also facing same issue. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/mauron85/cordova-plugin-background-geolocation/issues/619#issuecomment-691029600, or unsubscribe https://github.com/notifications/unsubscribe-auth/AB2PYTFAC43VNIMXLF3LQC3SFH76BANCNFSM4IU5AYGQ. ok - after further tests I can confirm this woks on IOS 13.7 on my daughter's iphone and ipad. Here are the relevant settings in BackgroundGeolocation.configure BackgroundGeolocation.configure({ locationProvider: BackgroundGeolocation.RAW_PROVIDER, desiredAccuracy: BackgroundGeolocation.HIGH_ACCURACY, pauseLocationUpdates: false, // need this for IOS I think (default was false anyhow but maybe not) stationaryRadius: 20, distanceFilter: 20, // docs say this is dinamically changed when locationProvider = DISTANCE_FILTER_PROVIDER }); you do NOT need to add the extra lines <config-file parent="UIBackgroundModes" target="*-Info.plist"> <array> <string>location</string> </array> <array> <string>processing</string> </array> </config-file> unless you require to app to refresh other stuff in the background, other than location updates and posting Hi, there are two ways to works geolocation in background, the first is as you commented with RAW_PROVIDER and the second is setting up in locationProvider: backgroundGeolocation.DISTANCE_FILTER_PROVIDER and call in your event location the method switchMode to force foreground and with this continue reporting in base to option.desiredAccuracy and option.distanceFilter BackgroundGeolocation.on('location', function(location) { // handle your locations here // switch to FOREGROUND mode BackgroundGeolocation.switchMode(BackgroundGeolocation.FOREGROUND_MODE); }) But I have a question, in the two ways this works and I can see blue in the status bar that means an app is actively using your location. My question is this can drain the battery? is not sending location until you reach your distance filter but this is always in the status bar. I hope anyone see this
2025-04-01T06:39:31.548237
2016-12-07T19:02:52
194137831
{ "authors": [ "azellsworth" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8152", "repo": "mavenlink-solutions/netsuite", "url": "https://github.com/mavenlink-solutions/netsuite/pull/5" }
gharchive/pull-request
Add search for ExpenseCategory to NetSuite Records pulls in all categories from NetSuite @ahuth @shirish-pampoorickal @AnnaMhairi @shirish-pampoorickal comments addressed
2025-04-01T06:39:31.573211
2024-02-29T21:25:07
2162128297
{ "authors": [ "maxcountryman", "thor314" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8153", "repo": "maxcountryman/tower-sessions-stores", "url": "https://github.com/maxcountryman/tower-sessions-stores/issues/8" }
gharchive/issue
Diesel session store support I noticed a discussion in tower-sessions between the maintainer of this repo and a contributor intending to add support for Diesel, which seems to have stalled at the end of last year. I'd like to have support for Diesel, so am proposing to create a pull request with the relevant parts, if that would be welcome. Please let me know if there's anything I might like to take into consideration. We don't vendor session stores so please publish this as an independent crate. There's a number of examples listed in the docs and README which you might reference. Leaving a note here for others to find regarding diesel. I spent about 2 hours looking into writing my own session store and evaluated it as not worth it, and will be migrating my project to sqlx instead, which seems generally better supported both by tower-sessions and other ecosystem projects I'm relying on (shuttle). It was mentioned to me elsewhere that Diesel's has been losing support relative to sqlx/seaORM on the basis of one or more of: of slow compilation times inconsistent async support slower developer experience than just writing compile-time checked SQL a necessarily complex trait ecosystem creating friction for open source maintainers to support diesel
2025-04-01T06:39:31.577983
2015-07-14T13:41:26
94947537
{ "authors": [ "alvinreuben", "luxis" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8154", "repo": "maxdrift/RMImagePicker", "url": "https://github.com/maxdrift/RMImagePicker/issues/2" }
gharchive/issue
having a problem without knowing how to fix this. Hi. I am currently trying to use your imagePicker in my own project, but I get this error even in my test that I created with your code. Do you have any idea? I have a problem in installing via cocoapods. fingents-Mac-mini-3:HoneyBadger fingent$ pod install Updating local specs repositories CocoaPods 0.39.0.beta.4 is available. To update use: gem install cocoapods --pre [!] This is a test version we'd love you to try. For more information see http://blog.cocoapods.org and the CHANGELOG for this version http://git.io/BaH8pQ. Analyzing dependencies [!] Unable to satisfy the following requirements: RMImagePicker (~> 0.1.4) required by Podfile Here is the podfile. Appreciate your help. Uncomment this line to define a global platform for your project platform :ios, '6.0' target 'HoneyBadger' do pod "RMImagePicker","~> 0.1.4" end target 'HoneyBadgerTests' do end
2025-04-01T06:39:31.581522
2017-10-27T23:06:23
269260007
{ "authors": [ "Tim77277", "manolisep" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8155", "repo": "maxep/MXParallaxHeader", "url": "https://github.com/maxep/MXParallaxHeader/issues/84" }
gharchive/issue
Floating Section Header not scrolling with tableView after using MXParallaxHeader for the tableHeaderView So I am facing this problem on my project. Somehow we must use UITableView with Plain Style and then dynamically adding table section header in to the tableView. And for the TableHeaderView of course I am using MXParallaxHeader. However, when I scroll the tableView up, my section header view got stock at the origin position while all cells are scrolling up. It is a very weird behavior I never seen before. I know #65 asked the same question but there is no a correct response. Please help. Thank you So I tried #65 solution again and it worked. Basically what you should do to make UITableView Section Header works with MXParallaHeader is to add UITableView into MXScrollView, assign your MXParallaxHeader (View) to the scrollView.parallaheader.view. And then set the correct scrollView.contentSize otherwise it won't be scrollable. This works well for my project, but the upcoming issue is that scrollView.parallaheader.minimumHeight doesn't work in this case. If someone know the solution, please share. Thanks @Tim77277 I did that like you say , it worked . But It have a problem, when I scroll the list to the bottom , I can't keep scroll down although I set scrollView.bounces = true and collectionView.bounces = true. Did you get this problem? Same problem here @Tim77277 I did that like you say , it worked . But It have a problem, when I scroll the list to the bottom , I can't keep scroll down although I set scrollView.bounces = true and collectionView.bounces = true. Did you get this problem? Same problem here
2025-04-01T06:39:31.588537
2019-02-13T14:36:39
409835168
{ "authors": [ "Sebastian-Hirsch", "maximal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8156", "repo": "maximal/audio-waveform-php", "url": "https://github.com/maximal/audio-waveform-php/issues/5" }
gharchive/issue
Return value of getWaveform function I tried your package, but when i used the function in the title, I get a boolean (true) instead of the image I expected to get. The getWaveformData function on the other hand return the expected waveform data. I used Version 1.2.0 of the library. I used PHP 7.2.15 and the GD-Extension is enabled. See the readme file and the PhpDoc for getWaveform() method: /** * Get waveform from the audio file. * @param string $filename Image file name * @param int $width Width of the image file in pixels * @param int $height Height of the image file in pixels * @param bool $onePhase `true` to get positive values only, `false` to get both phases * @return bool Returns `true` on success or `false` on failure. * @throws \Exception */ getWaveform() saves the image to $filename and returns boolean (whether the operation succeded).
2025-04-01T06:39:31.596236
2021-02-09T13:22:49
804546856
{ "authors": [ "Ayfri", "maximousblk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8157", "repo": "maximousblk/ghlog", "url": "https://github.com/maximousblk/ghlog/issues/4" }
gharchive/issue
Suggestion: Add multiple template with template selection I think that you should make multiple differents templates, ones with categories, ones without, one without emojis, etc, and let us select it through the CLI (with a --template argument for example). This would greatly improve this project! I knew someone would ask 😃 I originally wanted this to be more of a DIY solution where you have the info and you can arrange it as you desire and do whatever with the output. I would add multiple templates but there is so much info and so many ways to create changelogs and everyone wants it to be a certain way. so I only have one (opinionated) template by default and export all the info (sorted and unsorted) for people to create their own. for example, in the default template, I put a link to the commit with the short hash. some like to have the link on the commit text, some like to put it on the end, some only like to put the hash but not a link, some people would prefer not to have it altogether. this would either require me to add every requested template, or create a templating language which is dynamic enough. I'd rather that language be javascript/typescript than creating my own. Mh I see, but for example: When I tried to modify this to only have only 1 "Changelog" category, I struggled a lot because code is very axed for multiple categories, maybe you should change this, like regrouping every part of the configuration in 1 ts file? if you just want a list of commits, you can do something like: //<EMAIL_ADDRESS>const { _meta }= await getChangeLog(repo, base, head, config); const changelog = `CHANGELOG ${ _meta.commits.all.map(({ message }) => `- ${message}` ).join("\n") } ` // write `changelog` to file I'm trying to think of a good way for how to structure the out put of getChangeLog() so that it's as flexible as possible. Mh I see okay, I was just doing it badly!
2025-04-01T06:39:31.599501
2024-08-07T17:29:34
2453983830
{ "authors": [ "maxin-cn", "toyxyz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8158", "repo": "maxin-cn/Cinemo", "url": "https://github.com/maxin-cn/Cinemo/issues/11" }
gharchive/issue
omegaconf.errors.ConfigAttributeError: Missing key ckpt Thanks for the great work! When I ran video_editing.sh, I got the following error. I checked the sample.yaml and there is no ckpt part. Am I missing something? The gardio app works fine. (cinemo) C:\Users\toyxy\Cinemo>bash pipelines/video_editing.sh C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\diffusers\utils\outputs.py:63: FutureWarning: `torch.utils._pytree._register_pytree_node` is deprecated. Please use `torch.utils._pytree.register_pytree_node` instead. torch.utils._pytree._register_pytree_node( C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\huggingface_hub\file_download.py:1150: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`. warnings.warn( C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\diffusers\models\modeling_utils.py:109: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. return torch.load(checkpoint_file, map_location="cpu") Traceback (most recent call last): File "C:\Users\toyxy\Cinemo\pipelines\video_editting.py", line 198, in <module> main(OmegaConf.load(args.config)) File "C:\Users\toyxy\Cinemo\pipelines\video_editting.py", line 87, in main state_dict = find_model(args.ckpt) ^^^^^^^^^ File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\dictconfig.py", line 355, in __getattr__ self._format_and_raise( File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\base.py", line 231, in _format_and_raise format_and_raise( File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\_utils.py", line 899, in format_and_raise _raise(ex, cause) File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\_utils.py", line 797, in _raise raise ex.with_traceback(sys.exc_info()[2]) # set env var OC_CAUSE=1 for full trace ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\dictconfig.py", line 351, in __getattr__ return self._get_impl( ^^^^^^^^^^^^^^^ File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\dictconfig.py", line 442, in _get_impl node = self._get_child( ^^^^^^^^^^^^^^^^ File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\basecontainer.py", line 73, in _get_child child = self._get_node( ^^^^^^^^^^^^^^^ File "C:\Users\toyxy\ANACON~1\envs\cinemo\Lib\site-packages\omegaconf\dictconfig.py", line 480, in _get_node raise ConfigKeyError(f"Missing key {key!s}") omegaconf.errors.ConfigAttributeError: Missing key ckpt full_key: ckpt object_type=dict Hi, thanks for your interest. I have fixed this bug. the parameter ckpt will not be used. You can follow this to change your code: https://github.com/maxin-cn/Cinemo/blob/d0e99565dc60084f6afc6c8767463ffc5337b082/pipelines/video_editing.py#L90
2025-04-01T06:39:31.613780
2017-06-20T19:10:45
237314393
{ "authors": [ "kevintmcg", "maxvu" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8159", "repo": "maxvu/git-demo", "url": "https://github.com/maxvu/git-demo/pull/4" }
gharchive/pull-request
Added name to list Added name to list, please merge into master branch. Hey Kevin, I think this might the wrong repo. stevecif/git-demo eventually derives from this one but now lists a different class. Reach out to Steve to confirm. If you're sure you want to merge into this one, please restage your changes and squash any intermediary commits. I'll close this request in the meantime. Let me know if you need any help, Max Thanks. Will do. On Wed, Jun 21, 2017 at 8:52 PM, Max Vu<EMAIL_ADDRESS>wrote: Hey Kevin, I think this might the wrong repo. stevecif/git-demo https://github.com/stevecif/git-demo eventually derives from this one but now lists a different class. Reach out to Steve to confirm. If you're sure you want to merge into this one, please restage your changes and squash any intermediary commits. I'll close this request in the meantime. Let me know if you need any help, Max — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/maxvu/git-demo/pull/4#issuecomment-310245037, or mute the thread https://github.com/notifications/unsubscribe-auth/AbaZ-Xu6ueb1X8zV97QBbAm6J7D6t0aYks5sGbrrgaJpZM4OABph .
2025-04-01T06:39:31.623046
2016-08-11T20:34:24
170739520
{ "authors": [ "KaijenHsiao", "aisch" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8160", "repo": "mayfieldrobotics/ros-webrtc", "url": "https://github.com/mayfieldrobotics/ros-webrtc/issues/67" }
gharchive/issue
Ansible playbook missing stuff Trying to run the ansible-playbook command in the README.md, it seems to be missing most of its roles: kaijen@daisy:~/ros_webrtc_ws/src/ros-webrtc/test/provision$ ansible-playbook -i 'localhost,' -c local -K dev.yml SUDO password: ERROR! the role 'chrome' was not found in /home/kaijen/research/ros_webrtc_ws/src/ros-webrtc/test/provision/roles:/home/kaijen/research/ros_webrtc_ws/src/ros-webrtc/test/provision:./roles The error appears to have been in '/home/kaijen/ros_webrtc_ws/src/ros-webrtc/test/provision/dev.yml': line 16, column 5, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: roles: {role: chrome, become: yes} ^ here @KaijenHsiao sorry, can u do: ~/code/ros-webrtc/test/provision$ ansible-galaxy install -r requirements.yml just before you run: ~/code/ros-webrtc/test/provision$ ansible-playbook -i 'localhost,' -c local -K dev.yml that should install dependent roles. i'll update the docs if that fixes the issue. Ah, yup, that fixed it. Thanks!
2025-04-01T06:39:31.624035
2024-02-08T20:20:01
2126005311
{ "authors": [ "maypok86", "narqo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8161", "repo": "maypok86/otter", "url": "https://github.com/maypok86/otter/pull/49" }
gharchive/pull-request
Update README to make the wording friendlier This PR comes with a small rephrasing of the README's "Motivation" section, making the overall text simper and friendlier. Thanks for the pull request!
2025-04-01T06:39:31.633947
2017-09-14T09:18:41
257650046
{ "authors": [ "2rcsanyi", "mazen160" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8162", "repo": "mazen160/struts-pwn_CVE-2017-9805", "url": "https://github.com/mazen160/struts-pwn_CVE-2017-9805/issues/4" }
gharchive/issue
Echo doesn't work Hi! I've created a vulnerable server using Tomcat9, Struts2 REST Showcase webapp (v2.5.12) and Ubuntu 16.04 in Docker - and your exploit works like a charm (I can touch and rm files), however echo fails for me. The good old ysoserial has some issues with complex commands (pipes and I/O redirection) also, but since you're using echo in your example, maybe I'm just missing something - do you have any ideas how to fix it? Hi @2rcsanyi , Thanks for reporting! It seems that piping and I/O redirections are not working properly in the current exploit payload. I have updated the instructions and examples to use touch instead of echo. I would say it would be better to drop a file-based payload on the box, then execute it. It would be the easiest and most reliable way to do it in my opinion. Thanks again! Best, Mazin
2025-04-01T06:39:31.649937
2022-03-15T23:46:04
1170374435
{ "authors": [ "fletcherm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8163", "repo": "mbbsemu/MBBSEmu", "url": "https://github.com/mbbsemu/MBBSEmu/issues/556" }
gharchive/issue
[WCCMMUD] Majormud - Re-entering the game causes your character to drop half his stuff and lose some health Module Information Identifier WCCMMUD Module Name Majormud Version 1.11p Describe the bug Re-entering the game after a hangup causes your character to drop stuff on the group and lose some health. To Reproduce Steps to reproduce the behavior: Create a character, enter the game, go to the armor shop, buy some gear, and equip all of it. Ensure you are at full health. Click the hangup button in Megamud. (or force disconnect the session via whatever your client is) Reconnect and login to the game See a bunch of your gear on the ground. Also see you are missing some health. Expected behavior You should re-enter the game with the same amount of health as you had, and still have your loot equipped. Screenshots Before disconnect above, after disconnect below Software Information: OS: mbbsemu docker Telnet Client Majormud Nevermind. Realized that there is a hangup penalty configured in the WCCMMUD.MSG file.
2025-04-01T06:39:31.724908
2024-03-12T06:27:18
2180833229
{ "authors": [ "aclayton555", "vpchung" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8165", "repo": "mc2-center/data-models", "url": "https://github.com/mc2-center/data-models/issues/92" }
gharchive/issue
[docs] Fill out the Home and Next Steps pages The docs site could use some love on the landing page and the Next Steps page. CC @aclayton555 for awareness Once we have the first iteration of the community curation docs up on the scroll viewport, would be good to have a "return to curation docs" button within the standards pages to streamline navigation back to these. Change "Contribute" tab to "Contribute to the Data Model"
2025-04-01T06:39:31.790973
2019-07-15T01:14:38
467906874
{ "authors": [ "BuddermanTheAmazing", "mccreery" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8166", "repo": "mccreery/better-hud", "url": "https://github.com/mccreery/better-hud/issues/69" }
gharchive/issue
Twilight Forest Clock Bed? I just hopped into Twilight Forest and a white Bed icon appeared over the game clock. I tried to see if it was an issue with the placement but it's attached directly to the clock. Is there a way to disable this icon? The bed indicates that it's nighttime and therefore possible to sleep. You're right, I should probably add a way to disable it.
2025-04-01T06:39:31.940164
2016-07-28T18:25:53
168163713
{ "authors": [ "JaymerJaymer", "jane0815", "mchristopher", "tinarazzburton" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8169", "repo": "mchristopher/PokemonGo-DesktopMap", "url": "https://github.com/mchristopher/PokemonGo-DesktopMap/issues/354" }
gharchive/issue
Change method of notification/filter list The current method of adding names to the "hide common" and "notify" list is pretty cumbersome. I would suggest pulling up a list of all names and being able to check/uncheck names. If we could export and import lists, that would be helpful too as my list occasionally disappears. i agree, i had a long list of Pokemon ID#s that was EASY to copy/paste into view or hide. While some "do gooder" thought it was "cool" to modernize the UI, it doesn't work easily. You can't copy/paste a list either. I agree it is not very userfriendly. For me this is the most important feature of the map. @tinarazzburton @JaymerJaymer @jane0815 I'd recommend putting an issue in at https://github.com/AHAAAAAAA/PokemonGo-Map. This project simply makes that one easy to use, and I'm sure they'd welcome the feedback.
2025-04-01T06:39:31.944355
2024-10-15T00:26:14
2587342641
{ "authors": [ "AshtakaOOf", "HellerCommaA", "mcmonkey4eva" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8170", "repo": "mcmonkeyprojects/SwarmUI", "url": "https://github.com/mcmonkeyprojects/SwarmUI/pull/325" }
gharchive/pull-request
320: Add x to filter params for https://github.com/mcmonkeyprojects/SwarmUI/issues/320 uhhh also an x to clear should only show up if there's content to be cleared, not while empty closing for now, don't feel like fighting css. Turn it into a Draft?
2025-04-01T06:39:31.969512
2016-09-27T15:19:49
179524571
{ "authors": [ "jsumners", "teomurgi" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8171", "repo": "mcollina/pino", "url": "https://github.com/mcollina/pino/issues/96" }
gharchive/issue
pino cli discards not properly formatted log lines When piping pino in the console, pino discards the lines that are not formatted exactly as it expects. Bunyan instead simply prints those line not formatted. pull request: #95 Duplicate of #94
2025-04-01T06:39:31.996015
2024-10-01T19:07:12
2559988606
{ "authors": [ "Suvadip-sana", "mdazfar2" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8172", "repo": "mdazfar2/Ezyshop", "url": "https://github.com/mdazfar2/Ezyshop/issues/28" }
gharchive/issue
I want to add more category Hi! I am Suvadip Sana, I want to work on this project ISSUE I think in shopping website has more categories, For now it has only 3 Suggestion I suggest for this categories for adding - Home & Living Beauty & Personal Care Sports & Fitness Books & Media Toys & Games Please assign me this issue if you think it's an issue need to fix @Suvadip-sana go ahead, but remember to upload AI-generated images. Please don't copy from Google or any other website. Could you please provide me the images? or shall i add myself. Could you please provide me the images? or shall i add myself. You can add yourself @Suvadip-sana Ok. Thank you. @mdazfar2 @thakuratul2 I add this three more row of categories for 9 new and essential categories that might be presnt on shopping site. Please see! If you are happy then I create a new branch and make a pull request. @Suvadip-sana That's amazing! Do one thing: for every three columns, add a load more button. After clicking it, more columns will become visible, and after clicking load more again, it will show another set. Hope you understand. hope you understand. Don't forget to join our official Discord server. And also follow us on Linkedin And please ⭐ our repo Ok. Try my best. https://github.com/user-attachments/assets/32c1bd3a-84a1-4d96-be5f-3ec4551756e8 Check it. Should you please give me max label for this. It take way more time. Actually I work on two issue. So Please! If you give me I would be very grateful! @Suvadip-sana, that's amazing! Please create a pull request and update your sync fork your branch. Could you please give me max label for this? Added @Suvadip-sana You miss hactoberfest label You miss hactoberfest label. please add this. Then i raise a pull request Done! Anything else which i forget please let me know Should you close this issue? Because it merge already!
2025-04-01T06:39:31.998752
2024-06-13T11:44:10
2350930642
{ "authors": [ "RamakrushnaBiswal", "mdazfar2" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8173", "repo": "mdazfar2/HelpOps-Hub", "url": "https://github.com/mdazfar2/HelpOps-Hub/issues/389" }
gharchive/issue
Feat: Adding particle.js in landing page Describe the feature Adding particle.js in landing page Add ScreenShots https://vincentgarreau.com/particles.js/ chek this Record [X] I agree to follow this project's Code of Conduct [X] I'm a GSSOC'24 contributor [X] I want to work on this issue @RamakrushnaBiswal go ahead 🚀 solved with PR #401
2025-04-01T06:39:31.999759
2023-10-11T10:04:19
1937367888
{ "authors": [ "cafrisoft", "yiszza" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8174", "repo": "mdc-maui/mdc-maui", "url": "https://github.com/mdc-maui/mdc-maui/issues/32" }
gharchive/issue
[enhancement TextField] ReadOnly mode Please, add readonly mode in TextField Copy. It exists now.
2025-04-01T06:39:32.002587
2017-11-29T17:18:35
277838178
{ "authors": [ "bmammon" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8175", "repo": "mddub/nightscout-osx-menubar", "url": "https://github.com/mddub/nightscout-osx-menubar/issues/10" }
gharchive/issue
Delta Value switches to +0.0 Delta value is correct for ~1 minute after a new reading, but then it becomes "+0.0". I deleted and reinstalled and now it's working again! The problem was that I had iosxdripreader sending SGV values directly to Nightscout and also to dexcom share servers. I had the bridge variable enabled so after fetching from dexcom share, both readings would be equal and the slope would become 0.0. Once I deleted the "bridge" variable from my ENABLE line in Heroku, everything worked great. Thank you to @miguelkennedy for figuring this out!
2025-04-01T06:39:32.005986
2023-12-30T12:21:55
2060732226
{ "authors": [ "sector101010" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8176", "repo": "mdelobelle/metadatamenu", "url": "https://github.com/mdelobelle/metadatamenu/issues/504" }
gharchive/issue
[BUG] Canvas Group: Moving note outside all groups breaks the update feature Obsidian v1.5.3 Metadata-Menu 0.6.11 priority is an inherited canvas-group field Note is inside of a group in the canvas and the priority field has the correct group name: Move note outside the group Move note back into a group - field is not updated and console shows no error Sometimes it works, sometimes it doesn't. It's real strange Updated description: removed the console error, as that was caused by something else. There is no error shown in the console now. Sometimes it works, sometimes not.
2025-04-01T06:39:32.007825
2023-02-16T15:23:45
1587881387
{ "authors": [ "mdhiggins", "tchirou" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8177", "repo": "mdhiggins/PlexAutoSkip", "url": "https://github.com/mdhiggins/PlexAutoSkip/issues/31" }
gharchive/issue
Feature request : Credits detection auto skip Hi Plex has released the credit detection : https://support.plex.tv/articles/credits-detection/ Any chance you could implement the auto skip like with the intro ? Thanks a lot for all your work Already supported. Just add "credits" to your tags Fantastic ! Thanks a lot
2025-04-01T06:39:32.013256
2021-12-08T12:48:36
1074375153
{ "authors": [ "kennfoxx", "mdhiggins" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8178", "repo": "mdhiggins/sickbeard_mp4_automator", "url": "https://github.com/mdhiggins/sickbeard_mp4_automator/issues/1503" }
gharchive/issue
How to use with GPU? Hi everyone, I've been looking for information on using sickbeard_mp4_automator on a GPU server. The machine is Linux Ubuntu 18 + RTX 2070, could someone guide me how to configure? Lots of discussion on previous issues about this, just search for nvenc for closed issues on this repo to see what other people have done First step is you'll need to understand hardware decoding / encoding for FFMPEG to determine what options you want since this is actually what drives everything https://trac.ffmpeg.org/wiki/HWAccelIntro and specifically additional documentation for NVIDIA https://docs.nvidia.com/video-technologies/video-codec-sdk/ffmpeg-with-nvidia-gpu/ Once you have a working build of FFMPEG and can get hardware encoding working manually then you're ready to configure the script to automate it The main options you'll want for decoding are hwaccels hwaccel-decoders hwdevices hwaccel-output-format and for encoding you'll just want to set your codec to one of the nvenc encoders depending on if you want h264 or hevc/h265 So just set your first codec to h264_nvenc or hevc_nvenc, though you should include hevc or h264 after that so that remuxing can still be enabled Example [Video] codec = hevc_nvenc, hevc That should be more than enough to get you started https://github.com/mdhiggins/sickbeard_mp4_automator/issues/1444 https://github.com/mdhiggins/sickbeard_mp4_automator/issues/1499 https://github.com/mdhiggins/sickbeard_mp4_automator/issues/1463 https://github.com/mdhiggins/sickbeard_mp4_automator/issues/1421 Very grateful for the help.
2025-04-01T06:39:32.015027
2018-10-19T18:48:11
372074231
{ "authors": [ "jeff47", "mdhiggins" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8179", "repo": "mdhiggins/sickbeard_mp4_automator", "url": "https://github.com/mdhiggins/sickbeard_mp4_automator/issues/944" }
gharchive/issue
manual.py: Exception while retrieving poster "NoneType" object has no attribute 'strip'. Hello, I'm seeing this error when running manual.py. Is this related to the download-artwork option in autoProcess.ini? I have that set to Thumbnail. Can you post the whole error?
2025-04-01T06:39:32.257270
2018-08-29T20:54:05
355329114
{ "authors": [ "chrisdavidmills", "jmswisher" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8180", "repo": "mdn/sprints", "url": "https://github.com/mdn/sprints/issues/450" }
gharchive/issue
Work on content for Hack on MDN a11y event Acceptance criteria [x] List of suggested tasks and other supporting material produced for starting Tasks If relevant, you can make a checklist for tasks. [ ] Get existing material from Janet. [ ] Look over Estelle's a11y content audit. [ ] Refine material and produce list of suggested tasks for working on in the sprint. [ ] Other supporting material researched; what else do we need? Ask Florian what he did for the BCD HoM. Closing this because the Acceptance Criteria is met.
2025-04-01T06:39:32.289059
2016-12-14T17:31:59
195595219
{ "authors": [ "Dzol", "meadsteve" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8181", "repo": "meadsteve/white-bread", "url": "https://github.com/meadsteve/white-bread/pull/81" }
gharchive/pull-request
HTML Output for White Bread Nearing completion so this PR will help to review. Code looks 👍 so far. Could you add something to the README about how to switch to this kind of output? Certainly!
2025-04-01T06:39:32.292396
2015-09-26T17:53:13
108483754
{ "authors": [ "codydaig", "ilanbiala", "lirantal", "simison" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8182", "repo": "meanjs/mean", "url": "https://github.com/meanjs/mean/issues/938" }
gharchive/issue
Node-pre-gyp package? @lirantal Where are we using the node-pre-gyp package? @ilanbiala maybe you should ask Cody: 5d5d1b7b (Cody Daig 2015-07-17 01:09:08 -0700 50) "node-pre-gyp": "0.6.4", @codydaig do you know what the node-pre-gyp package is doing? I don't see it being required anywhere. Running npm ls shows it being required in a couple of places, notably: ├─┬<EMAIL_ADDRESS>│ ├─┬<EMAIL_ADDRESS>│ │ └──<EMAIL_ADDRESS>│ ├──<EMAIL_ADDRESS>│ └─┬<EMAIL_ADDRESS>│ ├─┬<EMAIL_ADDRESS>│ │ ├─┬<EMAIL_ADDRESS>│ │ │ └─┬<EMAIL_ADDRESS>│ │ │ ├──<EMAIL_ADDRESS>│ │ │ └─┬<EMAIL_ADDRESS> @simison that's fine that it is depended on, but there is no reason to have it as a top level dependency of our app. @ilanbiala Sorry for the delay. That came a PR I did to get rid of the npm warnings for the future upgrade of npm 3. It had something to do with dependency resolution. It's not used anywhere though, do you remember the PR? @ilanbiala just submit a PR to remove it and we'll see if it fails or not in the CI :) #1001
2025-04-01T06:39:32.309993
2016-11-03T15:01:20
187084707
{ "authors": [ "denisnazarov", "vyzo", "yusefnapora" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8183", "repo": "mediachain/aleph", "url": "https://github.com/mediachain/aleph/pull/66" }
gharchive/pull-request
Use jq for pretty printing Pipes json through jq for pretty printing instead of using console.dir, so we get valid json output with quotes and such. Defaults to printing in color if stdout is a TTY, but you can override with prettyPrint(obj, {color: false}) Will add the flag for pretty printing per @vyzo's suggestion; do we want it to be enabled or disabled by default? Pretty output by default seems nicer for end users... @denisnazarov, thoughts? I like pretty print by default @parkan what do you think fine by me either way.
2025-04-01T06:39:32.331129
2023-11-22T13:35:08
2006355898
{ "authors": [ "Mindbowser", "bnapora", "luis901101" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8184", "repo": "medplum/medplum", "url": "https://github.com/medplum/medplum/issues/3363" }
gharchive/issue
Login issue We setup self hosted solution and ran on ubuntu. On login page we got stuck, the error says Cannot read properties of undefined (reading 'digest') We're have same issue on self-hosted installation in Google Cloud. What was the resolution? This issue is because of the following call made on Sign in: crypto.subtle.digest('SHA-256', new TextEncoder().encode(str)); If you are not accessing by http://localhost then you should configure to use https to be able to login, because most browsers implementations requires a secure origin in order to allow using WebCrypto API:
2025-04-01T06:39:32.342946
2023-09-05T09:39:03
1881598201
{ "authors": [ "medyo", "yefreescoding" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8185", "repo": "medyo/hackertab.dev", "url": "https://github.com/medyo/hackertab.dev/issues/163" }
gharchive/issue
Disable transitions on theme toggle I've noticed that when switching to light mode and vice versa, the page transition causes the colors of the components to change at different times. This might be caused by varying transition values for each component. There's a simple and easy way to solve this problem, and this article explains it quite effectively. Link: https://paco.me/writing/disable-theme-transitions Looks interesting, could you spare a moment to submit a pull request?
2025-04-01T06:39:32.363603
2018-03-27T13:14:04
308963461
{ "authors": [ "pdellaert", "rygos", "sublabat" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8186", "repo": "meetfranz/franz", "url": "https://github.com/meetfranz/franz/issues/850" }
gharchive/issue
Franz 5.0.0. beta drives my MBP fans mad franz 5.0.0. beta drives my 2015er retina mbp fans mad. has anybody a similiar issue? For me the same. Franz 5.0.0-beta.17.834 macOS 10.13.4 Beta(17E197a) Running Services with Franz: WhatsApp, Telegram, Discord, Github, Skype and Riot. CPU load of Franz: 20% the whole time its open. some peaks above if I switch services. I'm still seeing this issue. Mac OS 10.14 (18A391), 3.1 GHz Intel Core i7, 16 GB 1867 MHz DDR3 Franz 5.0.0-beta.24.1110 CPU usage from activity monitor: Franz Helper: 52.5% Franz Helper: 29.2% Franz: 27.8% Franz Helper: 26.3% Franz Helper: 18.2% Services used: Messenger, WhatsApp, 4x Slack It basically overloads my whole system, blocking almost any action, slowing down all programs. Franz has become useless for me, because of this...
2025-04-01T06:39:32.366012
2017-11-20T13:39:42
275358923
{ "authors": [ "adlk", "pluma" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8187", "repo": "meetfranz/plugins", "url": "https://github.com/meetfranz/plugins/issues/59" }
gharchive/issue
[Deploy] FastMail Service FastMail Link to your repository https://github.com/foss-haas/franz-fastmail Additional Information A version compatible with franz4 also exists in a separate branch. Deployed, thanks for your contribution. :tada: Thank you for your awesome work on Franz!
2025-04-01T06:39:32.383290
2023-05-13T16:41:13
1708679006
{ "authors": [ "TentacleCat", "transpchan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8188", "repo": "megvii-research/IJCAI2023-CoNR", "url": "https://github.com/megvii-research/IJCAI2023-CoNR/issues/13" }
gharchive/issue
I use the colab version and try to run the example, but failed Note that --use-env is set by default in torchrun. If your script expects --local-rank argument to be set, please change it to read from os.environ['LOCAL_RANK'] instead. See https://pytorch.org/docs/stable/distributed.html#launch-utility for further instructions warnings.warn( usage: train.py [-h] [--world_size WORLD_SIZE] [--local_rank LOCAL_RANK] [--dataloader_imgsize DATALOADER_IMGSIZE] [--batch_size BATCH_SIZE] [--model_name MODEL_NAME] [--dataloaders DATALOADERS] [--mode {train,test}] [--test_input_person_images TEST_INPUT_PERSON_IMAGES] [--test_input_poses_images TEST_INPUT_POSES_IMAGES] [--test_checkpoint_dir TEST_CHECKPOINT_DIR] [--test_output_dir TEST_OUTPUT_DIR] [--test_output_video TEST_OUTPUT_VIDEO] [--test_output_udp TEST_OUTPUT_UDP] [--test_pose_use_parser_udp TEST_POSE_USE_PARSER_UDP] train.py: error: unrecognized arguments: --local-rank=0 ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 2) local_rank: 0 (pid: 993) of binary: /usr/bin/python3 Traceback (most recent call last): File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/usr/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/usr/local/lib/python3.10/dist-packages/torch/distributed/launch.py", line 196, in main() File "/usr/local/lib/python3.10/dist-packages/torch/distributed/launch.py", line 192, in main launch(args) File "/usr/local/lib/python3.10/dist-packages/torch/distributed/launch.py", line 177, in launch run(args) File "/usr/local/lib/python3.10/dist-packages/torch/distributed/run.py", line 785, in run elastic_launch( File "/usr/local/lib/python3.10/dist-packages/torch/distributed/launcher/api.py", line 134, in call return launch_agent(self._config, self._entrypoint, list(args)) File "/usr/local/lib/python3.10/dist-packages/torch/distributed/launcher/api.py", line 250, in launch_agent raise ChildFailedError( torch.distributed.elastic.multiprocessing.errors.ChildFailedError: train.py FAILED Failures: <NO_OTHER_FAILURES> Root Cause (first observed failure): [0]: time : 2023-05-13_16:38:58 host : 38b080a41a03 rank : 0 (local_rank: 0) exitcode : 2 (pid: 993) error_file: <N/A> traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html 抱歉,由于部分作者只对水文章感兴趣,v1已经没人维护了。建议改用 https://github.com/transpchan/Live3D-v2 。
2025-04-01T06:39:32.434968
2021-05-06T12:13:09
877431608
{ "authors": [ "RonitRay", "jjmanglani01" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8189", "repo": "mehulcse/findmyslot", "url": "https://github.com/mehulcse/findmyslot/issues/3" }
gharchive/issue
Cloudfront rate-limiting not handled Cowin APIs are behind a CloudFront CDN that after a certain number/frequency of requests blocks the device for some time. I'm not sure how long since it only happened to me once, but it is likely around half an hour without any requests. When this limit is exceeded, all GET requests to the API return a 403 Forbidden error 403 Forbidden error is not handled by your client. There is no message or alert that a block has been imposed for a while and the user should try again in a while. To the user it seems like the site is running as expected and refreshing at the defined frequency. It's possible that if they keep the tab open the rate limit will persist. This is an error from open APIs which we are using to fetch data. We are currently using useQuery which is taking care of retry logic for API. Although we will add try...catch block to avoid error If this happens again let us know. Thanks for pointing that out. Hi, This issue is fixed in this commit https://github.com/mehulcse/findmyslot/commit/42cb4353cf2a2458986c8cc7ee7287254bc3c309
2025-04-01T06:39:32.440867
2021-11-30T15:17:56
1067358457
{ "authors": [ "gmourier" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8190", "repo": "meilisearch/MeiliSearch", "url": "https://github.com/meilisearch/MeiliSearch/issues/1948" }
gharchive/issue
Add missing searchable_attributes.total metric to the Settings Updated telemetry event According to the telemetry specification, it should send searchable_attributes.total when the Settings Updated occurs. Implemented by https://github.com/meilisearch/MeiliSearch/pull/1941
2025-04-01T06:39:32.444928
2022-10-12T21:09:28
1406813576
{ "authors": [ "dolcy", "ferdi05" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8191", "repo": "meilisearch/awesome-meilisearch", "url": "https://github.com/meilisearch/awesome-meilisearch/pull/24" }
gharchive/pull-request
Add Laravel 9/Scout tutorial to Community Guides Hope all is well; a somewhat recent, beginner-friendly Laravel 9/Scout tutorial. Cheers. Pull Request Related issue Fixes #<issue_number> What does this PR do? Add new tutorial to Community Guides PR checklist Please check if your PR fulfills the following requirements: [x ] Does this PR fix an existing issue, or have you listed the changes applied in the PR description (and why they are needed)? [x] Have you read the contributing guidelines? [x] Have you made sure that the title is accurate and descriptive of the changes? Thank you so much for contributing to Meilisearch! Thanks for the addition @dolcy If you are participating in Hacktoberfest, and you would like to receive some swag from Meilisearch too, please complete this form. Hey there @ferdi05 - thank you!
2025-04-01T06:39:32.459565
2022-12-16T21:59:04
1500926210
{ "authors": [ "brunoocasali", "mmachatschek", "wucherpfennig" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8192", "repo": "meilisearch/meilisearch-symfony", "url": "https://github.com/meilisearch/meilisearch-symfony/issues/215" }
gharchive/issue
Case mismatch between loaded and declared class names: "MeiliSearch\Client" vs "Meilisearch\Client". Description See latest release of https://github.com/meilisearch/meilisearch-php/releases/tag/v0.26.1 Environment (please complete the following information): Meilisearch version: 0.26.1 meilisearch-symfony version: 0.9 symfony 6.2.2 Good lord, Symfony is applying custom logic to namespace and class casings 🙄 @mmachatschek can you give a look at #217? I'm also thinking of a way to handle the meili_search.yaml/meilisearch.yml change to not expose it as a breaking change. The bug is fixed and released as 0.10.0 unfortunately, I didn't have enough time to understand better how I could solve the meili_search: to meilisearch: root key in the config file automatically. If someone does, I would love to learn that :) Sorry for the inconvenience.
2025-04-01T06:39:32.468350
2023-02-16T08:48:50
1587274835
{ "authors": [ "curquiza", "mech" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8193", "repo": "meilisearch/meilisearch", "url": "https://github.com/meilisearch/meilisearch/issues/3503" }
gharchive/issue
Not finding documents that has word occurring on both documents Describe the bug I have 2 documents and when I try to query one word from each document, only the first word will trigger document retrieval. I am not sure if this is a fundamental understanding on my part on the relevant document, but I have asked around my colleague and all says the correct results should be both documents will display. Expected behavior When I query for jet and Satellite I am expecting 2 documents to be returned. Screenshots Meilisearch version: v1.0.1 This might be my own misunderstanding of how searching is supposed to work in enterprise/elasticsearch/Algolia/Meilisearch vs the Google Public Internet way of searching. Because when google, the result might be vast and it is showing all documents with "jet" word in it as well as with "satellite" word in it. Hello @mech The results you got are expected, it's not a bug 😊 Indeed, Meilisearch does prefix search, like Algolia. It means, when you type "jets satellite" Meilisearch first tries to retrieve documents containing jets AND satellite, and then, documents with only jets. This is the behavior of one of the ranking rules: the word ranking rules. Unfortunately, you cannot disable this behavior at the moment. You can customize the relevancy in Meilisearch by using matchingStrategy parameter during the search, but so far, you cannot ask for Meilisearch to retrieve all documents containing at least one word of each query word. Pinging our product team @gmourier for your information 😇 Want to ask further. The ordering is like so: jets AND satellite (will find) jets (will still find) satellite and sequence wording (will no longer try to find) The point 3, am I right? If you type jets satellite, yes. I'm not sure we are meaning the same thing for the third point, so let me give you a second example with 3 words. If you type `Hello everyone here", Meilisearch will try to find First, documents containing hello AND everyone AND here Then, documents containing hello AND everyone Finally documents containing hello If you change matchingStrategy to all (default is last and leads to the behavior I described above) Meilisearch will only retrieve documents containing hello AND everyone AND here, nothing else. Tks @curquiza, now I can see why it is called prefix search 😀. Tks for the explanation.
2025-04-01T06:39:32.474782
2024-05-16T15:28:23
2300719164
{ "authors": [ "irevoire" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8194", "repo": "meilisearch/meilisearch", "url": "https://github.com/meilisearch/meilisearch/issues/4639" }
gharchive/issue
[v1.8.0] Relevancy changes Motivation Implementing https://github.com/meilisearch/meilisearch/issues/4484 gives us the ability to finally untie the importance (or weight) of each field with their order. Usage Nothing to do. Changes When no searchable attributes are declared When no searchable attributes are declared, all the fields have the same importance instead of randomly giving more importance to the field we've encountered « the most early » in the life of the index. This means that before this PR, send the following json: [ { "id": 0, "name": "kefir", "color": "white" }, { "id": 1, "name": "white", "last name": "spirit" } ] Would make the field name more important than the field color or last name. This means that searching for white would make the document 1 automatically higher ranked than the document 0. After this PR, all the fields have the same weight, and none are considered more important than others. When a nested field is made searchable The second behavior change that happened with this PR is in the case you're sending this document, for example: { "id": 0, "name": "tamo", "doggo": { "name": "kefir", "surname": "le kef" }, "catto": "gromez" } Previously, defining the searchable attributes as: ["tamo", "doggo", "catto"] was actually defining the « real » searchable attributes in the engine as: ["tamo", "doggo", "catto", "doggo.name", "doggo.surname"], which means that doggo.name and doggo.surname were NOT where the user expected them and had completely different weights than doggo. In this PR all the weights have been unified, and the « real » searchable fields look like this: [ "tamo", "doggo", "doggo.name", "doggo.surname", "catto"] ^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ^^^^^ Weight 0 Weight 1 Weight 2 Impacted teams Implemented in https://github.com/meilisearch/meilisearch/pull/4631
2025-04-01T06:39:32.479944
2019-12-23T12:48:05
541740606
{ "authors": [ "herrmannplatz", "meinaart", "olaf89", "omelhoro" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8195", "repo": "meinaart/cypress-plugin-snapshots", "url": "https://github.com/meinaart/cypress-plugin-snapshots/issues/85" }
gharchive/issue
update button does not work(bug) Describe the bug When snapshot does not match trying to update it from cypress test runner does not work To Reproduce Steps to reproduce the behavior: Click update snapshot button Expected behavior Snapshot is updated Desktop (please complete the following information): OS: Linux & OSX Cypress Version: latest Additional context Add any other context about the problem here. I can't reproduce this, I am also working in OSX. No errors on the command line? I ran into a similar problem. While setting up the plugin i added the following properties in cypress config "env": { ... "cypress-plugin-snapshots": { "imageConfig": { "threshold": 0.1, "thresholdType": "pixel" }, "serverEnabled": true, "serverHost": "localhost", "serverPort": 2121, } }, In this case the websocket connection string is missing the token. When pressing the update button the save snapshot call just fails silently. Hope that helps. It's not documented anywhere but for me it worked to add token property to the config. No idea why but it's something I found from looking at the code. "serverEnabled": true, "serverHost": "localhost", "serverPort": 2121, "token": "1234" It's not documented anywhere but for me it worked to add token property to the config. No idea why but it's something I found from looking at the code. "serverEnabled": true, "serverHost": "localhost", "serverPort": 2121, "token": "1234"
2025-04-01T06:39:32.484114
2023-02-10T21:27:45
1580359576
{ "authors": [ "GalGend", "dror-ziv" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8196", "repo": "meirim-org/meirim", "url": "https://github.com/meirim-org/meirim/issues/651" }
gharchive/issue
Rename old users In the old website, we did not have usernames, and we kindly asked users to add their names. Yet, when we migrated the users from the old website, we renamed them 'a user from the old website' + their old ID. We no longer want that, but rather want to use their own names. This task is about creating a script that updates user names. If the user name contains 'משתמש מהאתר הישן', then we should rename it to their private name if exists. If not, we should rename them just 'user' + id. For example, the following name משתמש מהאתר הישן 533 will be renamed to משתמש 533 Hey there! To address the issue, I have a few questions regarding the script's placement and structure. Where should we place the script within the "meirim-org/meirim" project? Can it function independently, or should it be integrated into the meirim's CLI? In terms of implementation, should the script be a playbook with SQL queries, or a standalone script that connects to the database and performs the necessary operations? Lastly, I'd like to confirm if Python is an acceptable language for this task. Thanks in advance for your clarification! Hi @dror-ziv thanks for reaching out. We use Knex as a migration manager, the migrations files are at server/migration and are generated automatically by knex commands. The script IMO should be a SQL query, and you may see examples on the other migrations files there. For development, as well as for migrations, we use JS. We're here if you have more questions
2025-04-01T06:39:32.501636
2015-10-24T02:38:02
113130096
{ "authors": [ "SubstituteCS", "meladroit", "suloku" ], "license": "isc", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8197", "repo": "meladroit/svdt", "url": "https://github.com/meladroit/svdt/issues/17" }
gharchive/issue
Error when copying a file that is above x Mb I believe this is more or less an issue with memory allocation (and running out of memory) than with svdt it's self. However, assuming we know the memory allocation limit, what's to stop svdt from only allocating (and then copying) those bytes, and then releasing (or reusing) the allocated memory recursively for the length of the file being copied? Additionally, I do not know when the error begins, but it is between 1.1 Mb -> 10 Mb, since 1 Mb works fine, but 10 causes svdt to throw an error. I'm actually a little surprised that you're able to even copy a file that large. 1 MB or 1 Mb? What game are you testing this under? I wouldn't expect many games to even allocate that much space for the save data partition. 1 Megabyte, I tested with Mario Kart 7 (it has a decent amount of storage because of the replays folder.) When I tried the 10 Megabyte file, it causes SVDT to "crash" (tell me I have to exit due to an error), but when relaunched shows the file. I have not tested to see if the 10 MB file is 1:1 correct. I used Standard Units for my sizes (1 Megabyte = 1 * 10^6.) That is quite odd. Last time I tested trying to move too much data, usually the changes didn't end up sticking. It would be interesting to see if the 10 MB file is empty, partially copied, or just plain broken. I'll check after I finish dumping a friend's dead HDD. Also, Sm4sh is 100% working, so :smile: File is copied, but no bytes are written. Does SVDT alloc the entire sizeof the file? Additionally, is it possible to detect free space in the game cart? I think I've figured out a basic way of cart size. Assuming they're all the same per game, we could find the cart size ourselves using recursion. Then store that in a local db used by SVDT (via the title id) and simply compare bytes on cart vs total allowed bytes on cart to get the free bytes. I think we use ctrulib functions (i.e. FS:USER functions) when actually writing files to the save data, but you're right—we do malloc the entire size of the file (within copyFile() in main.c). That's probably not the smartest thing to do, but there should be enough memory available to us to copy most save files. Given that the 10 MB file on save data is empty, I'd say it's just that you ran out of free space—the OpenFile call probably succeeded, and then the WriteFile call probably failed. The save data size is probably the same for any given title. Are you saying to keep writing bytes to the save data until we run out of free space? It seems rather time-intensive to do this for every game. Write 100mb -> on fail move to Write 10mb -> on fail move to Write 1mb -> on fail move go Etc. Until we get to 1byte Sure, but doing this for every game seems a bit much. Keep in mind too that every time you run out of space, the filesystem becomes inoperable and svdt has to restart to access the filesystem properly again. There must be a less brute-force way to do this. @suloku, do you know if we can use FS:GetFreeBytes on the save data archive? [Not able to test it myself at the moment.] I can try to build this myself, If I clone the source directory (and make file) and add the FS:GetFreeBytes snippet, I can compile correct? I have the 3DS toolchain, but my linux box only has make and gcc/g++ If you'd like to clone the git repo in full, and you already have ctrulib, devkitPro, etc., then it should work. Honestly, though, I'd recommend making a smaller test program if all you want is to test that snippet. Can't provide too much more support at the moment—sorry! About fs:getfreebytes, I guess it should work? Profi2000 thinks so too. I'll try later if substituteCS doesn't do it before me. ps: zlib dependency is for pokemon rumble world de/compresion. Posting this because probably edits don't appear as notifications (see my previous comment) Ah, I guess the zlib dependency is something I added on—sorry. I think I confused it with the blowfish stuff, which is in source rather than as another library. 512-byte blocks makes sense (encryption scheme? flash memory sector size? either way ...), so perhaps we should indicate space used/available in multiples of 512 B. I made some test, they really seem to be 512 byte blocks. 512B blocks might be either because of The encryption method is using chunks of 512 bytes (and is padding data), which would be weird since it looks like it's Blowfish, and that uses 8 byte blocks. Or the Cartridge FS format is just really whack? Eitherway, 512B sizes are better since each cartridge is guaranteed to have at least 512 * n where n = ceiling(totalsize / 512)
2025-04-01T06:39:32.509842
2017-02-08T16:59:15
206259256
{ "authors": [ "dpethes", "mellinoe", "wizzard0" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8198", "repo": "mellinoe/ImGui.NET", "url": "https://github.com/mellinoe/ImGui.NET/issues/8" }
gharchive/issue
MIT license? what's the license for this repo? i'm willing to <-> integrate between my imgui wrapper which also boasts software rendering support, but I need the license to be MIT so I can reuse this for my dayjob I've been meaning to put up the MIT license for the project. @dpethes , do you have any objections to that? You and I are the only contributors at this point. Fine by me. Thanks a lot! Um, and could you please add the license file then? :) @wizzard0 Yes, sorry 😄. I will add it later today. https://github.com/mellinoe/ImGui.NET/commit/7cc420b16bc540beb0930714239b7437dad2d5e5
2025-04-01T06:39:32.543438
2023-08-14T10:03:36
1849449240
{ "authors": [ "TGlide", "ivoilic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8199", "repo": "melt-ui/melt-ui", "url": "https://github.com/melt-ui/melt-ui/pull/384" }
gharchive/pull-request
Clarified the warning in introduction.md Updated the under construction warning in the introduction based on this info: https://discord.com/channels/1116682155809067049/1140452054888873994 Once version 1.0 is released, semantic versioning will be followed, (limiting breaking changes to major releases). Until then, breaking changes might happen on non-major releases. ^ this - we're prerelease. We do bump minor/patch versions. So right now, minor could consist of a breaking change where patch changes should not. This makes it a lot clearer where breaking changes might happens and what the next goalpost is. Great idea. Thank you!
2025-04-01T06:39:32.553293
2021-05-04T23:41:26
1251992274
{ "authors": [ "MeltyBot" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8200", "repo": "meltano/hub", "url": "https://github.com/meltano/hub/issues/20" }
gharchive/issue
Add pages for taps, targets, and singer on the hub Migrated from GitLab: https://gitlab.com/meltano/hub/-/issues/21 Originally created by @tayloramurphy on 2021-05-04 23:41:26 We want pages for the following: [x] hub.meltano.com/taps [x] hub.meltano.com/targets [x] hub.meltano.com/singer [x] hub.meltano.com/singer/spec The /singer page will focus on our commitment to the Singer ecosystem and community. We will detail: [x] The Open API we intend to maintain for broader community access to the clean yaml definitions of every tap and target we find [x] Link to the spec [x] Link to the JSON schemas [x] Discuss the Tap and Target SDK with links to resources on those [x] Link to the /taps and /targets pages The /taps and /targets page will: [x] list all of the available taps/targets we're aware of [x] detail how to do a basic pip install [x] give easy links to the YAML definition and how to contribute to it [x] Link to JSON schema [ ] Have a Table of Contents to make Settings more discoverable: https://gitlab.com/meltano/hub/-/issues/30 View 10 previous comments from the original issue on GitLab
2025-04-01T06:39:32.554400
2022-08-12T18:22:32
1337564225
{ "authors": [ "aaronsteers", "edgarrmondragon" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8201", "repo": "meltano/sdk", "url": "https://github.com/meltano/sdk/pull/908" }
gharchive/pull-request
refactor(targets): Allow overriding the bulk insert statement in SQLSink Small improvement towards #9 Scratch my comment about the PR title. It looks great as is. I was reading from my mobile app and was looking in the wrong place.
2025-04-01T06:39:32.555300
2021-12-05T01:01:55
1071342758
{ "authors": [ "jabelone", "mshafer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8202", "repo": "membermatters/MemberMatters", "url": "https://github.com/membermatters/MemberMatters/pull/151" }
gharchive/pull-request
Add mysql install step to macOS setup instructions Hey there, just getting the backend running locally and had to install mysql for the pip install to work. Not sure if universal but some updated docs in case anyone else hits it. Perfect, thank-you!
2025-04-01T06:39:32.557894
2019-11-18T13:13:42
524359674
{ "authors": [ "kanes115", "mat-hek" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8203", "repo": "membraneframework/membrane-core", "url": "https://github.com/membraneframework/membrane-core/pull/216" }
gharchive/pull-request
Fix EOS and unlinking in bins problems This PR fixes: Wrong elements' behaviour on duplicated EndOfStream events Not notifying parents about EndOfStream event Not generating EndOfStream when elements went from :playing to :prepared @mat-hek also tests from test/membrane/integration/bin_test.exs cover some of this (especially the communication of EOS to parents). More tests are quite difficult to accomplish as this is a fix for RC. @kanes115 if they cover, why didn't they catch that? maybe they need a fix? I don't see any RC in the description of the PR Yea, okey, so there's part of the issue that is RC (receiving duplicate EOS events) and not RC part (not generating EOS in the first place).
2025-04-01T06:39:32.591478
2023-07-09T10:43:06
1795341082
{ "authors": [ "shay23b" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8204", "repo": "memphisdev/memphis.net", "url": "https://github.com/memphisdev/memphis.net/issues/96" }
gharchive/issue
Add connection ID and req version in RemoveProducerRequest and RemoveConsumerRequest add to the requests: ConnectionId RequestVersion = 1 duplicate
2025-04-01T06:39:32.631609
2023-02-24T05:56:28
1597951707
{ "authors": [ "ConnorLand", "sonam2207" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8205", "repo": "mendix/docs", "url": "https://github.com/mendix/docs/issues/5688" }
gharchive/issue
part-7-send-your-first-test-push-notification Please use the form below, leaving the prefilled data to help us. Thank you. Page link: part-7-send-your-first-test-push-notification Document link: notif-send-test.md My Issue/Suggestion Hi, i have an issue with the step, if the app is opened, i am not receiving the notification on my phone. It's only working if app is running in background or closed. Hello @sonam2207 , Apologies for the delayed response. The push notification process can be difficult indeed. For development help, I recommend searching or posting a question in the Mendix Forum. If you difficulties persist, please file a Mendix Support Ticket and they will help identify parts of the product that need to be improved. After that, I'll be notified and I will alter the documentation accordingly. Thanks for contacting Mendix Documentation, Connor
2025-04-01T06:39:32.638239
2020-12-28T05:45:06
775209351
{ "authors": [ "Dushusir", "mengshukeji" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8206", "repo": "mengshukeji/Luckysheet", "url": "https://github.com/mengshukeji/Luckysheet/issues/406" }
gharchive/issue
[Feature request]Optimization: The data is too large Is your feature request related to a problem? Please describe. An excel document (200K) with more than 1,000 rows, after being stored as a luckysheet, the db field reached more than 13 million characters (more than 60 times enlarged). Describe the solution you'd like The style of each cell is stored separately and takes up a lot of space. Consider separating repeated cell configurations 问题: 一个1000多行的excel文档(200K),存储成luckysheet后,db字段达到了1300多万个字符(扩大了60多倍)。 解决计划: 每个单元格的样式都是单独存储的,占用了很大空间。考虑分离重复的单元格配置 Duplicate of #418
2025-04-01T06:39:32.661404
2023-02-23T05:57:56
1596269728
{ "authors": [ "luoww1992", "undcloud" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8207", "repo": "menyifang/DCT-Net", "url": "https://github.com/menyifang/DCT-Net/issues/32" }
gharchive/issue
face flicker When the character is not moving, the picture will flicker Could you share some pictures or videos?
2025-04-01T06:39:32.681744
2023-05-24T18:20:00
1724524193
{ "authors": [ "andlaus", "louwersj" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8208", "repo": "mercedes-benz/odxtools", "url": "https://github.com/mercedes-benz/odxtools/pull/129" }
gharchive/pull-request
code cleanup Removed a number of unused import statements Removed a number of unused parameters / variables @andlaus I will have a chat with the "sheriff" and come back with some changes :) this PR is now obsolete: we now run ruff as part of the CI, ensuring that all imported modules are used. closing.
2025-04-01T06:39:32.688445
2017-11-05T23:48:16
271326424
{ "authors": [ "meridius", "pranasblk" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8209", "repo": "meridius/confluence-to-markdown", "url": "https://github.com/meridius/confluence-to-markdown/issues/1" }
gharchive/issue
Does it support to md conversion I'm getting raw html .. in the converted .md file Yes it does. But it is limited by what Pandoc can do. For example it has to leave the table in html format in case any styles or custom attributes are present.
2025-04-01T06:39:32.692845
2021-12-07T06:47:06
1072978879
{ "authors": [ "bjeyabalan", "mrtrkmn", "ping2balaji" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8210", "repo": "merkez/ubuntu-packer", "url": "https://github.com/merkez/ubuntu-packer/issues/1" }
gharchive/issue
ubuntu 20.04 creation using packer got stuck Hi @mrtrkmn , Thanks for this repo with detailed info about creating ubuntu VM with packer. I have tried the steps you have mentioned but got stuck in "waiting for ssh". Tried rdp and following is the output: Please let me know any way to resolve this. Thanks, ...Balaji.J are you trying to create the vm in a vm ? @ping2balaji Hi @mrtrkmn , The issue got resolved after i moved my host os from ubuntu server to ubuntu desktop. Not sure what fixed it though. However im facing following issue with virtualbox now while packer is provisioning the VM: "VBoxManage: error: Details: code NS_ERROR_FAILURE (0x80004005), component MachineWrap, interface IMachine" The VM abruptly stops with above error from VirtualBox and this is happening at random times like after 5mins sometme and after 30mins sometimes. Any input here? I checked internet and updated virtualbox to latest version. My host is Ubuntu 20.04. Please let me know ur suggestion. Thanks, ...Balaji Hi @mrtrkmn , the above issue is fixed after updating the vbox extension to same version as vbox. thanks, ...Balaji
2025-04-01T06:39:32.718008
2022-07-12T17:51:30
1302433575
{ "authors": [ "mert-donmez", "zgtibu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8211", "repo": "mert-donmez/GoogleReviewBot", "url": "https://github.com/mert-donmez/GoogleReviewBot/issues/1" }
gharchive/issue
Can we change to 1*? Can we change to 1*? It is going to review page but not selecting it then it is opening next new window for new id Hi. Sorry for the late reply. for 1 star review you should try this code instead of current code: line 52 elem=WebDriverWait(self.driver, 10).until(EC.element_to_be_clickable((By.CSS_SELECTOR,"#yDmH0d > c-wiz > div > div > div > div > div.O51MUd > div.l5dc7b > div.DTDhxc.eqAW0b > div.euWHWd.aUVumf > div > div:nth-child(1)"))) it may not be working properly after some updates. i will take a look later IndentationError: unindent does not match any outer indentation level It is logging in and going to the review page but not doing anything. it is opening another tab with next id..
2025-04-01T06:39:32.720594
2021-05-02T05:04:22
873839757
{ "authors": [ "mertasan", "tvartom" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8212", "repo": "mertasan/tailwindcss-variables", "url": "https://github.com/mertasan/tailwindcss-variables/issues/4" }
gharchive/issue
Support for PostCSS 7? Would it work to run with PostCSS Thanks for a nice plugin! The plugin is dependent on PostCSS ^8.2.9. Vue (vue-cli) is stil dependent on PostCSS 7, so Tailwindcss needs to be used with Tailwind's PostCSS 7 compatibility build. (Added with npm install -D tailwindcss@npm:@tailwindcss/postcss7-compat postcss@^7 autoprefixer@^9, see https://tailwindcss.com/docs/installation#post-css-7-compatibility-build) Is PostCSS 8 strict needed for this plugin, or could the dependency by changed to be PostCSS 7 or 8 instead? if the tests are successful, I will publish a new version. thank you. @tvartom please upgrade the plugin to v1.1.3 👍
2025-04-01T06:39:32.724923
2017-03-13T11:28:24
213739643
{ "authors": [ "mertenats", "samabsalom" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:8213", "repo": "mertenats/Sonoff", "url": "https://github.com/mertenats/Sonoff/issues/4" }
gharchive/issue
reset/restart error while compiling Hello Sam, I'm having a problem compiling the firmware in the arduino ide. /Users/samabsalom/Downloads/Sonoff-master/sonoff/sonoff.ino: In function 'void reconnect()': sonoff:237: error: 'restart' was not declared in this scope restart(); ^ sonoff:256: error: 'reset' was not declared in this scope reset(); ^ exit status 1 'restart' was not declared in this scope Do you have any idea what it is i'm doing wrong? I can't work it out! so far all i have altered to the standard firmware you provided was to comment out the TLS support Thanks, Sam Hi Sam, I didn't have this problem before, it's strange. I cannot verify it now. Maybe, you can try to declare the two function at the top of the file (before the reconnect() function), with: void reset(); void restart(); Sam yeah that has allowed me to compile it. do you think the restart and reset functions still work ? thanks again Sam Yes! If my knowledge is C are correct ;). Sam I wouldn't doubt that for a second! all up and running now. Thanks so much for your help Sam