added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:35:45.123015
2015-03-30T17:20:07
65259006
{ "authors": [ "cicku", "paulfurley", "toofishes" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11568", "repo": "toofishes/python-pgpdump", "url": "https://github.com/toofishes/python-pgpdump/pull/16" }
gharchive/pull-request
Add entry_point to setup.py This makes pip install pgpdump to install the pgpdump script into the user's $PATH (or .local/bin/) See https://chriswarrick.com/blog/2014/09/15/python-apps-the-right-way-entry_points-and-scripts/ Hi @toofishes is this repo still alive? Yes it is! I've been super busy with other projects though so have been neglecting pull requests here. I'll try and take a look at this sometime this week. On Mar 30, 2015, at 5:12 PM, Paul M Furley<EMAIL_ADDRESS>wrote: Hi @toofishes is this repo still alive? — Reply to this email directly or view it on GitHub. Totally understand! You can test locally by doing cd python-pgpdump; pip install --user -e . - you should see the pgpdump script appear in ~/.local/bin/ :) I'm against this change. This will conflict with the pgpdump from http://www.mew.org/~kazu/proj/pgpdump/en/ in /usr/bin. I maintain both packages in Fedora. Also, this is just a "parser library", I found no reason of having executable. @cicku Fair enough about conflicting namespace. I can't agree that it's just a library though - else why would it offer a main function at all? My understanding of representing main in the code is for tests ONLY, pgpdump imports and execute tests without side effects.
2025-04-01T04:35:45.125188
2017-07-02T02:17:06
239983878
{ "authors": [ "GeorgLink" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11569", "repo": "toolkit-for-ynab/toolkit-for-ynab", "url": "https://github.com/toolkit-for-ynab/toolkit-for-ynab/pull/837" }
gharchive/pull-request
fix default value in accounts-display-density Github Issue (if applicable): #XXX Trello Link (if applicable): Forum Link (if applicable): Explanation of Bugfix/Feature/Enhancement: @dbaldon commented on #828 that the default value should match the default option value. Recommended Release Notes: I'm only fixing this because it is the 'template' I had looked at while working on #828.
2025-04-01T04:35:45.130810
2024-06-09T15:54:06
2342348663
{ "authors": [ "pixtur", "rstecca" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11570", "repo": "tooll3/t3", "url": "https://github.com/tooll3/t3/issues/446" }
gharchive/issue
Bypassed nodes are not bypassed on load Describe the bug Nodes that are marked as bypassed still contribute to the graph when the operator is loaded. Please complete the following information): OS: Win11 Version: 3.9.1 Graphics card and resolution: Nvidia RTX 4070 To Reproduce Create an empty project. Don't delete the template nodes. Add a Transform node between Text and RenderTarget, set an offset (for example change Translation's Y to 0.250) Select the Transform node and mark it for bypass (SHIFT+B) Save Close T3 Open T3 and load the test project. See that despite the Transform is visually marked as bypassed, it still contributes to the graph. Expected behavior On load, bypassed nodes should be initialized so they are effectively bypassed. Screenshots Additional context It doesn't seem to depend on the type of node as I tested with a few and this problem is consistent. Thanks for this excellent bug report. I thought this issues was fixed with f329a2871d5827eb46e08a4be0f51a9bb957bd46 I will look into this. I tried to reproduce this issue with the current dev branch and the issue appears to be fixed for the upcoming release. That's great! Looking forward to the next release! Here you go: https://github.com/tooll3/t3/releases/tag/v3.9.2
2025-04-01T04:35:45.634130
2017-06-05T04:26:44
233491724
{ "authors": [ "kevingriffin", "pyromaniac" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11571", "repo": "toptal/chewy", "url": "https://github.com/toptal/chewy/pull/514" }
gharchive/pull-request
Repairs class check in rake helper. The previous fix broke string index names. Can you add a spec please? :) I was thinking something similar—I'll take a look at that this afternoon. Sorry for the breakage! I'm merging it since it was found out that it breaks our stuff as well, but will wait for specs :)
2025-04-01T04:35:45.653546
2015-05-14T10:21:58
76311542
{ "authors": [ "davecranwell", "gasman", "kaedroho" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11572", "repo": "torchbox/wagtail", "url": "https://github.com/torchbox/wagtail/pull/1303" }
gharchive/pull-request
Reorganised reference section. Also updated requirements, removed conflicting theme css file, updated README What's the rationale for moving the tutorial-style docs into 'reference'? To me, 'reference' implies things like API docs - things that people will refer back to after they've learned their way around the system, not the original learning materials. Maybe we should split the tutorial style docs out of references into a separate "usage guide" section. (this is actually how Djangos docs are structured) Definitely. The way the docs are structured at the moment, the 'How to' section (which is a much more tutorial-y sounding name) appears first, which means we're talking about things like multi-lingual configuration before we've even told users how to create pages.
2025-04-01T04:35:45.657632
2024-11-19T15:37:47
2672605375
{ "authors": [ "EdTheBearded" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11573", "repo": "torizon/meta-toradex-torizon", "url": "https://github.com/torizon/meta-toradex-torizon/pull/178" }
gharchive/pull-request
rac: enable DBus client by default Related-to: TOR-3637 Just checking, do we not need to update the rac recipe as well? I saw on the respective rac PR that we may need to update the packaging: torizon/rac#20 Actually wait that PR isn't even merged yet it seems. Or maybe I'm misunderstanding the connection between these. Yup, you're right! I thought by now this would've been merged, but yeah, we would also need to bump de recipe. I'll leave this PR in draft until it gets merged. @jsrc27 RAC hash has been bumped as part of the monthly release: https://github.com/torizon/meta-toradex-torizon/pull/188 And since the release is now finished, we can go ahead with this MR \o/
2025-04-01T04:35:45.722200
2022-07-12T06:07:29
1301582754
{ "authors": [ "codecov-commenter", "jinlingxu06" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11575", "repo": "towhee-io/towhee", "url": "https://github.com/towhee-io/towhee/pull/1547" }
gharchive/pull-request
add maxvit model Signed-off-by: xujinling<EMAIL_ADDRESS> Codecov Report Merging #1547 (22d4222) into main (89f49a9) will increase coverage by 0.04%. The diff coverage is 84.82%. @@ Coverage Diff @@ ## main #1547 +/- ## ========================================== + Coverage 70.03% 70.07% +0.04% ========================================== Files 351 358 +7 Lines 20535 20650 +115 Branches 3301 3318 +17 ========================================== + Hits 14381 14471 +90 - Misses 5245 5265 +20 - Partials 909 914 +5 Impacted Files Coverage Δ towhee/models/max_vit/max_vit.py 64.51% <64.51%> (ø) towhee/models/layers/mbconv.py 86.95% <100.00%> (+4.34%) :arrow_up: towhee/models/layers/relative_self_attention.py 100.00% <100.00%> (ø) towhee/models/max_vit/__init__.py 100.00% <100.00%> (ø) towhee/models/max_vit/configs.py 100.00% <100.00%> (ø) towhee/models/max_vit/max_vit_block.py 100.00% <100.00%> (ø) towhee/models/max_vit/max_vit_stage.py 100.00% <100.00%> (ø) towhee/models/max_vit/max_vit_utils.py 100.00% <100.00%> (ø) towhee/models/utils/gelu_ignore_parameters.py 100.00% <100.00%> (ø) towhee/models/utils/get_relative_position_index.py 100.00% <100.00%> (ø) ... and 2 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 89f49a9...22d4222. Read the comment docs.
2025-04-01T04:35:45.726763
2016-11-10T23:41:09
188648978
{ "authors": [ "foorjdev", "toxicFork" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11576", "repo": "toxicFork/react-three-renderer", "url": "https://github.com/toxicFork/react-three-renderer/issues/124" }
gharchive/issue
Compatibility with<EMAIL_ADDRESS> I've managed to update r3r to be compatible with the ES6 versions of three.js (specifically 0.82.1, though anything >=0.80.0 should work fine apart from failing tests that check the console output for the exact version, e.g. "THREE.WebGLRenderer 82"). You can take a look at my fork here. The vast majority of modified files only had import THREE from 'three'; changed to import * as THREE from 'three'; since there's no default export now. I also fixed a PropTypes issue having to do with TubeGeometry.FrenetFrames moving to Curve.computeFrenetFrames. Unfortunately I ran into a bit of a stumbling block on two of the tests for TextureLoader. The tests are written to replace THREE.ImageLoader with a stub (for example, imageLoaderLoadStub = sinon.stub(THREE, 'ImageLoader', ImageLoaderMock); in tests/src/descriptors/Texture.jsx), but since the THREE namespace is no longer used internally, the replacement has no effect on the function called internally by TextureLoader. The result is that the ImageLoaderMock test function is never called, leading to an error being thrown by expect(imageLoaderInstances.length).to.equal(1);. If anyone can offer some guidance on how I should proceed, I'd really appreciate it. I have very little experience with test frameworks like mocha, so it's difficult for me to discern the correct approach at times. Cool, good work! :D I really appreciate it! Start a pull request, and I can do PRs to your fork to show how to approach fixing the tests if you like. I should have some time this weekend :)
2025-04-01T04:35:45.766484
2021-09-30T13:07:48
1012158006
{ "authors": [ "scala-steward" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11577", "repo": "tpolecat/natchez", "url": "https://github.com/tpolecat/natchez/pull/428" }
gharchive/pull-request
Update munit-cats-effect-3 to 1.0.6 Updates org.typelevel:munit-cats-effect-3 from 1.0.3 to 1.0.6. GitHub Release Notes - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.typelevel", artifactId = "munit-cats-effect-3" } ] labels: test-library-update, semver-patch Superseded by #459.
2025-04-01T04:35:45.794511
2023-09-07T10:02:21
1885569445
{ "authors": [ "scala-steward" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11578", "repo": "trace4cats/trace4cats-zio", "url": "https://github.com/trace4cats/trace4cats-zio/pull/198" }
gharchive/pull-request
Update sbt-github-actions to 0.16.0 About this PR 📦 Updates com.codecommit:sbt-github-actions from 0.14.2 to 0.16.0 Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "com.codecommit", artifactId = "sbt-github-actions" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "com.codecommit", artifactId = "sbt-github-actions" } }] labels: sbt-plugin-update, early-semver-major, semver-spec-minor, commit-count:n:2 Superseded by #204.
2025-04-01T04:35:45.799967
2023-09-14T04:58:50
1895668936
{ "authors": [ "scala-steward" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11579", "repo": "trace4cats/trace4cats-zio", "url": "https://github.com/trace4cats/trace4cats-zio/pull/202" }
gharchive/pull-request
Update scala-library to 2.13.12 About this PR 📦 Updates org.scala-lang:scala-library from 2.13.8 to 2.13.12 📜 GitHub Release Notes - Version Diff Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scala-lang", artifactId = "scala-library" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "org.scala-lang", artifactId = "scala-library" } }] labels: library-update, early-semver-patch, semver-spec-patch, commit-count:n:2 Superseded by #222.
2025-04-01T04:35:45.822153
2020-03-30T15:20:16
590376233
{ "authors": [ "timocov" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11580", "repo": "tradingview/lightweight-charts", "url": "https://github.com/tradingview/lightweight-charts/pull/384" }
gharchive/pull-request
Refactoring dateFormat Type of PR: enhancement PR checklist: [x] Addresses an existing issue: fixes #368 [x] Includes tests [x] Documentation update Needs to add unit-tests for formatDate function.
2025-04-01T04:35:45.832879
2018-07-09T11:21:39
339408837
{ "authors": [ "ABeltramo", "ItalyPaleAle", "Okeanos", "Queuecumber", "Yajo", "ataraxus", "dduportal", "nmengin", "pxeger", "tpdownes", "vladimirtiukhtin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11581", "repo": "traefik/traefik", "url": "https://github.com/traefik/traefik/issues/3575" }
gharchive/issue
SSL client authentication on per frontend basis Do you want to request a feature or report a bug? Feature What do you expect to see? We need to authenticate clients using SSL on per frontend basis. Let's say I have an ingress defined and CA as kubernetes secret. How can I enable SSL authentication on particular path on that ingress? Thanks Hello @tiukhtinvladimir , Many thanks for your interest in the project. All the TLS configuration is mapped to the entrypoints and not to the frontends but you can make the CA optional in your incoming request as described in the documentation. Thus, you can check if the CA is mandatory (or not) directly into your backends. WDYT? I have the same issue. Would like to enforce client authentication for a single frontend. In my case, the backend application isn't able to validate the client certificate, so making CA optional and passing the certificate to the backend wouldn't help. In short, I have multiple services behind Traefik: some are public, some are private, and for the private ones I'd like Traefik to request a client certificate and validate the client. I agree that this would be helpful for me. Part of the reason is that, even if you make the verification optional, a browser that has certificates in it will prompt you to select one when you first browse to any site on the host. For others, I think a reasonable approach would be to have an entrypoint on a non-standard port like 8443. It's certainly not ideal -- and might not work given firewall realities -- but client certificates are a funny business to be in anyhow. This is wrongly labeled as a question. The answer is "you can't", and this is an important topic, so this should be a feature request IMHO. Is it possible to enforce the TLS by passing optional=false but still make it work with LetsEncrypt? Example config: [entryPoints] [entryPoints.https] address = ":443" [entryPoints.https.tls] [entryPoints.https.tls.ClientCA] files = ["<redacted>"] optional = false [acme] caServer = "https://acme-v02.api.letsencrypt.org/directory" email = "<redacted>" storage = "acme.json" entryPoint = "https" onHostRule = true [acme.tlsChallenge] Because in this case the tls challenge from LetsEncrypt will fail because they are not providing the client certificate. News on this? It's a useful feature and nginx ingress can already do it Hi, for information the version v2.0 of Traefik will have improvements over this, as the tlsOptions have moved in the configuration: https://docs.traefik.io/v2.0/https-tls/overview/ (they are now defined at the Router, which is the renamed - and improved- object that was referred as "frontend" in Traefik v1.x as per https://blog.containo.us/back-to-traefik-2-0-2f9aa17be305#3f17). This new Traefik is currently alpha, so wait and see :) Hey folks, so this I haven't a chance yet to try this out, but it seems to be resolved for traefik >v2? I guess this issue could be closed? @dduportal if you have a link to a working example for client certs per router (frontend in v1) it would be nice if you would share it. Any updates, especially now that 2.0 is de facto? The comments above hint that it's possible in 2.0, so could that be clarified? I stumbled across this as well in the past few days. What I need is the following: The ability to specify mTLS (mutual TLS / client certificate) authentication with distinct CAs for different paths on the same host address. The ability to use RequireAndVerifyClientCert in its current form is insufficient because I want the different paths to be treated differently with separate CAs applying to each. "Overlapping" CAs, i.e. I don't know which of them is actually used for verification, are not a solution to my problem. I also cannot implement mTLS in the backend applications (which would render the mTLS option in Traefik obsolete anyway). Related threads / discussions that I found that also require this or something similar: https://community.traefik.io/t/default-tls-options-without-host/13238 https://community.traefik.io/t/many-found-different-tls-options-for-routers-on-the-same-host-errors/13852 https://community.traefik.io/t/restrict-http-url-path-using-mtls/10650 https://community.traefik.io/t/multiple-mtls-options-based-on-path-rules/4602 #6362 None of these have an actual answer :(
2025-04-01T04:35:45.837062
2016-07-14T11:26:16
165539094
{ "authors": [ "Joe4545", "MadhuriHR", "anhengchangyua", "cooldyj", "jmukiibi", "yumi0629" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11582", "repo": "traex/CalendarListview", "url": "https://github.com/traex/CalendarListview/issues/47" }
gharchive/issue
Calendar Not Displayed The GitHub demo is not displaying Calendar. It is not giving any error or exceptions. I am using Android 4.3. have the same issue Same Issue. @Joe4545 changed the year to 2017, but still it aint working I have the same issue by using Android 4.2 add the following code, and then Calendar shows DayPickerView dayPickerView = (DayPickerView) findViewById(R.id.calendar_picker); dayPickerView.setController(this); @cooldyj I have added the code,but it still doesn't work. I solved it!! Modify the follwing code: @Override public int getMaxYear() { return 2017; } The max year should be bigger than your system year! Change year to 2017 , it still doesn't work. The max year should be bigger than your system year! thx!
2025-04-01T04:35:45.864705
2022-10-20T04:08:43
1415909205
{ "authors": [ "fikimaul", "kdpuvvadi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11583", "repo": "trakt-tools/universal-trakt-scrobbler", "url": "https://github.com/trakt-tools/universal-trakt-scrobbler/issues/208" }
gharchive/issue
FER: Add hotstar and Amazon.in hotstar essentially Disney+ of India. I was work on hotstar.com service. But series not yet work, after all work i will merge to this repo. Maybe wanna test the build. chrome.zip any build for firefox? firefox.zip this, but i never test the firefox build Hi @kdpuvvadi i've new build and fix the series on hotstar. Visit https://github.com/trakt-tools/universal-trakt-scrobbler/pull/212 to test the build
2025-04-01T04:35:45.866817
2022-05-13T06:49:07
1234812268
{ "authors": [ "gkHoplite" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11584", "repo": "tranek/GASDocumentation", "url": "https://github.com/tranek/GASDocumentation/issues/79" }
gharchive/issue
Any plan to add Tags for UE4? Do you have a plan to add "git-project-tag" for UE 4.27? After converting to UE5, It's quite hard to get UE4 version. I think some beginners would have to pay lots of time on this Someone struggling with How to use this project on UE4, Use this command git clone git reset --hard 756524f2055a6b9964e5af4bb417201b4666f796 you can see above hash (756524....) on git log Anyway, thank you for making this opensource-project. Awesome! Oh you have branch for this, sorry. I am used to it so much ( It support downloads on git hub) so didn't check that never mind.
2025-04-01T04:35:45.868186
2019-09-16T12:56:47
494033013
{ "authors": [ "sgu-fai", "sheshankdebugged" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11585", "repo": "tranhungt/okcupidjs", "url": "https://github.com/tranhungt/okcupidjs/issues/48" }
gharchive/issue
who-you-like Hello, is it possible to get the list of users in: https://www.okcupid.com/who-you-like Thank you Same query
2025-04-01T04:35:45.888102
2018-12-13T07:15:05
390547446
{ "authors": [ "stangah", "tinyels" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11586", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/1457" }
gharchive/pull-request
WIP: Office invoice flow [delivers #161689746] Description Kind of a lot going on here, sorry: Adds GetShipmentInvoices endpoint, retrieves all invoices with a matching ShipmentID Pulls code into a DeliverAndPriceShipment service, so that action can be performed on a test shipment when standing up data using db_populate_e2e Makes sure our test data doesn't have dates that fall on a weekend (this can blow up the rate engine I think) Adds Invoice to our normalized entity schema Adds selectors to fetch shipmentLineItems for a given invoice ID Refactors some of our invoice components a bit: InvoicePanel still does most of the orchestration LineItemTable just shows a table of line items and a total UnbilledTable has both the "Approve Payment" widget and a LineItemTable of unbilled line items InvoiceTable shows invoice info and a LineItemTable of line items for that invoice InvoicePayment uses both internal component state and entities methods to manage UI Note: Until #1451 merges, you'll notice that all accessorials are marked as invoiced, whether they are approved or priced. This PR doesn't go into fixing that. Missing: React testing around invoice components Setup Part of this PR is making sure that the move with locator "DOOB" (not my name :p) has all the expected ShipmentLineItems priced and ready to go to be invoiced, so test on that. You'll need ot re-run: make db_populate_e2e Code Review Verification Steps [ ] Code follows the guidelines for Logging [ ] The requirements listed in Querying the Database Safely have been satisfied. Any new migrations/schema changes: [ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys) [ ] Have been communicated to #dp3-engineering [ ] There are no aXe warnings for UI. [ ] This works in IE. Any new client dependencies (Google Analytics, hosted libraries, CDNs, etc) have been: [ ] Communicated to @willowbl00 [ ] Added to the list of network dependencies [ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores) [ ] Request review from a member of a different team. [ ] Have the Pivotal acceptance criteria been met for this change? References Pivotal story for this change Screenshots This looks great! Let's create a chore for adding more tests and cleaning up the dead code.
2025-04-01T04:35:45.889552
2020-01-29T21:59:14
557141636
{ "authors": [ "chrisgilmerproj" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11587", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/3429" }
gharchive/pull-request
Update certificates for move.mil from DISA Description Similar to #1014 and #1042 - this is an update to the existing certificates. It ought to operate as a no-op for us but will also ensure new partners connecting over mTLS can talk to us. @rdhariwal - just realized I didn't deploy this to Experimental. I'll do that first before merging.
2025-04-01T04:35:45.892944
2020-11-17T23:03:42
745152198
{ "authors": [ "robot-mymove", "travelar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11588", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/5256" }
gharchive/pull-request
MB-5238: Pushing build failures to #prac-engineering vs. #on-call Description A minor revision to change the CI Build failure notifications to #prac-engineering vs. #on-call. Once this is merged, I will update the slack URL in the circleci project config. Code Review Verification Steps [x] Request review from a member of a different team. [x] Have the Jira acceptance criteria been met for this change? References Jira story for this change Messages :book: :link: MB-5238 Generated by :no_entry_sign: dangerJS against 3f525b6fe8ef097b092327153cdbbfd9963eb5ac
2025-04-01T04:35:45.896621
2022-01-25T00:56:19
1113302320
{ "authors": [ "robot-mymove", "ronaktruss" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11589", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/8076" }
gharchive/pull-request
MB-11140: WIP draft for PPM DB ADR MB-11140 Summary After doing some research on the proposed ideas, there is still no clear best choice. I tentatively have a slight preference for option 3 - Creating a new table. I have two open questions that will likely be the deciding factors in this decision. What is the future scope in terms of new shipments? What is the likelihood of having bandwidth and prioritizing updating the db to match the newly proposed pattern? Is an ADR the best medium for this decision? It seems appropriate to me, but the lack of previous DB specific decisions caught my eye. It seems like many of the backend decisions were captured in what is now the backend mymove docs. Verification Steps for Reviewers These are to be checked by a reviewer. Database Any new migrations/schema changes: [ ] Have been communicated to #g-database Messages :book: :link: MB-11140 Generated by :no_entry_sign: dangerJS against bcf418b813fc17b4ee56426fd09b01b0a88055a9
2025-04-01T04:35:45.909320
2022-06-13T16:56:33
1269709818
{ "authors": [ "duncan-truss", "felipe-lee", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11590", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/8758" }
gharchive/pull-request
[MB-1200] PPM closeout Weight Tickets page component Jira ticket for this change Summary Is there anything you would like reviewers to give additional scrutiny? this article explains more about the approach used. Setup to Run Your Code 💻 You will need to use three separate terminals to test this locally. Terminal 1 Start the Storybook locally. make storybook Terminal 2 Start the UI locally. make client_run Terminal 3 Start the Go server locally. make server_run Additional steps Access the Login as a Verification Steps for Author These are to be checked by the author. [ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores) [ ] Request review from a member of a different team. [ ] Have the Jira acceptance criteria been met for this change? Verification Steps for Reviewers These are to be checked by a reviewer. Frontend [ ] User facing changes have been reviewed by design. [ ] There are no aXe warnings for UI. [ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge). [ ] There are no new console errors in the browser devtools [ ] There are no new console errors in the test output [ ] If this PR adds a new component to Storybook, it ensures the component is fully responsive, OR if it is intentionally not, a wrapping div using the officeApp class or custom min-width styling is used to hide any states the would not be visible to the user. Backend [ ] Code follows the guidelines for Logging [ ] The requirements listed in Querying the Database Safely have been satisfied. Database Any new migrations/schema changes: [ ] Follows our guidelines for Zero-Downtime Deploys [ ] Have been communicated to #g-database [ ] Secure migrations have been tested following the instructions in our docs Screenshots Warnings :warning: This PR does not include changes to unit tests, even though it affects app code. :warning: Files located in legacy directories (src/shared or src/scenes) have been edited. Are you sure you don’t want to also relocate them to the new file structure? View the frontend file org ADR for more information Messages :book: :link: MB-1200 Generated by :no_entry_sign: dangerJS against cb5ab9410885e0b348d3cc865175d205acc80933 Ah, I meant to put in my review comment that none of my comments or questions are blocking, but mainly because some will maybe be easier to clear out once we do the full integration piece with the BE changes in place.
2025-04-01T04:35:45.912393
2023-01-19T16:38:02
1549527204
{ "authors": [ "ahobson", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11591", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/9919" }
gharchive/pull-request
Disable unneeded logging in iampostgres to avoid test race condition Jira ticket for this change Summary Remove unnecessary logging that causes race conditions in the server tests Setup to Run Your Code 💻 You will need to use one terminal to test this locally. Terminal 1 go test ./pkg/iampostgres/... Warnings :warning: Please add the JIRA issue key to the PR title (e.g. MB-123) Generated by :no_entry_sign: dangerJS against c3fac88854482af0a3689f4b4c289ee6dfa84bf2
2025-04-01T04:35:45.925787
2016-10-07T21:28:26
181771451
{ "authors": [ "irees", "laidig" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11592", "repo": "transitland/mapzen-gtfs", "url": "https://github.com/transitland/mapzen-gtfs/pull/23" }
gharchive/pull-request
enables zipfile compression by default. I spotted a minor issue with the make_zip function-- zipfile needs to be told to deflate explicitly. Looks good, and I appreciate the test coverage! Thanks! Thanks. If you can make a release with this next week, I'd appreciate it. I'll make a new release next week. Thanks again for your contributions.
2025-04-01T04:35:45.927366
2022-01-13T03:11:28
1101119524
{ "authors": [ "mathcolo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11593", "repo": "transitmatters/t-performance-dash", "url": "https://github.com/transitmatters/t-performance-dash/pull/132" }
gharchive/pull-request
Version bump tsunami (but primarily Python 3.9) ⚠️ For after bus goes in! Python 3.9 node 16 Make sure Pipfile and requirements.txt match @nathan-weinberg nice ^ lol Closing because we're going to do python and js parts separately per @austinjpaul
2025-04-01T04:35:46.020778
2018-10-20T13:43:27
372213158
{ "authors": [ "arron-green", "codecov-io", "travisbrown" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11604", "repo": "travisbrown/catbird", "url": "https://github.com/travisbrown/catbird/pull/75" }
gharchive/pull-request
bumps finagle to 18.10.0 Bumps finagle to 18.10.0 to match version in upcoming finagle/finch 0.25.0 release https://github.com/finagle/finch/pull/1007 Codecov Report Merging #75 into master will increase coverage by 33.7%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #75 +/- ## ========================================== + Coverage 52.47% 86.17% +33.7% ========================================== Files 9 9 Lines 202 123 -79 Branches 3 3 ========================================== Hits 106 106 + Misses 96 17 -79 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 68f1f66...e5006b6. Read the comment docs. Excellent, thanks @arron-green! Release is on its way to Maven Central: https://github.com/travisbrown/catbird/releases/tag/v18.10.0 @travisbrown thank you!
2025-04-01T04:35:46.044618
2018-04-03T21:14:07
311001149
{ "authors": [ "candlerb", "justone", "travisjeffery" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11605", "repo": "travisjeffery/jocko", "url": "https://github.com/travisjeffery/jocko/issues/111" }
gharchive/issue
Cluster example out of date The cluster example uses command line flags which are no longer valid: Error: unknown flag: --debug Error: unknown flag: --log-dir (should be --data-dir ?) Error: unknown flag: --prometheus-addr Error: unknown flag: --serf-members (should be --join or --join-wan ?) So I tried running it like this: ./jocko broker --data-dir=/tmp/jocko0 --broker-addr=<IP_ADDRESS>:9001 --raft-addr=<IP_ADDRESS>:9002 --serf-addr=<IP_ADDRESS>:9003 --id=1 >broker0.out 2>&1 & ./jocko broker --data-dir=/tmp/jocko1 --broker-addr=<IP_ADDRESS>:9101 --raft-addr=<IP_ADDRESS>:9102 --serf-addr=<IP_ADDRESS>:9103 --join=<IP_ADDRESS>:9003 --id=2 >broker1.out 2>&1 & ./jocko broker --data-dir=/tmp/jocko2 --broker-addr=<IP_ADDRESS>:9201 --raft-addr=<IP_ADDRESS>:9202 --serf-addr=<IP_ADDRESS>:9203 --join=<IP_ADDRESS>:9003 --id=3 >broker2.out 2>&1 & These options are accepted; but no broker is listening on ports 9001, 9101 or 9201, nor is serf listening on 9003, 9103 or 9203. # netstat -natp Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 <IP_ADDRESS>:9002 <IP_ADDRESS>:* LISTEN 1413/jocko tcp 0 0 <IP_ADDRESS>:9102 <IP_ADDRESS>:* LISTEN 1422/jocko tcp 0 0 <IP_ADDRESS>:9202 <IP_ADDRESS>:* LISTEN 1430/jocko tcp 0 0 <IP_ADDRESS>:22 <IP_ADDRESS>:* LISTEN 436/sshd tcp6 0 0 :::8301 :::* LISTEN 1413/jocko tcp6 0 0 :::36275 :::* LISTEN 1413/jocko tcp6 0 0 :::22 :::* LISTEN 436/sshd # netstat -naup Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name udp 0 0 <IP_ADDRESS>:51041 <IP_ADDRESS>:6831 ESTABLISHED 1413/jocko udp 0 0 <IP_ADDRESS>:68 <IP_ADDRESS>:* 250/dhclient udp 0 0 <IP_ADDRESS>:41643 <IP_ADDRESS>:6831 ESTABLISHED 1430/jocko udp 0 0 <IP_ADDRESS>:43866 <IP_ADDRESS>:6831 ESTABLISHED 1422/jocko udp6 0 0 :::8301 :::* 1413/jocko Captured output: ==> broker0.out <== 2018/04/03 20:58:05 Initializing logging reporter 2018-04-03T20:58:05.431Z INFO jocko/broker.go:107 hello {"id": 1, "broker addr": "", "serf addr": "<IP_ADDRESS>:9003", "raft addr": "<IP_ADDRESS>:9002", "id": 0, "raft addr": "<IP_ADDRESS>:9002"} 2018/04/03 20:58:05 [INFO] raft: Initial configuration (index=0): [] 2018/04/03 20:58:05 [INFO] raft: Node at <IP_ADDRESS>:9002 [Follower] entering Follower state (Leader: "") 2018/04/03 20:58:05 [INFO] serf: EventMemberJoin: builder :: 2018/04/03 20:58:05 [WARN] serf: Failed to re-join any previously known node 2018-04-03T20:58:05.528Z INFO jocko/serf.go:66 adding LAN server {"id": 1, "broker addr": "", "serf addr": "<IP_ADDRESS>:9003", "raft addr": "<IP_ADDRESS>:9002", "id": 0, "raft addr": "<IP_ADDRESS>:9002", "meta": {"ID":0,"Name":"","Bootstrap":false,"Expect":0,"NonVoter":false,"Status":1,"RaftAddr":"<IP_ADDRESS>:9002","SerfLANAddr":"%!b(string=<IP_ADDRESS>:9003):8301","BrokerAddr":"<IP_ADDRESS>:9001"}} 2018-04-03T20:58:05.529Z INFO jocko/server.go:71 hello {"id": 1, "broker addr": "", "serf addr": "<IP_ADDRESS>:9003", "raft addr": "<IP_ADDRESS>:9002", "node id": 0, "addr": "<IP_ADDRESS>:9001"} 2018/04/03 20:58:06 [WARN] raft: no known peers, aborting election ==> broker1.out <== 2018/04/03 20:58:43 Initializing logging reporter 2018-04-03T20:58:43.325Z INFO jocko/broker.go:107 hello {"id": 2, "broker addr": "", "serf addr": "<IP_ADDRESS>:9103", "raft addr": "<IP_ADDRESS>:9102", "id": 0, "raft addr": "<IP_ADDRESS>:9102"} ==> broker2.out <== 2018/04/03 20:59:13 Initializing logging reporter 2018-04-03T20:59:13.172Z INFO jocko/broker.go:107 hello {"id": 3, "broker addr": "", "serf addr": "<IP_ADDRESS>:9203", "raft addr": "<IP_ADDRESS>:9202", "id": 0, "raft addr": "<IP_ADDRESS>:9202"} I tried running the first process under strace. Here are all the lines matching htons: connect(4, {sa_family=AF_INET, sin_port=htons(6831), sin_addr=inet_addr("<IP_ADDRESS>")}, 16) = 0 getsockname(4, {sa_family=AF_INET, sin_port=htons(59060), sin_addr=inet_addr("<IP_ADDRESS>")}, [16]) = 0 getpeername(4, {sa_family=AF_INET, sin_port=htons(6831), sin_addr=inet_addr("<IP_ADDRESS>")}, [16]) = 0 bind(6, {sa_family=AF_INET, sin_port=htons(9002), sin_addr=inet_addr("<IP_ADDRESS>")}, 16) = 0 getsockname(6, {sa_family=AF_INET, sin_port=htons(9002), sin_addr=inet_addr("<IP_ADDRESS>")}, [16]) = 0 bind(10, {sa_family=AF_INET6, sin6_port=htons(0), inet_pton(AF_INET6, "::1", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, 28) = 0 bind(11, {sa_family=AF_INET6, sin6_port=htons(0), inet_pton(AF_INET6, "::ffff:<IP_ADDRESS>", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, 28) = 0 bind(10, {sa_family=AF_INET6, sin6_port=htons(8301), inet_pton(AF_INET6, "::", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, 28) = 0 getsockname(10, {sa_family=AF_INET6, sin6_port=htons(8301), inet_pton(AF_INET6, "::", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, [28]) = 0 bind(11, {sa_family=AF_INET6, sin6_port=htons(8301), inet_pton(AF_INET6, "::", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, 28) = 0 getsockname(11, {sa_family=AF_INET6, sin6_port=htons(8301), inet_pton(AF_INET6, "::", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, [28]) = 0 bind(12, {sa_family=AF_INET6, sin6_port=htons(0), inet_pton(AF_INET6, "::", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, 28) = 0 getsockname(12, {sa_family=AF_INET6, sin6_port=htons(45202), inet_pton(AF_INET6, "::", &sin6_addr), sin6_flowinfo=0, sin6_scope_id=0}, [28]) = 0 (I don't see any attempt to open ports 9001 or 9003?) Here are the lines matching = -1: access("/etc/ld.so.nohwcap", F_OK) = -1 ENOENT (No such file or directory) access("/etc/ld.so.preload", R_OK) = -1 ENOENT (No such file or directory) access("/etc/ld.so.nohwcap", F_OK) = -1 ENOENT (No such file or directory) access("/etc/ld.so.nohwcap", F_OK) = -1 ENOENT (No such file or directory) epoll_ctl(5, EPOLL_CTL_ADD, 4, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188013312, u64=140576467611392}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 4, 0xc420055a8c) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_ADD, 4, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188013312, u64=140576467611392}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 4, 0xc42005599c) = -1 EPERM (Operation not permitted) newfstatat(AT_FDCWD, "/etc/mdns.allow", 0xc4200209f8, 0) = -1 ENOENT (No such file or directory) epoll_ctl(5, EPOLL_CTL_ADD, 4, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188013312, u64=140576467611392}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 4, 0xc420055654) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_ADD, 8, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188012896, u64=140576467610976}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 8, 0xc42019b0a4) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_ADD, 9, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188012896, u64=140576467610976}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 9, 0xc42019afec) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_ADD, 9, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188012896, u64=140576467610976}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 9, 0xc42019abb4) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_ADD, 9, {EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLET, {u32=2188012896, u64=140576467610976}}) = -1 EPERM (Operation not permitted) epoll_ctl(5, EPOLL_CTL_DEL, 9, 0xc42019afe4) = -1 EPERM (Operation not permitted) The EPERM issues are a bit worrying. Maybe this is a symptom of running within an lxd container (but then again, running in a docker container is supposed to work) Just tried it again. Unfortunately I cannot get either a single-node or multi-node setup running. Single node # ./jocko broker 2018/06/16 21:06:36 Initializing logging reporter 2018-06-16T21:06:36.095Z INFO jocko/broker.go:109 hello {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>:9094", "raft addr": "<IP_ADDRESS>:9093", "id": 0, "raft addr": "<IP_ADDRESS>:9093"} 2018/06/16 21:06:36 [INFO] raft: Initial configuration (index=0): [] 2018/06/16 21:06:36 [INFO] raft: Node at <IP_ADDRESS>:9093 [Follower] entering Follower state (Leader: "") 2018/06/16 21:06:36 [INFO] serf: EventMemberJoin: jocko :: 2018-06-16T21:06:36.138Z INFO jocko/server.go:71 hello {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>:9094", "raft addr": "<IP_ADDRESS>:9093", "server id": 0, "addr": "<IP_ADDRESS>:9092"} 2018-06-16T21:06:36.139Z INFO jocko/serf.go:74 adding LAN server {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>:9094", "raft addr": "<IP_ADDRESS>:9093", "id": 0, "raft addr": "<IP_ADDRESS>:9093", "meta": {"ID":0,"Name":"","Bootstrap":false,"Expect":0,"NonVoter":false,"Status":1,"RaftAddr":"<IP_ADDRESS>:9093","SerfLANAddr":"<IP_ADDRESS>:9094:8301","BrokerAddr":"<IP_ADDRESS>:9092"}} 2018/06/16 21:06:37 [WARN] raft: no known peers, aborting election I spy something dubious there: "SerfLANAddr":"<IP_ADDRESS>:9094:8301" (8301 exists in the source code as DefaultLANSerfPort) In another screen I try to create a topic: # ./jocko topic create --topic test error code: not controller Back in the broker screen I see: 2018/06/16 21:06:56 Reporting span 58a4a97d0a5b8bfe:152681c6a74b2a88:58a4a97d0a5b8bfe:1 2018/06/16 21:06:56 Reporting span 58a4a97d0a5b8bfe:29f53d9eec662474:58a4a97d0a5b8bfe:1 2018/06/16 21:06:56 Reporting span 58a4a97d0a5b8bfe:3ab651b5a79db080:58a4a97d0a5b8bfe:1 2018/06/16 21:06:56 Reporting span 58a4a97d0a5b8bfe:6215d9d3872fa0c8:58a4a97d0a5b8bfe:1 2018/06/16 21:06:56 Reporting span 58a4a97d0a5b8bfe:6cd74c36fd162c6:58a4a97d0a5b8bfe:1 2018/06/16 21:06:56 Reporting span 58a4a97d0a5b8bfe:58a4a97d0a5b8bfe:0:1 2018/06/16 21:07:12 Reporting span 8304d1cf33b493f:48fb487a39f54e3:8304d1cf33b493f:1 2018/06/16 21:07:12 Reporting span 8304d1cf33b493f:3119779192068843:8304d1cf33b493f:1 2018/06/16 21:07:12 Reporting span 8304d1cf33b493f:14c0ac963bda4e32:8304d1cf33b493f:1 2018/06/16 21:07:12 Reporting span 8304d1cf33b493f:61ef62b06b1e53f3:8304d1cf33b493f:1 2018/06/16 21:07:12 Reporting span 8304d1cf33b493f:41950f95b4c41d52:8304d1cf33b493f:1 2018/06/16 21:07:12 Reporting span 8304d1cf33b493f:8304d1cf33b493f:0:1 2018/06/16 21:07:13 ERROR: error when flushing the buffer: write udp <IP_ADDRESS>:60767-><IP_ADDRESS>:6831: write: connection refused I don't know what's supposed to be listening on port 6831; this number doesn't appear in the Jocko source code anywhere. And indeed nothing is listening on this port, although jocko has a connected UDP socket to send to 6831: # netstat -naup (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name udp 0 0 <IP_ADDRESS>:60767 <IP_ADDRESS>:6831 ESTABLISHED 15769/jocko udp 0 0 <IP_ADDRESS>:68 <IP_ADDRESS>:* 256/dhclient udp6 0 0 :::8301 :::* 15769/jocko Cluster The instructions in _examples/cluster/README.md have invalid flags. I changed them to: # ./jocko broker \ --data-dir="/tmp/jocko0" \ --broker-addr=<IP_ADDRESS>:9001 \ --raft-addr=<IP_ADDRESS>:9002 \ --serf-addr=<IP_ADDRESS>:9003 \ --id=1 2018/06/16 21:23:41 Initializing logging reporter 2018-06-16T21:23:41.179Z INFO jocko/broker.go:109 hello {"id": 1, "broker addr": "<IP_ADDRESS>:9001", "serf addr": "<IP_ADDRESS>:9003", "raft addr": "<IP_ADDRESS>:9002", "id": 1, "raft addr": "<IP_ADDRESS>:9002"} 2018/06/16 21:23:41 [INFO] raft: Initial configuration (index=0): [] 2018/06/16 21:23:41 [INFO] raft: Node at <IP_ADDRESS>:9002 [Follower] entering Follower state (Leader: "") 2018/06/16 21:23:41 [INFO] serf: EventMemberJoin: jocko :: 2018-06-16T21:23:41.223Z INFO jocko/server.go:71 hello {"id": 1, "broker addr": "<IP_ADDRESS>:9001", "serf addr": "<IP_ADDRESS>:9003", "raft addr": "<IP_ADDRESS>:9002", "server id": 1, "addr": "<IP_ADDRESS>:9001"} 2018-06-16T21:23:41.226Z INFO jocko/serf.go:74 adding LAN server {"id": 1, "broker addr": "<IP_ADDRESS>:9001", "serf addr": "<IP_ADDRESS>:9003", "raft addr": "<IP_ADDRESS>:9002", "id": 1, "raft addr": "<IP_ADDRESS>:9002", "meta": {"ID":1,"Name":"","Bootstrap":false,"Expect":0,"NonVoter":false,"Status":1,"RaftAddr":"<IP_ADDRESS>:9002","SerfLANAddr":"<IP_ADDRESS>:9003:8301","BrokerAddr":"<IP_ADDRESS>:9001"}} 2018/06/16 21:23:43 [WARN] raft: no known peers, aborting election (Note: same problem with SerfLANAddr having two ports) In another screen, trying to add a second broker: # ./jocko broker \ --data-dir="/tmp/jocko1" \ --broker-addr=<IP_ADDRESS>:9101 \ --raft-addr=<IP_ADDRESS>:9102 \ --serf-addr=<IP_ADDRESS>:9103 \ --join=<IP_ADDRESS>:9003 \ --id=2 2018/06/16 21:24:31 Initializing logging reporter 2018-06-16T21:24:31.488Z INFO jocko/broker.go:109 hello {"id": 2, "broker addr": "<IP_ADDRESS>:9101", "serf addr": "<IP_ADDRESS>:9103", "raft addr": "<IP_ADDRESS>:9102", "id": 2, "raft addr": "<IP_ADDRESS>:9102"} 2018/06/16 21:24:31 [INFO] raft: Initial configuration (index=0): [] 2018/06/16 21:24:31 [INFO] raft: Node at <IP_ADDRESS>:9102 [Follower] entering Follower state (Leader: "") error starting broker: Failed to create memberlist: Could not set up network transport: Failed to start TCP listener on "<IP_ADDRESS>:9103" port 8301: listen tcp :8301: bind: address already in use # This one exits because it tries to bind to 8301; that port is already in use by the first process. Analysis It seems that meta.SerfLANAddr / serf_lan_addr is assembled from SerfLANConfig.MemberlistConfig.BindAddr and SerfLANConfig.MemberlistConfig.BindPort $ grep -R serf_lan_addr . ./jocko/leader.go: "serf_lan_addr": meta.SerfLANAddr, ./jocko/metadata/metadata.go: SerfLANAddr: m.Tags["serf_lan_addr"], ./jocko/serf.go: config.Tags["serf_lan_addr"] = fmt.Sprintf("%s:%d", b.config.SerfLANConfig.MemberlistConfig.BindAddr, b.config.SerfLANConfig.MemberlistConfig.BindPort) However, BindAddr defaults to both address and port: ./cmd/jocko/main.go: brokerCmd.Flags().StringVar(&brokerCfg.SerfLANConfig.MemberlistConfig.BindAddr, "serf-addr", "<IP_ADDRESS>:9094", "Address for Serf to bind on") // TODO: can set addr alone or need to set bind port separately? And BindPort defaults to 8301, and AFAICS cannot be overridden. $ grep -R DefaultLANSerfPort . ./jocko/config/config.go: DefaultLANSerfPort = 8301 ./jocko/config/config.go: conf.SerfLANConfig.MemberlistConfig.BindPort = DefaultLANSerfPort ... although in the test suite, it is set explicitly: ./jocko/testing.go: config.SerfLANConfig.MemberlistConfig.BindPort = ports[2] ./jocko/testing.go: s1.config.SerfLANConfig.MemberlistConfig.BindPort) ./testutil/testutil.go: config.SerfLANConfig.MemberlistConfig.BindPort = ports[1] I can't see how this can possibly work outside the test suite. What I can do is force --serf-addr=<IP_ADDRESS> at which point at least we don't have duplicate ports in SerfLANAddr: # ./jocko broker --serf-addr=<IP_ADDRESS> 2018/06/16 21:31:42 Initializing logging reporter 2018-06-16T21:31:42.136Z INFO jocko/broker.go:109 hello {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>", "raft addr": "<IP_ADDRESS>:9093", "id": 0, "raft addr": "<IP_ADDRESS>:9093"} 2018/06/16 21:31:42 [INFO] raft: Initial configuration (index=0): [] 2018/06/16 21:31:42 [INFO] raft: Node at <IP_ADDRESS>:9093 [Follower] entering Follower state (Leader: "") 2018/06/16 21:31:42 [INFO] serf: EventMemberJoin: jocko <IP_ADDRESS> 2018-06-16T21:31:42.170Z INFO jocko/server.go:71 hello {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>", "raft addr": "<IP_ADDRESS>:9093", "server id": 0, "addr": "<IP_ADDRESS>:9092"} 2018-06-16T21:31:42.175Z INFO jocko/serf.go:74 adding LAN server {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>", "raft addr": "<IP_ADDRESS>:9093", "id": 0, "raft addr": "<IP_ADDRESS>:9093", "meta": {"ID":0,"Name":"","Bootstrap":false,"Expect":0,"NonVoter":false,"Status":1,"RaftAddr":"<IP_ADDRESS>:9093","SerfLANAddr":"<IP_ADDRESS>:8301","BrokerAddr":"<IP_ADDRESS>:9092"}} 2018/06/16 21:31:42 [WARN] serf: Failed to re-join any previously known node 2018/06/16 21:31:43 [WARN] raft: no known peers, aborting election However it still fails in the same same way as single node cluster (client says error code: not controller; broker fails writing to UDP port 6831) P.S. Looking in the source code of serf itself, it uses a helper to split addr:port into the separate components of MemberlistConfig func (c *Command) setupAgent(config *Config, logOutput io.Writer) *Agent { bindIP, bindPort, err := config.AddrParts(config.BindAddr) ... serfConfig.MemberlistConfig.BindAddr = bindIP serfConfig.MemberlistConfig.BindPort = bindPort And I found port 6831 in jaeger-client-go. Since this is for OpenTracing, the failure to send to this UDP port may not matter. It would of course be nice to turn off when not needed. ./vendor/github.com/uber/jaeger-client-go/transport_udp.go:const defaultUDPSpanServerHostPort = "localhost:6831" Working on fixes in #133 Thanks for working on this. I ran into the same issue with the ports conflicting. Let me know if I can help with testing or code review. Current status: you can start a one-node cluster with jocko broker --bootstrap --bootstrap-expect=1, and create a topic with jocko topic create --topic <name>. When I try to publish a message with confluent-kafka-python, it fails with the following error: %3|1529616101.104|PROTOERR|rdkafka#producer-1| [thrd:main]: localhost:9092/bootstrap: Protocol parse failure at 31/70 (rd_kafka_parse_Metadata:306) (incorrect broker.version.fallback?) %3|1529616101.104|PROTOERR|rdkafka#producer-1| [thrd:main]: localhost:9092/bootstrap: 65536 topics: tmpabuf memory shortage %4|1529616101.104|METADATA|rdkafka#producer-1| [thrd:main]: localhost:9092/bootstrap: Metadata request failed: connected: Local: Bad message format (1ms): Permanent You can start multiple nodes with e.g. --bootstrap-expect=3, but the cluster won't come up because the --join option currently does nothing. (I still haven't worked out why jocko needs both raft and serf. Maybe it's to allow a cluster where only a subset of nodes store the raft commit log?) Cluster startup now kind-of working: serf needs to have a unique node name, so I added a JOCKONODENAME environment variable to override it. This is something which should rarely be used, so I didn't make it a command line flag. There seems to be a problem with negative message transit times (!) 2018/06/22 08:04:21 [DEBUG] serf: messageJoinType: jocko1 2018/06/22 08:04:21 [DEBUG] serf: messageJoinType: jocko1 2018/06/22 08:04:21 [DEBUG] serf: messageJoinType: jocko1 2018/06/22 08:04:21 [ERR] serf: Rejected coordinate from jocko0: round trip time not in valid range, duration -7.035µs is not a positive value less than 10s And the client still has to know which node to connect to: # cmd/jocko/jocko topic create --topic weeble --broker-addr <IP_ADDRESS>:9201 error code: not controller # cmd/jocko/jocko topic create --topic weeble --broker-addr <IP_ADDRESS>:9101 error code: not controller # cmd/jocko/jocko topic create --topic weeble --broker-addr <IP_ADDRESS>:9001 created topic: weeble @candlerb thanks for the PRs, merged them. you need both serf and raft cause they do different things, serf does discovery and raft does consensus. right After latest push on branch candlerb/serfaddr (pull request #136), metadata response now works. Next problem is when publishing to a topic: 2018-06-23T21:17:07.518Z ERROR jocko/broker.go:427 produce to partition failed {"id": 0, "broker addr": "<IP_ADDRESS>:9092", "serf addr": "<IP_ADDRESS>:9094", "raft addr": "<IP_ADDRESS>:9093", "id": 0, "raft addr": "<IP_ADDRESS>:9093", "error": "no replica for topic mytopic partition 0"} github.com/travisjeffery/jocko/log.(*logger).Error /root/go/src/github.com/travisjeffery/jocko/log/logger.go:38 github.com/travisjeffery/jocko/jocko.(*Broker).handleProduce /root/go/src/github.com/travisjeffery/jocko/jocko/broker.go:427 github.com/travisjeffery/jocko/jocko.(*Broker).Run /root/go/src/github.com/travisjeffery/jocko/jocko/broker.go:146
2025-04-01T04:35:46.046783
2016-08-25T16:38:27
173252509
{ "authors": [ "jamie-sherriff", "patrickkunka" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11606", "repo": "travisjeffery/mocha-teamcity-reporter", "url": "https://github.com/travisjeffery/mocha-teamcity-reporter/pull/23" }
gharchive/pull-request
don't make commonjs export and hoisting to window mutually exclusive In mocha-casperjs for example, window will be defined at runtime, but reporters are still loaded via a commonjs require call I don't think this is applicable any more and will be refactored in upcoming changes. Please open an issue if you have problems
2025-04-01T04:35:46.052554
2016-12-08T08:50:29
194278387
{ "authors": [ "danielnorberg", "frsyuki" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11607", "repo": "treasure-data/digdag", "url": "https://github.com/treasure-data/digdag/pull/410" }
gharchive/pull-request
Add rename_tables option to td_ddl operator Syntax: td_ddl>: rename_tables: [{from: t1, to: t2}, {from: a1, to: a2}] How it works: make sure that all from tables exist. repeat renameTable REST API call with overwrite=true option until it succeeds or 404 Not Found happens. ignore 404 Not Found errors. This should be valid because step 1. confirmed that all tables exist. If 404 happens, it means that an operation is duplicated in somewhere (operator, client, load balancer, server, etc.). @danielnorberg PTAL Took the liberty of fixing the above issues. @danielnorberg Thank you for the additional updates!
2025-04-01T04:35:46.054757
2017-03-03T20:14:52
211790922
{ "authors": [ "frsyuki", "komamitsu" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11608", "repo": "treasure-data/digdag", "url": "https://github.com/treasure-data/digdag/pull/498" }
gharchive/pull-request
executor.task_ttl should not kill tasks BLOCKED, PLANNED or READY for long time BLOCKED and PLANNED states mean that there're other tasks running in the attempt (BLOCKED: running dependent tasks, PLANNED: running child tasks). Purpose of task_ttl is killing tasks without any progress for long time, killing BLOCKED / PLANNED tasks don't make sense. Instead, it should only kill leaf tasks. READY is also excluded because having task in READY state for long time means that the workflow is progressing expectedly but simply number of threads is insufficient. To kill long-running attempt, attempt_ttl should be used instead. 👍 , although I left a minor comment
2025-04-01T04:35:46.160491
2022-06-09T12:14:43
1266041381
{ "authors": [ "lukasbals" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11609", "repo": "treely/boemly", "url": "https://github.com/treely/boemly/issues/97" }
gharchive/issue
Remove NavItem component The NavItem component has been deprecated since it uses buttons and not links. Since frameworks like Next.js have their own link components it's hard to create links inside the component library that should be accessible to Google bots. That's why we remove the NavItem component. Done in https://github.com/treely/boemly/pull/98
2025-04-01T04:35:46.162348
2022-08-09T07:32:59
1332836857
{ "authors": [ "adipolak", "johnnyaug" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11610", "repo": "treeverse/lakeFS", "url": "https://github.com/treeverse/lakeFS/issues/3868" }
gharchive/issue
Surface Spark integrations in docs Many lakeFS users integrate it with Spark. To simplify the search experience of docs, Spark integrations should be a top-level category in our documentation. It is now the first element under Integrations.
2025-04-01T04:35:46.167705
2016-04-14T19:12:03
148455137
{ "authors": [ "PeEllAvaj", "mercersl" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11611", "repo": "trello/api-docs", "url": "https://github.com/trello/api-docs/pull/17" }
gharchive/pull-request
adding defaultLists param to board API docs. defaultLists was an unexposed param on the create board method for API, making it evident. @PeEllAvaj will you look over this just to make sure all is well? @mercersl Looks good to me. Merging now. The only improvement I could imagine would be to possibly explain what the parameter does (I've tried to always add more helpful narrative any time I touch the reference), but just having the parameter listed at all is a step in the right direction.
2025-04-01T04:35:46.172647
2022-12-19T14:54:40
1503093402
{ "authors": [ "cridus", "tozuf0x" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11612", "repo": "tremwil/SteamP2PInfo", "url": "https://github.com/tremwil/SteamP2PInfo/issues/24" }
gharchive/issue
Unhandled exception: Unauthorized Access So I've installed this app and selected a path to steam logs (changed it because I have installed steam elsewhere), however it is giving me access denied exception as I input command into console. I run it as admin and turned windows defender off to no avail. How do I grant access to that folder? Full error message app gives me is UnauthorizedAccessException: Access to the path "G:\SteamGames\logs" is denied. at System.IO.__Error.WinIOError(Int32 errorCode, String maybeFullPath) at System.IO.FileStream.Init(String path, FileMode mode, FileAccess access, Int32 rights, Boolean useRights, FileShare share, Int32 bufferSize, FileOptions options, SECURITY_ATTRIBUTES secAttrs, String msgPath, Boolean bFromProxy, Boolean useLongPath, Boolean checkHost) at System.IO.FileStream.ctor(String path, FileMode mode, FileAccess access, FileShare share) at SteamP2PInfo.SteamPeerManager.d__8.MoveNext() --- end of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Windows.Threading.ExceptionWrapper.InternalRealCall(Delegate callback, Object args, Int32 numArgs) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(Object source, Delegate callback, Object args, Int32 numArgs, Delegate catchHandler) and ipc_SteamClient.log is empty So, did you try manually changing the permissions as I said? Looks like both steamp2p and logs folder are in full access Changing app location to the same drive as steam seemed to solve the issue Then I guess we should check why it uses the wrong drive when the path explicitly defines the correct one.
2025-04-01T04:35:46.199046
2015-05-21T07:09:04
78866136
{ "authors": [ "Iafrus", "sargdavid" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11613", "repo": "trestletech/shinyRGL", "url": "https://github.com/trestletech/shinyRGL/issues/11" }
gharchive/issue
You must enable Javascript to view this page properly. I saw there's a closed issue with same title but no comments to that. I'm trying to use shinyRGL in an embedded shiny app in an rmarkdown doc and i get blank space and this message This is an ungoing issue, please see my post here: https://groups.google.com/forum/#!topic/shiny-discuss/GJQ6tJNU7jU
2025-04-01T04:35:46.208740
2020-09-24T08:49:49
707984153
{ "authors": [ "nirjanmunshi", "trevorwang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11614", "repo": "trevorwang/retrofit.dart", "url": "https://github.com/trevorwang/retrofit.dart/issues/248" }
gharchive/issue
Unable to get status code for a request I have created a model class and configured the basic structure of retrofit client @RestApi(baseUrl: "https://baseurl.in/") abstract class RetrofitService { factory RetrofitService(Dio dio, {String baseUrl}) = _RetrofitService; static RetrofitService create() { final dio = Dio(); dio.interceptors.add(PrettyDioLogger()); return RetrofitService(dio); } @GET("api/getConfiguration") Future<Config> getConfiguration(); } Since I am getting Future<Config> in return so I have created a class to get this data. Future<Config> getConfiguration() { return _client.getConfiguration(); } but the problem arises when I am calling this method from main(). API returns the data successfully but I am unable to find any status code for the request. I found if I include Future<Response<Config>> as a return type for the above mention function then it returns a response code but then the generated code by retrofit showing an error @override getConfiguration() async { const _extra = <String, dynamic>{}; final queryParameters = <String, dynamic>{}; final _data = <String, dynamic>{}; final Response<Map<String, dynamic>> _result = await _dio.request( 'api/easha/getConfiguration', queryParameters: queryParameters, options: RequestOptions( method: 'GET', headers: <String, dynamic>{}, extra: _extra, baseUrl: baseUrl), data: _data); // ERROR: Response<Config>.fromJson is not working here. which is logical final value = Response<Config>.fromJson(_result.data); return value; } How to fix this issue? try this https://github.com/trevorwang/retrofit.dart#get-orignal-http-reponse Ya alreadt stared the repo and, I saw that. That is OK. but it captures the status code on the catchError block. I want to get the status code after something like this if(response.statuscode == 200){ // do 200 work } else { // do other work } This library depends on dio which throws non 200 response as an error. Currently, you have to follow this
2025-04-01T04:35:46.211774
2023-05-16T15:31:48
1712263259
{ "authors": [ "Carapacik", "trevorwang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11615", "repo": "trevorwang/retrofit.dart", "url": "https://github.com/trevorwang/retrofit.dart/pull/571" }
gharchive/pull-request
Fix for build_runner ^2.4.4 Also add topics for pub.dev Link https://github.com/trevorwang/retrofit.dart/issues/570 https://github.com/dart-lang/build/issues/3509 @jakemac53 build.yaml(project) global_options: freezed: runs_before: - json_serializable json_serializable: runs_before: - retrofit_generator:retrofit build.yaml(retrofit) builders: retrofit_generator:retrofit: import: "package:retrofit_generator/retrofit_generator.dart" builder_factories: ["retrofitBuilder"] build_extensions: { ".dart": [".retrofit.g.part"] } auto_apply: dependents build_to: cache applies_builders: ["source_gen|combining_builder"] Right? @trevorwang Thanks for your contribution
2025-04-01T04:35:46.223061
2016-01-27T15:53:06
129183949
{ "authors": [ "agerber23", "lifeelement", "trgraglia" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11616", "repo": "trgraglia/jquery.horizonScroll.js", "url": "https://github.com/trgraglia/jquery.horizonScroll.js/issues/3" }
gharchive/issue
Mobile? Hey, does this plugin have any support for browsing on mobile devices? Swiping is supported via jquery.swipeTouch.js. It can be disabled if you like. (Check out the readme) Thanks, I meant when I viewed the demo (http://trgraglia.github.io/jquery.horizonScroll.js/demo/index.html) on mobile (iOS), everything was shrunk down and wasn't letting me swipe. I was wondering if something special had to be enabled? By the way, if anyone is curious, on ios9 there is a bug that causes this to render incorrectly. Here's the meta tag that got it to work for me, sorry for the late update: The swipe doesn't work in Chrome. Which version? @lifeelement 11 On Feb 16, 2016 11:43 PM, "agerber23"<EMAIL_ADDRESS>wrote: Which version? @lifeelement https://github.com/lifeelement — Reply to this email directly or view it on GitHub https://github.com/trgraglia/jquery.horizonScroll.js/issues/3#issuecomment-185019410 . The scrolling works by button click, but not the swiping. I'm using Windows 10 on a touch screen device. On Feb 17, 2016 9:00 AM, "Roy Oliver"<EMAIL_ADDRESS>wrote: 11 On Feb 16, 2016 11:43 PM, "agerber23"<EMAIL_ADDRESS>wrote: Which version? @lifeelement https://github.com/lifeelement — Reply to this email directly or view it on GitHub https://github.com/trgraglia/jquery.horizonScroll.js/issues/3#issuecomment-185019410 . Could you provide an example of the 'scrollTo' method? Using the $(document).horizon('scrollTo', i); doesn't seem to be working. With the demo provided and adding a button to that demo to navigate to the second index, how should this be done? With the mouse it works... I dont have a touchscreen win10 device. Can anyone else reproduce this?
2025-04-01T04:35:46.227422
2024-05-14T10:17:24
2295036391
{ "authors": [ "Starttoaster", "joanmp-ndtx", "jsmanrique" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11617", "repo": "trickstercache/trickster", "url": "https://github.com/trickstercache/trickster/issues/718" }
gharchive/issue
Is Trickster open to new 1.1.x releases? Hi, Currently we are using Trickster 1.1.5 and want to contribute with some fixes or minor improvements: Backfill tolerance is not applied when a prometheus query includes an "offset". Round start and end time for prometheus label values request to the most recent minute for increasing cacheability. (included in 2.0.0-beta2) Query params are not used as cache key in POST requests. (included in 2.0.0-beta2) Repeated query params (like match[] in prometheus label values) are discarded. The 1.1.5 was released 3 years ago. Is the branch v1.1.x alive and open to accept new PR's and releases? Any comment about this? Having spoken a bit with the maintainers here in the past, they appear to be very busy either with other projects, their paying jobs, or their personal lives. I wouldn't count on a comment coming soon. That said, having looked a bit at the legacy v1 branches in this repo, they're almost different codebases from what is in the default branch today. It might make more sense to try to update to the v2 beta and continue forward with those fixes, in my opinion. With their already very limited bandwidth, I imagine it being tricky for them to actively maintain those legacy branches. That said, I don't speak for the maintainers. That's just my hot take.
2025-04-01T04:35:46.236313
2015-08-22T00:43:32
102488855
{ "authors": [ "chepazzo", "coxley", "jathanism" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11618", "repo": "trigger/ansible-trigger", "url": "https://github.com/trigger/ansible-trigger/issues/2" }
gharchive/issue
Commando plugins support Does this support being able to specify Commando plugins to run, such as contrib.commando.plugins,gather_info? Could be useful to create custom plugins you'd want to run before hand or ones you often use, and then using Ansible to initiate it much like you would with the XMLRPC service. @chepazzo On Fri, Aug 21, 2015 at 17:43 Codey Oxley<EMAIL_ADDRESS>wrote: Does this support being able to specify Commando plugins to run, such as contrib.commando.plugins,gather_info? Could be useful to create custom plugins you'd want to run before hand or ones you often use, and then using Ansible to initiate it much like you would with the XMLRPC service. — Reply to this email directly or view it on GitHub https://github.com/trigger/ansible-trigger/issues/2. Can you describe how you would want this to work? If you give an example Ansible task (i.e. how would you want to call the module), and the expected return values (for use with debugging), I can tinker around and see what I can come up with. In theory, yes, it is possible, even if we just added a new module to the library. The idea, here, is to mimic what Juniper and Arista have done so that in my Ansible playbooks, my tasks would use the vendor-specific modules when available and these Trigger module otherwise. The current module included (trigger_command) is a very basic implementation that accepts a single command and a single device. Of course, we welcome any pull requests.
2025-04-01T04:35:46.572480
2021-06-06T20:42:34
912943566
{ "authors": [ "jbr", "prabirshrestha" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11619", "repo": "trillium-rs/trillium", "url": "https://github.com/trillium-rs/trillium/issues/23" }
gharchive/issue
Error handling Would it be possible to add error handling example in https://trillium.rs/ ? Doc are great by the way for a new project. Doc are great by the way for a new project. Thanks! The absence of error handling is not entirely a mistake, but I should address it directly in the documentation. Currently, the error handling story is the least-developed aspect of Trillium, but we are limited by what the language can express at the moment. In the future, we'll be able to have a ErrorHandler::new(async move |conn: &Conn| → Result<(), MyResult> { Ok(()) }) that impl's Handler, but until the language develops that capability, it's a matter of picking the least-bad workaround. There are two macros available to make this easier: trillium::conn_try! and trillium::conn_unwrap!, but I expect there'll be further improvements. Regardless of how it turns out, error handling is going to end up looking very different from tide. In particular, I think there's a distinction between "returning a non-200 response" and a rust Err, and that the code to translate between them should live entirely in the application. I think most applications will want to have a single thiserror-style enum Error that also implements Handler. I was actually looking into some sort of error handling sort of middleware (handler) that auto converts errors to json/html bad requests or 500 or thiserror to some json represntation. trillium_smol::run( Router::new() .get("/users", {conn-> Err('blah') )) .get("/posts", posts_api), ErrorHandler::new(async move |conn: &Conn| -> { conn.with_status(500).with_body(json!(conn.err())) }); ); Yeah, that's not possible yet in async rust I'm going to convert this to a discussion because I believe this will be a long-running topic of discussion, but will re-create an issue to represent the short term need for better documentation
2025-04-01T04:35:46.585463
2023-10-27T16:49:11
1965869003
{ "authors": [ "lambrospetrou", "oneonestar", "wewark" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11620", "repo": "trinodb/trino-gateway", "url": "https://github.com/trinodb/trino-gateway/issues/81" }
gharchive/issue
[Bug] The query history UI should use externalUrl instead of backendUrl Hello folks, we hit a bug using the new trino-gateway when opening query details from the gateway's UI displaying query history. The trino-gateway uses backendUrl instead of externalUrl in the links it shows in the UI (see code) which means when opened from our internal users, they do not resolve properly since it's trying to use the internal endpoints that the gateway uses to communicate with the clusters. The previous presto-gateway was using the gateway's URL for this (see code), which works so far in our case, and then it routes the UI calls to fetch the query details to the right coordinator cluster using the query mapping table/cache. At some point there is a request http://<gateway-url>/ui/api/query/<query_id>, which is matched by the request handler and gets sent to the right coordinator that has the query details. Solution I suggest one of the following: Use the externalUrl of the backend cluster for the links in the query history, and then we will need to set those to something resolvable by the expected users of the admin UI. Revert to using the gateway's URL and let the routing take care of finding the right backend cluster. I can provide a PR to fix this as long as we decide on which of the two solutions you would like to use. Maybe @mosabua can chime in with a preference or route to the right folks? Hello, any ETA for this? It seems like even setting externalUrl to the gateway URL has no effect. It is probably using the backend URL anyway https://github.com/trinodb/trino-gateway/blob/49c5c82dbc18a50b9104d001917dc3139088a0da/webapp/src/components/history.tsx#L57C30-L57C47 This issue is a bit complicated. Currently, we only store the backend_url in the query history table. Both backend_url and external_url can change at any time. This can lead to problems when the URLs are changed and the old backend_url in the query history no longer matches with the current setting. Storing the foreign key gateway_backend.name in the query history table could solve the issue. We can get the latest backend_url and external_url for the cluster. For audit and debugging purposes, backend_url should also be kept for reference. CREATE TABLE IF NOT EXISTS gateway_backend ( name VARCHAR(256) PRIMARY KEY, routing_group VARCHAR (256), backend_url VARCHAR (256), external_url VARCHAR (256), active BOOLEAN ); CREATE TABLE IF NOT EXISTS query_history ( query_id VARCHAR(256) PRIMARY KEY, query_text VARCHAR (256), created bigint, backend_url VARCHAR (256), user_name VARCHAR(256), source VARCHAR(256) ); @oneonestar thanks for detailed response. Do we have any ETA for that fix?
2025-04-01T04:35:46.618285
2016-02-04T11:07:44
131314391
{ "authors": [ "vedarthk" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11621", "repo": "tripit/slate", "url": "https://github.com/tripit/slate/pull/443" }
gharchive/pull-request
Add documentation for JS API This documents how to consume JS API in website or any browser based application. Also we have made fundamental change of separation of APIs into Frontend and Backend. This will also add documentation for Openings Frontend API. More can be found at http://developers.recruiterbox.com @tariqislam kindly please delete this, it was created by mistake on this repository.
2025-04-01T04:35:46.644036
2022-04-05T02:26:16
1192523639
{ "authors": [ "nllerandi3lift", "patrickloughrey" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11622", "repo": "triplelift-internal/Prebid.js", "url": "https://github.com/triplelift-internal/Prebid.js/pull/47" }
gharchive/pull-request
TL-19850 Finished log error logic around floors functionality Added try catch functionality around price floors logic Added corresponding test to make sure other currencies other than USD would be false @patrickloughrey I've merged into our forked master. Feel free to PR into the main repo
2025-04-01T04:35:46.696612
2016-12-01T15:15:29
192873657
{ "authors": [ "erikthedeveloper", "trueadm" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11626", "repo": "trueadm/inferno", "url": "https://github.com/trueadm/inferno/pull/497" }
gharchive/pull-request
GitHub links infernojs/inferno -> trueadm/inferno Caught another one via global Find!! Because @trueadm is ridiculously too fast and merged https://github.com/trueadm/inferno/pull/496 before I could push this one onto it! 😜 👍 💨 Actual Footage: I waited this time in case there were more commits :P Thanks again BTW, love the gif haha :)
2025-04-01T04:35:46.801973
2018-11-21T22:22:41
383312051
{ "authors": [ "rawadrifai" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11627", "repo": "trufflesuite/truffle-hdwallet-provider", "url": "https://github.com/trufflesuite/truffle-hdwallet-provider/issues/78" }
gharchive/issue
Account Not Found error var coinbase = "0x..."; var provider = new HDWalletProvider("....", "http://localhost:8545"); const truffleContract = TruffleContract(ERC20TokenABI); truffleContract.setProvider(provider); // create new ERC20 const result = await truffleContract.new( { from: coinbase, gas: gasLimit } ); It fails when I call a function of an initiated truffle contract. Error: the string "Account not found" was thrown, throw an Error :) at process._tickCallback (internal/process/next_tick.js:68:7) @joepagan - which version of web3 are you using? I'm using 0.20.x and the await web3.eth.getAccounts() is returning undefined for me. I wonder if it's the web3 version. So the only thing I am doing differently is hardcoding the account address as opposed to calling getAccounts. Do you know why this is different? Furthermore, are you able to do something similar with a private key instead of mnemonic? Thank you and I will upvote you on stackexchange. more info - If I follow @joepagan instructions it will throw this error: Error: Web3ProviderEngine does not support synchronous requests. ok I figured it out.. just needed to do coinbase.toLowerCase(). It now finds the account.
2025-04-01T04:35:46.818488
2024-07-09T19:08:42
2398971386
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11634", "repo": "trunk-io/mergequeue-staging", "url": "https://github.com/trunk-io/mergequeue-staging/pull/22973" }
gharchive/pull-request
affectivity, guavina This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 100 sleep for: 300s close stale after: 4 hours [pullrequest] requests per hour: 0 deps=[a,g] /trunk merge
2025-04-01T04:35:46.820183
2024-07-11T05:08:32
2402260803
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11635", "repo": "trunk-io/mergequeue-staging", "url": "https://github.com/trunk-io/mergequeue-staging/pull/23770" }
gharchive/pull-request
ethynyl, boodles This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 100 sleep for: 300s close stale after: 4 hours [pullrequest] requests per hour: 0 deps=[b,e] /trunk merge
2025-04-01T04:35:46.821682
2024-07-13T21:08:32
2407164650
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11636", "repo": "trunk-io/mergequeue-staging", "url": "https://github.com/trunk-io/mergequeue-staging/pull/24914" }
gharchive/pull-request
cushion, analphabete This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 100 sleep for: 300s close stale after: 4 hours [pullrequest] requests per hour: 0 deps=[a,c] /trunk merge
2025-04-01T04:35:46.846533
2024-05-09T23:52:00
2288648027
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11647", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/48758" }
gharchive/pull-request
drowsed, gumihan This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 2100s close stale after: 24 hours [pullrequest] requests per hour: 100 deps=[d,g] /trunk merge
2025-04-01T04:35:46.848323
2024-05-17T01:50:16
2301654239
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11648", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/53394" }
gharchive/pull-request
boracic, allegretto This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 2100s close stale after: 24 hours [pullrequest] requests per hour: 100 deps=[a,b] /trunk merge
2025-04-01T04:35:46.851480
2024-05-28T06:41:43
2320254926
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11650", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/60097" }
gharchive/pull-request
dichroic, egol This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 2100s close stale after: 24 hours [pullrequest] requests per hour: 100 deps=[d,e] /trunk merge
2025-04-01T04:35:46.852990
2024-05-29T18:59:07
2324037266
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11651", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/61257" }
gharchive/pull-request
griffade, argolic This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 2100s close stale after: 24 hours [pullrequest] requests per hour: 100 deps=[a,g] /trunk merge
2025-04-01T04:35:46.854533
2024-05-29T20:52:14
2324229150
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11652", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/61354" }
gharchive/pull-request
compulsorily, decarbonate This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 2100s close stale after: 24 hours [pullrequest] requests per hour: 100 deps=[c,d] /trunk merge
2025-04-01T04:35:46.861165
2024-06-15T17:58:16
2355081690
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11656", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/76357" }
gharchive/pull-request
ambivalent This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 600s close stale after: 24 hours [pullrequest] requests per hour: 20 deps=[a] /trunk merge
2025-04-01T04:35:46.862789
2024-06-20T21:31:49
2365312677
{ "authors": [ "EliSchleifer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11657", "repo": "trunk-io/mergequeue", "url": "https://github.com/trunk-io/mergequeue/pull/81241" }
gharchive/pull-request
dessil, enddamaging This pull request was generated by the 'mq' tool [test] flake rate: 0.1 logical conflict every: 1000 sleep for: 600s close stale after: 24 hours [pullrequest] requests per hour: 20 deps=[d,e] /trunk merge
2025-04-01T04:35:46.876890
2024-02-27T22:04:28
2157690479
{ "authors": [ "NOTz00m", "amirslayer", "truong-bui" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11663", "repo": "truong-bui/AsyncLoadingScreen", "url": "https://github.com/truong-bui/AsyncLoadingScreen/issues/92" }
gharchive/issue
Async Loading Screen Shows Black Screen on Startup I'm using the Async Loading Screen plugin in Unreal Engine 5.3.2 and I'm encountering an issue where the screen just shows black on startup (regardless of if i'm playing the game in Standalone version or if it's actually packaged for shipping) and then simply goes to my main menu. I've tried updating my graphics driver, revisiting the setup instructions, looking at fixes for other issues, etc... but the issue persists. I've tried it in different projects, disabled all other unused plugins, different mp4 files (including the ones provided in the instructions), but it doesn't work for some reason. Any help would be greatly appreciated, thanks. Hi, Have you carefully followed the tutorial to understand how it works yet? Usually, this issue is caused by the movie path not being correct. Thanks for the response, My bad, I accdientally closed the issue. Thank you for the response, but unfortunately. the movie path doesn't seem to be the issue. It stays on the black screen for about 3 seconds and then loads into the main menu. Is there anything I can do about this? I recommend following the plugin's tutorial to see how it works. If the movie clip in the tutorial works (which I am sure works fine), there may be something wrong with your clip format or path, or your project has a plugin that conflicts with AsyncLoadingScreen. @truong-bui i can verify it, the video is appearing after a long long time! @truong-bui I can say with 100% certainty that I have followed the plugin's tutorial to see how it works and have done the exact steps mentioned there. I made a whole new blank project, enabled ONLY AsyncLoadingScreen (disabled other plugins), followed the steps, put the clip path exactly as mentioned (I used the clip in the tutorial) and it still comes up as just black for 3 seconds and then brings me to the Open Level Button. I also tried the default loading screen setting and that works just fine, however, the startup loading screen does not work. Before adding the movie clip, it does display a black screen with a white border at the bottom, but that doesn't help me considering I need a video at startup. Well, it is very weird. Because it just works fine on my side. Here you can check the Async Loading demo that I packaged recently with Unreal version 5.3.2 https://drive.google.com/file/d/1b7rRsqwR1rl_1TJpnLYUYTImA31SyA3I/view?usp=drive_link Can you please provide a link/screenshot/video to the issue? So I can take a look at it. I cannot help if I can't reproduce the issue on my side. Thank you for the support, I managed to find the issue out through trial and error. Apparently, something to do with the permissions for my Windows Explorer/Account was not allowing Unreal Engine/AsyncLoadingScreen to see the file and therefore was just playing nothing. Atleast for now, I managed to fix it and now it finally works which is great. Again, thank you for the help and I'm sorry for any inconveniences caused.
2025-04-01T04:35:46.908526
2018-06-19T14:08:24
333688743
{ "authors": [ "berndbohmeier", "jensneuber" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11664", "repo": "trustlines-network/relay", "url": "https://github.com/trustlines-network/relay/issues/112" }
gharchive/issue
Send events via push notifications service Use the firebase integration to push events @jensneuber Do you need them as notification or as data message or as notification with some data payload? https://firebase.google.com/docs/cloud-messaging/concept-options -> as notification with some data payload Okay, so what should the title, body and data be. Is data just the event dict? title for first iteration is "new ", body is just "click for more details".
2025-04-01T04:35:46.927017
2022-04-22T09:58:55
1212137638
{ "authors": [ "catenocrypt", "hewigovens", "serin-sebastian", "serinkuriapilly1997" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11665", "repo": "trustwallet/wallet-core", "url": "https://github.com/trustwallet/wallet-core/issues/2156" }
gharchive/issue
Doge signing error I have signed a doge transaction and when I try to broadcast the transaction , I got the following error try reversing your txid try reversing your txid I have already tried that , but the error still exist. when i decode the transaction the txid is shown as interchanged ( adjacent characters interchanged). for example f30944c3f6a0670f4e732a98079d2ddfa22b312fbb4c96fad225b21cfbf5c235 is shown as 3f90443c6f0a76f0e437a28970d9d2fd2ab213f2bbc469af2d522bc1bf5f2c53 while d ecoding. how did you reverse it? you need to decode the tx id as bytes then reverse I tried this but when I try to decode then server error (500) occurred . I am attaching my code sample here . The mentioned TX has only 1 output, as far as I can see, so the outputIndex 7 is wrong. https://blockchair.com/dogecoin/transaction/af8ff35aa7237b1b59c790826bca62fcef4776cd699e6de8b5b4cf704760fac1 Thank you it works but transaction is not confirmed yet. transaction hash is : 2f4a97555523b1e0eae8d97366573151b1cb7bf98c954dbd1cefb3cfdeadc5e4 . Is there any other issue with my code? . There are multiple txids so, Is there any change in this code ? Transaction is successfully broadcasted but it showing minor preference low . Transaction is not yet confirmed . How can i increase the minor preference in my code You set byte fee as 1 which is too low at current network situation https://blockchair.com/dogecoin/charts/median-transaction-fee-usd Then what is the sufficient value for byte fee ?
2025-04-01T04:35:46.962915
2024-10-08T19:09:37
2573991118
{ "authors": [ "Fryingpannn", "itzamanjain" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11666", "repo": "trypear/pear-landing-page", "url": "https://github.com/trypear/pear-landing-page/pull/318" }
gharchive/pull-request
[WIP] Blog Submission - No-code Blog Feature Description This PR implements the blog submission form as part of the no-code blog feature. need to create a blog table first in prod with these cols CREATE TABLE blog ( id SERIAL PRIMARY KEY, title VARCHAR(255) NOT NULL, url VARCHAR(255) NOT NULL, excerpt TEXT, date DATE NOT NULL, author VARCHAR(255) NOT NULL, tag VARCHAR(255), content TEXT NOT NULL, readingTime INT, status VARCHAR(50) DEFAULT 'pending' ); Related Issue Issue #305 Changes Made Summary: This PR introduces a blog submission form that allows users to submit blogs by providing details such as the blog title, content, tags, and author information. The next step is to integrate the form with the backend for storing blog data and to implement an approval system for moderators. Changes: Added /submit-blog route with a submission form. Form fields include title, content (Markdown), author, email, tags, and more. Next Steps: Integrate the form with the backend to store blogs in the database. Implement an admin/moderator approval system. Screenshots Checklist [ ] I have tagged the issue in this PR. [ ] I have attached necessary screenshots. [ ] I have provided a short description of the PR. [ ] I ran yarn build and the build is successful. [ ] My code follows the style guidelines of this project. [ ] I have added necessary documentation (if applicable). 👀 @Fryingpannn I’ve added some more changes to make blog submission no code . Now, users can submit a post by going to /blog/submit-blog using the blog form. The blog will be added to the blog table in Supabase with the status set to "pending". Admins can go to /secret-admin-page to see all blogs in the pending state and can approve or reject them, which will update the status in the database. Things to be done: Ensure only authenticated users can submit a blog. (can be done using withAuth() ) Restrict access to the admin page so that only admins can approve or reject blogs. Set up a flow to display approved blogs on the blog page. What do you think of this approach? Any suggestions or feedback?
2025-04-01T04:35:46.971044
2024-10-09T20:12:11
2576861285
{ "authors": [ "nickgarlis", "toteto" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11668", "repo": "ts-rest/ts-rest", "url": "https://github.com/ts-rest/ts-rest/issues/705" }
gharchive/issue
Add rawBody option to disable body parsing on server Problem There are use-cases where it would be useful to disable body parsing even if the content-type is application/json. An example would be verifying Stripe's webhook signatures. From their docs: Stripe requires the raw body of the request to perform signature verification. If you’re using a framework, make sure it doesn’t manipulate the raw body. Any manipulation to the raw body of the request causes the verification to fail. Proposal Adding a rawBody flag on the contract could allow users to disable this automatic behavior when necessary. Implementation I was thinking of modifying the following function to something like this. const evaluateContent = async (request: TsRestRequest) => { if (request.method !== 'GET' && request.method !== 'HEAD') { if (request.headers.get('content-type')?.includes('json') && !schema.rawBody) { request['content'] = await request.json(); } else if (request.headers.get('content-type')?.startsWith('text/') || schema.rawBody) { request['content'] = await request.text(); } } }; If this is of interest then I could open a Pull Request with the proposed changes and some tests. Additional context I have only used the serverless Next handler and admittedly I haven't given this too much of a thought but perhaps this is something that the other handlers could benefit from ? Personally I have modified the JSON parsing of Express, not sure if the same approach is applicable for you. application.use( express.json({ verify: (req, _res, buf) => { if (req.url && options.rawBodyEndpoints?.includes(req.url)) { Object.assign(req, { rawBody: buf.toString() }); } }, }), ); options.rawBodyEndpoints includes list of endpoints that should receive the rawBody. On the consumer side, I have method for getting the rawBody param export function requestRawBody(req: Request): string | Buffer | undefined { if (!('rawBody' in req)) return undefined; if (typeof req.rawBody === 'string' || Buffer.isBuffer(req.rawBody)) return req.rawBody; console.warn('rawBody is not a string or Buffer'); return undefined; }
2025-04-01T04:35:46.980229
2023-06-07T20:29:22
1746630321
{ "authors": [ "csturiale", "tsawler" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11669", "repo": "tsawler/signer", "url": "https://github.com/tsawler/signer/pull/1" }
gharchive/pull-request
using urlParser instead of split to fix creation of longer signed urls Updated the library in order to be able to sign url where the path is longer. Added a specific test where module was not working properly Ie. using url: <EMAIL_ADDRESS> Thanks for this!
2025-04-01T04:35:46.986802
2012-01-17T09:24:33
2866218
{ "authors": [ "guyf", "tschellenbach" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11670", "repo": "tschellenbach/DjangoOpenInviter", "url": "https://github.com/tschellenbach/DjangoOpenInviter/issues/5" }
gharchive/issue
form clean issue? Hi, I have got this running - thanks for publishing! I have had to change line 15 in forms.py to be a form level clean not field level (ie. def clean not def clean_email) it seems that self.cleaned_data['password'] is out of scope if doing clean_email. This seems a pretty fundamental issue - so I am wondering if my understanding/usage is incorrect? Guy. I think the order is probably random, but quite stable. For me it's always working. It should be refactored to the clean method. The only hard bit is to attach the resulting error to the email field. Cheers, Thierry On Tue, Jan 17, 2012 at 10:24 AM, Guy < <EMAIL_ADDRESS> wrote: Hi, I have got this running - thanks for publishing! I have had to change line 15 in forms.py to be a form level clean not field level (ie. def clean not def clean_email) it seems that self.cleaned_data['password'] is out of scope if doing clean_email. This seems a pretty fundamental issue - so I am wondering if my understanding/usage is incorrect? Guy. Reply to this email directly or view it on GitHub: https://github.com/tschellenbach/DjangoOpenInviter/issues/5
2025-04-01T04:35:47.001857
2022-05-03T20:13:47
1224580665
{ "authors": [ "lelegard", "lmoody1" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11671", "repo": "tsduck/tsduck", "url": "https://github.com/tsduck/tsduck/issues/989" }
gharchive/issue
[BUG] Output Stream with Null Input Streams Won't Execute Bug description: Output stream doesn't start if a null input stream is supplied How to reproduce: tsp --control-local localhost --control-port 8000 -v -b 41000000 -I null -j -O ip localhost:8999 -P merge -j --no-psi-merge --ignore-conflicts -p 0x1-0x20 -d 0x0 -d 0x21-0x1FFE "tsp -v -b 3000000 -I file -i C:/Files/Stream_MULx21.ts -P regulate" -P regulate -P merge -j -t "tsp -b 2000000 -v -I null -P regulate" -P pat -P pmt -c -P regulate Expected behavior: The merged streams (except for the null stream) should output to localhost: 8999 and be seen on (for example) tsReader. Expecting to see (at the minimum) "TOT", "SDT", "NIT", "CAT" and "TDT" PIDs. Not expected to see PAT. See "Merged Streams without Null Input Stream (Works as Expected)" in the attached document as an example of what we should be seeing. Errors and logs: The merged stream isn't being output. See "Merged Streams with Null Input Stream (Not Working as Expected)" in attached document. Environment: OS: Windows OS version: Windows 10 TSDuck full version: 3.30-2710 Installation type: official binary TSDuck-Win64-3.30-2710.exe Additional information: Output is as expected if the null input stream is restarted with new parameters (see Merged Streams with Null Input Stream Restarted with New Parameters (Works as Expected) in attached document): tspcontrol -t localhost:8000 restart -v 3 -j -t "tsp -b 2000000 -v -I file -i C:/Files/Stream_8E1.ts -P regulate" tspcontrol -t localhost:8000 restart -v 4 -c -a 0x8FF6/0x21 tspcontrol -t localhost:8000 restart -v 5 -a 0x8E1/0x5 -s 0x8FF6 --pcr-pid 0x1fff Output is as expected if the stream is executed without the null input stream (see Merged Streams without Null Input Stream (Works as Expected in attached document): tsp --control-local localhost --control-port 8000 -v -b 41000000 -I null -j -O ip localhost:8999 -P merge -j --no-psi-merge --ignore-conflicts -p 0x1-0x20 -d 0x0 -d 0x21-0x1FFE "tsp -v -b 3000000 -I file -i C:/Files/Stream_MULx21.ts -P regulate" -P regulate Merged Streams with Null Input Stream Bug.docx Hi @lmoody1 I do not know what is the exact issue here but I would recommend to start with a more rational usage of the regulate plugin. There is no need to put two of them in sequence, it is at best useless and at worst counterproductive. Moreover, the first merge is placed before the first regulate, meaning that the secondary stream is merged on an unregulated stream of null packets. Additionally, there is no need for -j options if you do not use "joint termination". I have no idea if this will fix the issue but you should start with a more rational regulation, something like: tsp -b xxxxx ....\ -I null \ -P regulate \ -P merge "tsp -I file ... -P regulate" \ -P merge "tsp -I file ... -P regulate" \ -P ... -O ip ... I've tried this with a multitude of variations of locations of the regulate plugin(s), as well as the addition, removal and ordering of pat & pmt plugins. I have also tried it without the -j option. I have not been able to get it to work. My goal is to be able to start and stop input streams at any time with new stream settings. I'm doing this by creating a muxed output stream with a file input stream merged with multiple null input streams that can be modified and restarted at any time with the users' requested configurations. This is the basic logical: input from a file input from null stream input from null stream input from null stream output to localhost In the above example, null input streams 2, 3, 4 are available to the users to restart at any time with user requested settings (i.e. input from file "c:\files\xxx.ts", with a bitrate of "2000000", to service "0x8FF1", and input PID "0x8e1" mapped to PID "0x8E2") What's happening is: If I create a merge stream with only the input from file stream (i.e. 1 above without 2, 3, or 4) I am seeing the output. If I create a merge stream with the input from a file merged with the null streams (i.e. 1 above merged with 2, 3 and 4). I am not seeing the output (I don't think it's executing the stream at all) If I create a merge stream with the input from a file merged with a null stream (i.e. 1 above merged with 2, 3 and 4) I am not seeing the output, but then I restart one of the null input streams with user requested settings (i.e. restart 2 above with input from file "c:\files\xxx.ts", with a bitrate "2000000", to service "0x8FF1", and input PID "0x8e1" mapped to PID "0x8E2"), I am seeing the output. One thing that does work is if I specify a service and PID for the null input streams. I.E. with a service starting at 0xA0 and PID starting at 0xB0: tsp --control-local localhost --control-port 8000 -v -b 41000000 -I null -j -O ip localhost:8999 -P merge -j --no-psi-merge --ignore-conflicts -p 0x1-0x20 -d 0x0 -d 0x21-0x1FFE "tsp -v -b 3000000 -I file -i C:/files/input_stream.ts -P regulate" -P regulate -P merge -j -t "tsp -b 100000 -v -I null -P regulate" -P pat -c -a 0xA0/0x21 -P pmt -c -a 0xB0/0x5 -s 0xA0 --pcr-pid 0x1fff -P merge -j -t "tsp -b 100000 -v -I null -P regulate" -P pat -c -a 0xA1/0x21 -P pmt -c -a 0xB1/0x5 -s 0xA1 --pcr-pid 0x1fff However, this requires us to create a dummy service and a dummy PID for each null input stream. We'd prefer not to have to do this. It would require us to reserve those values so the users won't be able to use them (and we're looking to reserve 32 null streams so that will eliminate 32 services and PIDs for users to use). Instead, we'd prefer to do something like this: tsp --control-local localhost --control-port 8000 -v -b 41000000 -I null -j -O ip localhost:8999 -P merge -j --no-psi-merge --ignore-conflicts -p 0x1-0x20 -d 0x0 -d 0x21-0x1FFE "tsp -v -b 3000000 -I file -i C:/files/input_stream.ts -P regulate" -P regulate -P merge -j -t "tsp -b 100000 -v -I null -P regulate" -P pat -P pmt -c -P merge -j -t "tsp -b 100000 -v -I null -P regulate" -P pat -P pmt -c I hope this all makes sense. I appreciate your help. -Lee
2025-04-01T04:35:47.007024
2022-05-13T09:50:05
1235010508
{ "authors": [ "WolfWFr", "abdukhashimov" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11672", "repo": "tsenart/vegeta", "url": "https://github.com/tsenart/vegeta/issues/595" }
gharchive/issue
go 1.18 compatibility: import latest github.com/streadway/quantile commit Proposal I would like to ask to update this repository by importing the latest commit of the github.com/streadway/quantile dependency. Currently, trying to import the tsenart/vegeta repository into a go v1.18 project, I get the following error on running go mod tidy: bitbucket.org/innius/api/test/throttling tested by bitbucket.org/innius/api/test/throttling.test imports github.com/tsenart/vegeta/v12/lib tested by github.com/tsenart/vegeta/v12/lib.test imports github.com/streadway/quantile tested by github.com/streadway/quantile.test imports .: "." is relative, but relative import paths are not supported in module mode This relative import statement in github.com/streadway/quantile appears to have been removed in their latest commit. Background Using go v1.18 and importing this repo yields errors from go mod commands and is preventing successful builds. Workarounds revert to older version of go. Is there any updates on this issue?
2025-04-01T04:35:47.011039
2015-08-31T10:45:36
104028988
{ "authors": [ "permagnus", "tsheaff" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11673", "repo": "tsheaff/SexyTooltip", "url": "https://github.com/tsheaff/SexyTooltip/issues/1" }
gharchive/issue
Not compatible with App Extensions Hi. This lib does not compile in an app extension project. References to UIApplication sharedApplication are not allowed. True, that wasn't a design goal or consideration when I built it. Feel free to submit a pull request. Thanks.
2025-04-01T04:35:47.042277
2020-11-01T13:28:46
733966366
{ "authors": [ "kolumb", "rexim" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11674", "repo": "tsoding/something", "url": "https://github.com/tsoding/something/pull/257" }
gharchive/pull-request
Stop all enemies in debug mode If you enter debug mode while enemy is moving towards you, he'll continue to move in that direction forever. @kolumb looks good to me! :+1: Thank you for your contribution!
2025-04-01T04:35:47.095348
2017-02-03T21:04:09
205268985
{ "authors": [ "faddee", "tuchk4" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11675", "repo": "tuchk4/storybook-readme", "url": "https://github.com/tuchk4/storybook-readme/issues/9" }
gharchive/issue
Why is @kadira/storybook-addons peer dependency? When installing with Yarn I get this warning: warning<EMAIL_ADDRESS>has unmet peer dependency "<EMAIL_ADDRESS>It's a dependency for @kadira/storybook but I was able to get it to work just fine without using @kadira/storybook-addons explicitly. So is it really necessary to have is as a peer dependency? I took one of existing storybook addong as the example. And @kadira/storybook-addons at most of other addons: storybook-addon-links storybook-addon-actions storybook-addon-knobs When should you use peer dependencies? When you are building a library to be used by other projects, and This library is using some other library, and You expect/need the user to work with that other library as well More info here: Peer Dependencies @ nodejs.org Why use peer dependencies in npm for plugins? I never faced such problem at all my apps with storybook and this addon. Can you show please your package.json and npm -v? You expect/need the user to work with that other library as well Which I don't. I'm not expected to use @kadira/storybook-addons to get this addon to work and therefor would I rather not to have it as dependency at all. I was generally just wondering if I missed something. But I guess it's there for another reason. @faddee @kadira/storybook-addons is dependecy of @kadira/storybook (package.json). Seems your get warning because of versions mismatch. That is why I asked to show your package.json Or try to update all to latest versions with npm-check npm-check -u Here's my dependencies... { "dependencies": { "@types/react": "^15.0.6", "@types/react-dom": "^0.14.22", "@types/react-redux": "^4.4.36", "react": "^15.4.2", "react-dom": "^15.4.2", "react-redux": "^5.0.2", "redux": "^3.6.0", "ts-node": "^2.0.0", "tslint": "^4.4.2", "typescript": "2.1.5" }, "devDependencies": { "@kadira/storybook": "^2.35.3", "awesome-typescript-loader": "^3.0.0-beta.18", "css-loader": "^0.26.1", "file-loader": "^0.10.0", "json-loader": "^0.5.4", "node-sass": "^4.5.0", "raw-loader": "^0.5.1", "react-svg-loader": "^1.1.1", "sass-loader": "^4.1.1", "storybook-readme": "^1.2.0", "style-loader": "^0.13.1", "webpack": "1.x" } }
2025-04-01T04:35:47.099938
2020-11-08T19:20:37
738535009
{ "authors": [ "Enrico204", "Hirama", "demget" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11676", "repo": "tucnak/telebot", "url": "https://github.com/tucnak/telebot/issues/334" }
gharchive/issue
Support error and params Telegram returns error Bad Request: group chat was upgraded to a supergroup chat with new chat id but bot cuts this info and I can't get a new chat id { "ok": false, "error_code": 400, "description": "Bad Request: group chat was upgraded to a supergroup chat", "parameters": { "migrate_to_chat_id": -1001432162429 } } I'm having the same issue too. I'm working on a pull request for a workaround. @tucnak I think that the error checking function is fundamentally broken. The Go JSON parser can be used to detect whether the JSON contains an error or not We can catch it similarly to the flood error like here: https://github.com/tucnak/telebot/commit/315d880be7eac113df384902ee6cfda3606146be. But it's getting too cumbersome to handle using regexes, so it'll be great to move to the JSON unmarshaling. PRs are welcome. I wrote the #335 for this. However I think that the code needs some major refactoring in that section #335 9805b1f622d8161ed5615d0738c8071066ed7dea
2025-04-01T04:35:47.100917
2023-11-19T17:00:39
2000942469
{ "authors": [ "gioagu" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11677", "repo": "tudelft3d/3DCityDB-Tools-for-QGIS", "url": "https://github.com/tudelft3d/3DCityDB-Tools-for-QGIS/issues/58" }
gharchive/issue
Can't delete solitary vegetation objets from layers It's a bug due to an error in the delete trigger functions that are invoked when trying to delete features from a layer. Solved in the upcoming v. 0.8.7
2025-04-01T04:35:47.127602
2018-12-04T15:38:37
387345837
{ "authors": [ "fgameiro", "tulios" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11678", "repo": "tulios/mappersmith", "url": "https://github.com/tulios/mappersmith/issues/123" }
gharchive/issue
Identifying 502 bad gateway errors When trying to understand some API calls returning a 502 bad gateway error, I found out that mappersmith returns, in this case, a default status code of 400 and the Network error message. So I am unable to filter out these failed calls through the lib. Is this an expected result or is it a bug? Is there a way to get a response object with the status code 502? @fgameiro sorry for the delay, mappersmith should only return 400 for network errors, but I can see a potential bug on the node http gateway, are you talking about node.js? https://github.com/tulios/mappersmith/blob/7381e226be5f7dde4fb1870b6e9f0113608305ef/src/gateway/http.js#L108-L113 This might be happening for 502, I can double check hey @tulios, thanks for the response! No, actually the problem is happening on browser, but after further checking, we found out that what is probably happening is that the cloudflare service we use is masking our server's 502. Then, we can't reach the cloudflare's 502 because of their lack of Access-Control-Allow-Credentials. After all this, we conclude that mappersmith is getting an errored response as we expect, but without any info, so we get some sort of default response (the 400 status code and the Network error message).
2025-04-01T04:35:47.152635
2023-10-10T23:44:26
1936478672
{ "authors": [ "Qubitium", "Tostino", "anujnayyar1", "turboderp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11679", "repo": "turboderp/exllamav2", "url": "https://github.com/turboderp/exllamav2/issues/103" }
gharchive/issue
Sliding Attention Window Dear @turboderp, Just wondering if you have been able to implement the sliding attention window for mistralai/Mistral-7B-v0.1 and its derivatives, It would be awesome to be able to use this model up to its full context window, with the speed of exllama!! All the very best and thank you so much for your continued amazing contributions to the community!! I've considered various ways of implementing it, and the problem isn't actually the SWA itself but rather creating an API for it that doesn't break everything else. Keep in mind, all it does (unless I'm still completely failing to understand the paper) is truncate the cache, as opposed to truncating and reevaluating the context. You don't actually get a 32k context size or 128k or whatever they advertise. If you have a 4k sliding window, any tokens more than 4k positions in the past are still forgotten and can no longer be attended to. They may still be vaguely reflected in the keys and values produced before that past was forgotten, but the actual keys are no longer present and there's only so much information that can be conveyed without them, especially with the shorter GQA keys/values of Mistral. So it's very questionable how valuable SWA is for context size, and after all Mistral works well past the 4k window size when used as a regular Llama model (good for at least 7k in my tests). As for speed, it is faster to truncate (or better yet, roll) the cache than to truncate the context, but it's not a huge difference in the end. You typically wouldn't truncate the context for every token added, but maybe for every 200 tokens or so. At that point, you're just resetting the cache and forwarding the truncated context as a prompt, which will process at some 14,000 tokens/second on a 4090. So it's not much of a stutter you get, and it's not all that often. One big problem with truncating/rolling the cache is that you lose the system prompt, character cards and whatever else you stuck at the beginning of the context. If you settle for the "regular approach" it's trivial to retain those parts of the context when truncating (both the CLI chat example in ExLlamaV2 and the web UI in V1 do this). I've yet to work out if there's some sort of compromise that would mark part of the cache to retain as the rest of it rolls. But either way it's going to complicate the interface, as there isn't really a way to do it transparently, I don't think. So yes, I do plan to do something with it, but I'm not sure what form it will take, and I'd advise everyone not to make too many assumptions about the benefits of SWA. @turboderp So the issue with the system prompt / character card sliding out of context is solved by locking a portion of the KV cache (first N tokens), and then truncate the sliding window that begins after that locked portion. With llamacpp you just use --keep [token count of system prompt] @Tostino But that means the sliding context is a disjointed/non-continuous franken merge of header + footer with middle cutoff. Not sure how this would degrade output. At least the current sliding window is a mostly continuous/coherent message. Have you tried this --keep method and gotten acceptable output? I think you misunderstood what I was trying to describe. I have tried it by setting --keep in llamacpp and it successfully continued generation after truncating the sliding window portion. Without recompiling, it discards half of the sliding window portion of the kv cache, but that is easy to change to discard whole conversation turns at a time instead. I don't quite understand what you mean by disjointed header, middle, and footer. You just have a fixed "header" section, and you add to the end of your sliding window until it is full, and then it starts discarding the tokens after your header section at the beginning of your sliding window. So I don't quite get how it could be disjointed. @Tostino I think we are talking about the same thing. Let me rewrite my end to be more clear. A -> B ->C . If A is fixed header in --keep, and B is the extra part that is > sliding window, you are asking model to generate D with A -> C. Generation of D (next part) having attentions to A and C with B chopped off. My worry is that the cutting the linear generated B that precedes C may cause issues. Could be wrong on this as you have said it is working correctly in your case. Ah, yeah it will only work with tasks that will not require referring back to information at the start of B by the time we need to truncate it. E.g. I can stuff a document almost to the entire context length of the LLM, and ask for a summary where the generated summary would overflow the available context window. After we have summarized the first paragraph, we don't really care if it is truncated in my experience. Or at least it hasn't mattered for me when I was testing on some of my tasks that would possibly work with this. Closing this as stale. Rolling context may be addressed in the future with continuous batching/paged attn at some point.
2025-04-01T04:35:47.175444
2018-09-07T13:15:41
358062775
{ "authors": [ "domchristie", "gingerlime", "joker-777" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11680", "repo": "turbolinks/turbolinks", "url": "https://github.com/turbolinks/turbolinks/issues/413" }
gharchive/issue
fetch page on back/forward (and keep caching)? Turbolinks looks awesome, and the caching implementation also seems so clever. Also the codebase is so nice and clean. It's a pleasure to read. We want to use caching, but when the user hits the back/forward, we hope to still be able to load the page (after the initial preview from cache -- just like normal clicks work). The main reason for this, is that we can fetch fresh data from gon, which is embedded on the body. Is there anything we can do to hook into events and detect/distinguish normal visits from back/forward visits? and if we do detect those, can we trigger a real visit before turbolinks:load fires? Otherwise, is there a way to make browser back/forward behave the same as normal turbolinks visits? Interesting problem! For pages that are likely to change I have tended to just set <meta name="turbolinks-cache-control" content="no-cache"> and take the performance hit, however refreshing on back/forward is intriguing. For restoration visits (i.e. back/forward), Turbolinks binds to the popstate event, so as a possible approach, we could also bind to popstate and reload the page with Turbolinks. This might look like: ;(function () { var reloadWithTurbolinks = (function () { var scrollPosition function reload () { scrollPosition = [window.scrollX, window.scrollY] Turbolinks.visit(window.location.toString(), { action: 'replace' }) } window.addEventListener('turbolinks:load', function () { if (scrollPosition) { window.scrollTo.apply(window, scrollPosition) scrollPosition = null } }) return reload })() window.addEventListener('popstate', function () { requestAnimationFrame(reloadWithTurbolinks) }) })() A couple of downsides. popstate is also fired when visiting same-page anchors e.g. <a href="#my_anchor">Jump to …</a>, which might be problematic. Secondly there seems to be a bug with the scroll position when navigating back, which is not ideal … but this is a start and it may work for you! Thanks for the detailed response and suggestion. We also considered hooking into the popstate, but we also have some other cases that hook into it (specifically, with Algolia search), and we weren't entirely sure about the best action to take. Looks like Turbolink.visit(..., {action: 'replace'}) would make this seamless, similarly to what happens with normal page visits. Looks like the anchor link behaviour might throw another spanner in the works in this case :-/ Re adding no-cache - in our case, gon acts as some kind of bridge between server and client, so it can get updated on almost any page. I'm curious if the different behaviour on back/forward compared to normal page visit is intentional, or due to how the browser behaves. Do you happen to know? (i.e. normal visit would preview from cache, and then fetch the page and update it. Back/Forward only shows the cached data). /cc @joker-777 We also considered hooking into the popstate, but we also have some other cases that hook into it (specifically, with Algolia search), and we weren't entirely sure about the best action to take. Looks like Turbolink.visit(..., {action: 'replace'}) would make this seamless, similarly to what happens with normal page visits, which is more or less what we want. Yes, this method is pretty much what happens internal in Turbolinks when performing an Application Visit to a page already in the cache: 1. load page from cache 2. visit same location and replace the result. Step 1 is already performed by Turbolinks on popstate Looks like the anchor link behaviour might throw another spanner in the works in this case unfortunately :-/ Yes this could get a bit tricky. Turbolinks currently makes a request to same-page anchors, so you might get away with it, although this is a bug (dealt with in #285), so its behaviour might change. Re adding no-cache - in our case, gon acts as some kind of bridge between server and client, so it can get updated on almost any page. I suppose if you're working on an app with lots of changes in realtime, by adding no-cache in lots of places, you'll be losing a key benefit of Turbolinks. FWIW I work on a messaging app. We use no-cache on the list of "chats", so that if someone is viewing a single chat then navigates Back, the list is updated with the latest excerpt (if that makes sense!). I'm curious if the different behaviour on back/forward compared to normal page visit is intentional, or due to how the browser behaves. Do you happen to know I think this is the intended behaviour and maybe inspired by the default browser behaviour: Turbolinks automatically initiates a restoration visit when you navigate with the browser’s Back or Forward buttons. Applications using the iOS or Android adapters initiate a restoration visit when moving backward in the navigation stack. If possible, Turbolinks will render a copy of the page from cache without making a request. Otherwise, it will retrieve a fresh copy of the page over the network. See Understanding Caching for more details. https://github.com/turbolinks/turbolinks#restoration-visits Thanks again @domchristie! I really appreciate it. We'll need to see what the best approach might be. The documentation is very clear, and matches what we saw in the browser. What I wasn't so sure about was the why, or whether this was the intention. Also, in case we missed something regarding the key events within the turbolinks lifecycle. I can definitely see why it makes sense to avoid another request if you're "just" browsing back into the history. In our case it might introduce some issues, so we were hoping for some degree of control. Looks like using popstate and .visit(..., {action: 'replace'}) are the best options at the moment. Thanks, @domchristie for your suggestion. This could definitely be a solution. Best would be though to have the possibility to define through a configuration if I want to have a "restore" visit when clicking on the back button or a normal visit. Would you guys be interested if we would create a pull request with a possible solution to configure this? /c @domchristie Feel free to explore this :) I'm not 100% sure this is something we'll accept because no-cache is satisfactory in many cases, and although it's not official, or quite the same as an Application visit, the suggested workaround might do the trick. I'll keep this issue open as a place to discuss this, and for others to share their views. Thanks! I understand this may seem like some kind of an edge case, but to some extent I think it's even a saner default. I'll try to give some reasoning. It can eliminate the need for no-cache almost entirely if you know that visiting a cache page would immediately load a fresh version. So you get all the benefits of perceived performance, and still get fresh data quickly enough ... Also it increases consistency in behaviour. Instead of caching behaving differently with back/forward and with clicks on cached pages, all pages behave the same. In any case, we'll look at using the workaround, but we just thought that maybe it's an improvement that others can enjoy (at least as some kind of a configurable option). I'm also very curious what others think about this.
2025-04-01T04:35:47.193239
2016-04-15T21:29:02
148773851
{ "authors": [ "Fudoshiki", "Startouf", "WaKeMaTTa", "dgilperez", "domchristie", "glennfu", "gregblass", "michaelphines", "nateberkopec" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11681", "repo": "turbolinks/turbolinks", "url": "https://github.com/turbolinks/turbolinks/pull/76" }
gharchive/pull-request
Support action: 'back' to navigate back in history I decided to take a stab at this to see how it would work after posting #74. There's still an issue remaining where in the case of calling Turbolinks.visit("http://site.dev/path/here", {"action":"back"}) Firebug reports that http://site.dev/path/here is called once, aborted, then called again. I'm having trouble figuring out why that is. On the turbolinks-ios side, in my ApplicationController.swift#presentVisitableForSession I had to add else if action == .Back { popViewControllerAnimated(true) } and in Turbolinks/Actions.swift I added case Back = "back". With those changes, everything seemed to work perfectly! So aside from the unexplained aborted request, this does exactly what I hoped it would do. I look forward to hearing what you think! If I'm on the right track, let me know and I'll try to put in some appropriate tests and another pull request to turbolinks-ios. I haven't yet gotten far with turbolinks-android but I hope to dive into that soon also. If anyone's trying to implement this on their own in the meantime, the https://github.com/turbolinks/turbolinks-rails gem needs redirection.rb#26 changed to action: ["advance", "replace", "back"].include?(action.to_s) ? action : "replace" So just wondering how I would go about using this. I'm trying: gem 'turbolinks', git: 'https://github.com/glennfu/turbolinks.git', branch: "navBack" But thats not working. Then I'd also have to modify turbolinks-rails redirection.rb#26 as well? @gregblass I actually just clone down the repo and run blade build. I then copy dist/turbolinks.js into the vendor/assets/javascripts folder of my app. This makes it easier to hack on/modify while you're investigating the changes. any tmp solution? @Fudoshiki My Rails app has gem 'turbolinks', github: 'glennfu/turbolinks-rails' in the Gemfile. Then I manually include this in my application.js instead of the original turbolinks.js: https://gist.github.com/glennfu/34bbe7e7b1dc5d4cfeb84c82f423658d I've been running this for 2 months now with no problems. I have next scenario: REST Open form with method: :put (edit action) Enter invalid data Submit Got rendered page with errors (update action) Press F5 request GET on update action (route error) Same issue? gem 'turbolinks', github: 'glennfu/turbolinks-rails' and https://gist.github.com/glennfu/34bbe7e7b1dc5d4cfeb84c82f423658d don't solved this I don't think there's ever a proper setup where F5 would resubmit a form. It's certainly not desirable behavior. At any rate, unless I'm misunderstanding your intent, I don't know what this has to do with this Issue, which is about adding the ability to navigate backward in the stack. In your described scenario, navigating back wouldn't even be the expected behavior. More likely you would, in your controller, tell Turbolinks to redirect with action: 'back' on a successful form submit, then the browser would drop you on the page you were on BEFORE you saw the form. I believe what you are seeing on rendering page errors should be action: 'replace' in order to update the same page you're on, which is already part of Turbolinks without this PR. Just realized one more important change. Line 215 of Session.swift in turbolinks-ios reads visitVisitable(visitable, action: .Restore). This should be changed to visitVisitable(visitable, action: .Back) I just pushed a commit that fixes a bug I discussed here: https://github.com/turbolinks/turbolinks-ios/issues/50 Hey @nateberkopec, any opinion on my last message about "back" vs "go_back" vs "refresh"? I'm less excited about "refresh" because this PR's behavior is about simulating the "back" button, and "refresh" is synonymous with "reload". Also, gotta throw in there that Webster says "back" is a verb too :) Let me know what word we should use here and I'll update this PR. @glennfu Probably just go_back "to back", back's verb form, as in "to back a candidate" doesn't really fit here IMO. but that's just my opinion, the maintainers of this repo I know usually have really strong opinions on readability/naming. It's also possible that they're thinking up some different solution entirely for this problem (though your PR still helps by proposing one). retreat? I just noticed the only other context this "verb" is used in the code is in the backHistoryWithLocationAndRestorationIdentifier which is analogous to the replaceHistoryWithLocationAndRestorationIdentifier and pushHistoryWithLocationAndRestorationIdentifier methods. With that in mine, rewindHistory and reverseHistory sound pretty good. I don't know that goBackHistory sounds good. retreatHistory is a solid "maybe" :) @sstephenson @gregblass @packagethief any opinions on this verb? I'm happy to update the entire working and rebased solution once I feel like we generally have a good name for it. It would be awesome if we can get it right the first time because this verb will exist in turbolinks, turbolinks-ios, turbolinks-android, and turbolinks-rails, so refactoring it later if we decide we don't like it will be really annoying. backward :question: (backwardHistory & backwardHistoryWithLocationAndRestorationIdentifier) What about rewind ? Has anyone suggested popHistory yet? +1 to rewind. go_back or step_back would be also OK IMO. any progress? With that in mind, rewindHistory and reverseHistory sound pretty good. I don't know that goBackHistory sounds good. retreatHistory is a solid "maybe" :) … any opinions on this verb? @glennfu I don't feel that back is that bad. It may not be consistent with the other verbs, but it is clear and straightforward. For me, the issue with rewind and reverse is that they sound like they could transport the user to any point in a continuous timeline (like rewinding a cassette, or reversing a car), rather than traveling to the last discrete step. However, with this in mind, I wonder if this approach could be expanded to navigate to any location in the stack, rather than just the previous one? For example, given the following stack: [Index] [Show] [Edit] then calling Turbolinks.visit('Index', { action: 'back' }) could result in the following stack: [Index] This could be a solution to the "reset" case mentioned here: https://github.com/turbolinks/turbolinks/issues/74#issuecomment-213499482, but I can't work out if this might cause confusion in some other cases.
2025-04-01T04:35:47.196408
2021-03-26T12:59:21
841936758
{ "authors": [ "e-gineer", "rajmohanty17" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11682", "repo": "turbot/steampipe-plugin-aws", "url": "https://github.com/turbot/steampipe-plugin-aws/issues/261" }
gharchive/issue
Decoding of secrets rendered in the encoded user_data for EC2. Is your feature request related to a problem? Please describe. Find secrets in EC2 User Data is one of the ISO-27001 requirement. Describe the solution you'd like Describe alternatives you've considered A clear and concise description of any alternative solutions or features you've considered. Additional context Autoscaling - https://docs.aws.amazon.com/autoscaling/ec2/userguide/create-launch-config.html EC2 - As an initial suggestion, you could use postgres regexp tests against the user_data field. Primitive, but would be a start.
2025-04-01T04:35:47.204479
2023-07-07T12:24:44
1793460533
{ "authors": [ "graza-io", "misterek" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11683", "repo": "turbot/steampipe-plugin-github", "url": "https://github.com/turbot/steampipe-plugin-github/issues/297" }
gharchive/issue
bypass_pull_request_allowance_* is not populated. Describe the bug The list of teams/people "Allow specified actors to bypass required pull requests" does not appear to be populated. Steampipe version (steampipe -v) Steampipe v0.20.7 Plugin version (steampipe plugin list) 0.28.1 (Also installed from repo) To reproduce Select "Allow specified actors to bypass required pull requests" in branch protections. Add a user to the list. Then steampipe query "select default_branch_ref from github_my_repository where name=reponame" (I see the same results with github_repository) Expected behavior Expect "bypass_pull_request_allowance_users" or "bypass_pull_request_allowance_teams" to be populated somewhere. (https://github.com/turbot/steampipe-plugin-github/blob/main/github/table_github_branch_protection.go#L58) Additional context It appears as if it's supposed to happen here: https://github.com/turbot/steampipe-plugin-github/blob/main/github/table_github_branch_protection.go#L305 If anyone can confirm, I can attempt to take a look when I have time, though I'm unfamiliar with the GraphQL API. For reference, this is the specific feature I am referring to: https://github.blog/changelog/2021-11-19-allow-bypassing-required-pull-requests/ Hi @misterek - sorry to hear you're hitting issues! Ideally, yes the github_branch_protection table would be used to get the information on branch protections, whilst some is available directly from the repo table for convenience, we omitted including anything which would need to page nested resources to obtain data such as users/apps/teams which can bypass PR for force push rules, etc. I added myself to the relevant section on a repo as a test: When querying the results I can see I am correctly added there. If you're seeing anything different would you be able to confirm your membership level on the repository in question as well as that your GH PAT has the following permissions: gist, read:org, read:user, repo, user:email Interesting! Apparently I was looking at the completely wrong table. When querying github_branch_protection, I do see it there. I think what I'd maybe say is it's slightly confusing, because when querying "default_branch_ref", it includes a number of branch protection rules (and is actually labeled "branch_protection_rule"). But, I'm now seeing that that's an incomplete set of branch protection rules. Anyway, I am completely unblocked and will close this issue. Thank yo so much!
2025-04-01T04:35:47.222395
2024-03-03T02:34:05
2165087501
{ "authors": [ "JamesChenX", "zhaieryuan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11684", "repo": "turms-im/turms", "url": "https://github.com/turms-im/turms/issues/1407" }
gharchive/issue
how to config plugin config Properties in yml , please for example ! 1、我把敏感词汇的配置放在 turms-plugin-antispam 的 resources 目录application.yaml下,服务启动turms_service ,报错: APPLICATION FAILED TO START Description: Binding to target [Bindable@5d8ab698 type = im.turms.server.common.infra.property.TurmsProperties, value = 'provided', annotations = array[@org.springframework.boot.context.properties.ConfigurationProperties(ignoreInvalidFields=false, ignoreUnknownFields=false, prefix="turms", value="turms")], bindMethod = JAVA_BEAN] failed: Property: turms.plugin.antispam.dictparsing.skipinvalidcharacter Value: "true" Origin: class path resource [application.yaml] - 55:31 Reason: The elements [turms.plugin.antispam.dictparsing.skipinvalidcharacter,turms.plugin.antispam.dictparsing.textfilecharset,turms.plugin.antispam.dictparsing.textfilepath,turms.plugin.antispam.enabled] were left unbound. Property: turms.plugin.antispam.dictparsing.textfilecharset Value: "UTF-8" Origin: class path resource [application.yaml] - 54:26 Reason: The elements [turms.plugin.antispam.dictparsing.skipinvalidcharacter,turms.plugin.antispam.dictparsing.textfilecharset,turms.plugin.antispam.dictparsing.textfilepath,turms.plugin.antispam.enabled] were left unbound. Property: turms.plugin.antispam.dictparsing.textfilepath 2、同上我把配置放在 turms-service 的 resources 目录application.yaml下,服务启动turms_service ,也是报错如上。 3、具体我需要配置的参数如下: turms: plugin: antispam: enabled: true dictParsing: textFilePath: /turms_ubuntu/turms/csv_files/in_use.csv textFileCharset: UTF-8 skipInvalidCharacter: true 4、请问当前服务抽层到 turms-server-common 服务后,idea 的自动加载配置,又需要怎么解决?作者有考虑么?按照官网的设置,不生效的,应为turms-server-common服务迁走了。 我把敏感词汇的配置放在 turms-plugin-antispam 的 resources 目录application.yaml下,服务启动turms_service ,报错: The error logs mean the properties you specified don't exist. We design turms servers to throw when encountering unknown properties so that users can know they have configured the wrong properties. It is correct that you configure turms-anti-spam properties under turms-service (because turms-anti-spam is a plugin for turms-service). But the properties name you configured is wrong. The correct naming style is shown in the following image: I checked the doc I wrote for turms-plugin-antispam, and I think you may followed the doc that uses the wrong naming style for properties. hhh. I will correct the doc later. 请问当前服务抽层到 turms-server-common 服务后,idea 的自动加载配置,又需要怎么解决?作者有考虑么?按照官网的设置,不生效的,应为turms-server-common服务迁走了。 It seems you think turms-server-common is a service (i.e. a server), but it isn't. turms-server-common is not a service, not an application, but is a library that is shared and used by turms servers (e.g. turms-gateway, turms-service, turms-ai-serving, and future servers), so turms-server-common itself cannot run. Thank you very much for your reply. ---- Replied Message ---- | From | @.> | | Date | 03/03/2024 11:39 | | To | turms-im/turms @.> | | Cc | zhaieryuan @.>, Author @.> | | Subject | Re: [turms-im/turms] how to config plugin config Properties in yml , please for example ! (Issue #1407) | 我把敏感词汇的配置放在 turms-plugin-antispam 的 resources 目录application.yaml下,服务启动turms_service ,报错: The error logs mean the properties you specified don't exist. We design turms servers to throw when encountering unknown properties so that users can know they have configured the wrong properties. It is correct that you configure turms-anti-spam properties under turms-service (because turms-anti-spam is a plugin for turms-service). But the properties name you configured is wrong. The correct naming style is shown in the following image: image.png (view on web) I checked the doc I wrote for turms-plugin-antispam, and I think you may followed the doc that uses the wrong naming style for properties. hhh. I will correct the doc later. 请问当前服务抽层到 turms-server-common 服务后,idea 的自动加载配置,又需要怎么解决?作者有考虑么?按照官网的设置,不生效的,应为turms-server-common服务迁走了。 It seems you think turms-server-common is a service (i.e. a server), but it isn't. turms-server-common is not a service, not an application, but is a library that is shared and used by turms servers (e.g. turms-gateway, turms-service, turms-ai-serving, and future servers), so turms-server-common itself cannot run. — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you authored the thread.Message ID: @.***>
2025-04-01T04:35:47.246893
2023-03-22T09:05:22
1635338819
{ "authors": [ "VishnuVelayuthan", "ipa-rar", "roni-kreinin" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11685", "repo": "turtlebot/turtlebot4_setup", "url": "https://github.com/turtlebot/turtlebot4_setup/issues/5" }
gharchive/issue
unable to SSH after running turtlebot_setup.sh script Please provide the following information: OS: (e.g. Ubuntu 22.04) ROS Distro: (e.g. Humble) installed: Expected behaviour ssh back into the turtlebot4 after installing and rebooting Actual behaviour static IP changes unable to establish ssh connection with the new IP address Robot does not respond after rebooting To Reproduce Provide the steps to reproduce: run wget -qO - https://raw.githubusercontent.com/turtlebot/turtlebot4_setup/humble/scripts/turtlebot4_setup.sh | bash in this line https://github.com/turtlebot/turtlebot4_setup/blob/038d6f881ed56748d5727bc43c449084a628e591/scripts/turtlebot4_setup.sh#L48 the netplan yaml is removed and this might cause the issue Hey, suffering from the same bug^^^^ I can't ssh as well. I'm also having an issue where the different ros packages are also not being able to be located by apt. I am unable to run turtlebot4-setup as my computer is saying it is not a command @VishnuVelayuthan @roni-kreinin I have found a workaround for this problem. I stopped following the documentation found in the readme and this helped me a lot. Especially this wget -qO - https://raw.githubusercontent.com/turtlebot/turtlebot4_setup/humble/scripts/turtlebot4_setup.sh | bash command caused the ssh to break as it simply removed the netplan yaml file sudo rm /etc/netplan/50-cloud-init.yaml. https://github.com/turtlebot/turtlebot4_setup/tree/humble-readme This provides more steps on configuring but this is also not fail proof. I will write a documentation if I succeed in setting this up and would send a PR. And turtlebot4-setup is not a command and its an alias which in turn should run ros2 run turtlebot4_setup turtlebot4_setup. You can check the aliases here. Honestly, turtlebot is something one can use out of box but this version is very buggy and not an out of the box usable robot or to be specific it is not really fit for multi robot use case. PS: I own four Turtlebot4 robots @ipa-rar The turtlebot4_setup.sh script is used to set up the actual humble image from a clean ubuntu 22.04 server image. Unless you want to make a modified version of the turtlebot4 image, I would recommend just downloading the built image from http://download.ros.org/downloads/turtlebot4/ and flashing that. I have flashed the prebuilt image using the documentation provided in the Readme but when the robot boots up it boots into BusyBox as shown here. Then flashing the card using Balena Etcher helped to boot up the robot with the new image.
2025-04-01T04:35:47.307030
2024-01-18T16:18:17
2088627087
{ "authors": [ "jwhite", "klousf" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11686", "repo": "tuya/tuya-smart-life", "url": "https://github.com/tuya/tuya-smart-life/issues/93" }
gharchive/issue
Air Humidifier CA-605B Smart (unsupported) by smartlife I added an CA-605B to the smart life app on my iPhone and that is working fine. I then tried to try the same to the smartlife integration on my Home Assistant (via HACS) and the device is added through the QR-code but as unsupported and with no sensors at all. I am really new to this, but is this fixable, or will it be added in the future? You should paste the diagnostic info in this thread. It might be possible for someone to create a pull request to support this.
2025-04-01T04:35:47.310357
2019-10-22T07:46:51
510484892
{ "authors": [ "Liys0558", "ex-yangguoqing002", "sara782" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11687", "repo": "twangnh/Distilling-Object-Detectors", "url": "https://github.com/twangnh/Distilling-Object-Detectors/issues/9" }
gharchive/issue
lack of nms_cuda.c hi, thanks for your contributions , and when i compile the code of lib and find the error that where is no nms_cuda.c in src of nms folder, please help me, thanks i have the same problem, have you solved i now? i have solved this problem just now! add https://github.com/jwyang/faster-rcnn.pytorch/blob/master/lib/model/nms/src/nms_cuda.c at the error place i have solved this problem just now! add https://github.com/jwyang/faster-rcnn.pytorch/blob/master/lib/model/nms/src/nms_cuda.c at the error place Hello! I know I'm late to this, but I've tried your solution and it's still not working. Any other suggestions?
2025-04-01T04:35:47.321171
2015-02-18T21:12:46
58128830
{ "authors": [ "cvrebert", "lukeberry99", "orrd", "troywarr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11688", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/15840" }
gharchive/issue
Input group addons are misaligned/mismatched height Input group addons render at a mismatched height, or are horizontally misaligned with the adjacent input fields (such that a double border is visible). This appears to affect both the input-group-addon and input-group-btn classes, with various types of text/icon content within the addons. To reproduce: In Chrome/Mac v40.0.2214.111, visit http://getbootstrap.com/components/#input-groups Scroll down to the examples under the "Basic example," "Sizing," "Button addons," "Buttons with dropdowns," or "Segmented buttons" headings Results: The examples in these sections shown misalignment in the addons. This occurs in my project as well (not just the Bootstrap docs), and the addons don't appear to be picking up any styling beyond what comes out of the box in Bootstrap. This issue may be related to #10936, although that appears to concern Glyphicon content. This issue is not limited to that. Screenshots: I'm not able to replicate this problem at all, also running Chrome 40.0.2214.111 on latest OSX @troywarr Can you please confirm that your browser zoom level is at 100% (i.e. no zoom)? This does not appear to be a Bootstrap issue. Closing, and sorry for the hassle. For posterity - this actually appears to be a Chrome zoom bug. Not sure if it's in vanilla Chrome, or caused by an extension I'm running. From my experience, this type of inconsistent misalignment is usually caused by browser zoom. In fact, that was the first thing I tried - using Command 0 in Chrome to zoom to actual size. Turns out that, at least when I'm zoomed in to 110%, Chrome's "Actual Size" doesn't do anything! (Or at least not what I'd expect - return to 100%.) I had to zoom out to 100% using Command -. I've been developing in zoom level 110% for who knows how long now. :-/ Yeah, as mentioned in our docs (http://getbootstrap.com/getting-started/#support-browser-zooming ), non-pinch browser page zooming breaks the Web generally, and Bootstrap doesn't attempt any heroics to try to support it. Definitely - and I think that's the only reasonable policy. :+1: Thanks for the help, and sorry again for the false alarm. Just for the benefit of other people who may find this same issue happening even when they are at 100% zoom mode, there is an issue with Bootstrap SASS that will cause a very similar problem. See https://github.com/twbs/bootstrap-sass/issues/409 for a solution involving setting the SASS precision.
2025-04-01T04:35:47.322468
2015-08-19T20:19:51
101989274
{ "authors": [ "MichaelRushton", "cvrebert" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11689", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/17055" }
gharchive/issue
Microsoft Edge support for v4 The supported browsers section doesn't reference Microsoft Edge. It didn't yet exist under its current name the last time we touched that part of the docs :smile: Fixed by #17059.
2025-04-01T04:35:47.324232
2016-03-27T08:17:49
143787635
{ "authors": [ "arkroy" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11690", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/19630" }
gharchive/issue
Bootstrap JS not working I'm trying to create Bootstrap project with BrowserSync, however it seems the JS isn't working. The project files can be found here<EMAIL_ADDRESS>I've tried pulling from bower_components directly, CDN and by just copying the individual files into the dist/js folder. Any help would be much appreciated. Sorry, I guess it was uncss removing style classes to be appended through JavaScript
2025-04-01T04:35:47.326670
2016-03-31T08:51:25
144830762
{ "authors": [ "RyanZim", "beingsagir", "cvrebert", "kutec" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11691", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/19645" }
gharchive/issue
Justified nav is not working on chrome ! Here is the output of a theme sample on Chrome! Example name : Justified nav If you are referring to a theme at http://themes.getbootstrap.com, please do not report it here. See: CONTRIBUTING.md#using-the-issue-tracker. Duplicate of #17723. Please search more next time. Not only Chrome in all browsers it's not working!
2025-04-01T04:35:47.384143
2017-04-09T18:02:47
220487496
{ "authors": [ "geminiX", "mdo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11692", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/22399" }
gharchive/issue
list-inline & col-offset not working I am a newbie to GitHub and programming in general, so I apologize for any errors in the following report. I am creating a page on CodePen and the list-inline and column offset classes are not working properly in v4.0.0 alpha 6, however if i switch to Bootstrap 3.3.7 it is working as expected. Links to Bootstrap CSS files I am using: https://maxcdn.bootstrapcdn.com/bootstrap/3.3.7/css/bootstrap.min.css https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0-alpha.6/css/bootstrap.min.css My CodePen project: https://codepen.io/geminix/pen/PpMNzz v4 has different HTML, CSS, and JS, so you might need to change a few things around. Take a look at the v4 docs at http://v4-alpha.getbootstrap.com. The grid classes have changed for offsets (now .offset-*), for example.
2025-04-01T04:35:47.385973
2018-07-02T16:15:10
337581220
{ "authors": [ "Maaacoooo", "mdo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11693", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/26789" }
gharchive/issue
Backwards Compatibility I'm sure everyone relates. But goddamn, why does it have to rename the classes? I have to rewrite almost everything just to apply the 4.1 Update. What classes were renamed from v4.0 to v4.1? @Maaacoooo Let us know if there's something we can help with here. Minor releases add new features, but shouldn't introduce breaking changes from the v4.0 stable release.
2025-04-01T04:35:47.388581
2018-09-27T08:22:52
364357537
{ "authors": [ "MartijnCuppens", "andresgalante", "brnrd87" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11694", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/27327" }
gharchive/issue
Form input does not inherit font-weight If the variable $font-weight-normal is set to a non-default value, e.g. "300", this value is not inherited by text inputs. The computed value of the font-weight property is found to be 'reset' to 400 in the developer tools (Firefox 62.0). When you add 'font-weight: inherit;' to the text input, this is solved. Another solution could be to add a $input-font-weight variable. The case where font weights of 300 and 400 are available might indeed be a situation where this might come in handy. If we continue on this, we need to make sure we'll also add this for .custom-select. The user agent in chrome will also reset the font weight to 400. I'll fix this.
2025-04-01T04:35:47.405414
2021-09-19T17:25:35
1000370574
{ "authors": [ "Nzamuyejoram", "XhmikosR", "baijifeilong", "bjermeland", "danielmatthew", "heydarm", "kris2kris" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11695", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/35018" }
gharchive/issue
Importing bootstrap.scss breaks SASS compilation Operating system and version: WIndows 10 Browser and version: Chrome 93 Suggested fix: Use math.max() instead of max(). I have create-react-app project (react-scripts@4.0.3). It works if I do npm start, but not when npm run build. This error occurs if you're using Dart SASS 1.40.0 and higher. SassError: 0.2rem and 0 are incompatible. ╷ 9 │ $return: append($return, max($value, 0)); │ ^^^^^^ 0.2rem │ ━ 0 ╵ node_modules\bootstrap\scss\mixins\_border-radius.scss 9:36 valid-radius() node_modules\bootstrap\scss\mixins\_border-radius.scss 20:20 border-radius() node_modules\bootstrap\scss\_reboot.scss 322:3 @import node_modules\bootstrap\scss\bootstrap.scss 17:9 @use src\styles\custom.scss 1:1 @use src\styles\index.scss 1:1 root stylesheet That error was supposedly resolved in 1.40.1. However, while that particular error has been resolved, something else within the Bootstrap partials results in a build fail across the following versions – when using CRA. 1.40.1 1.41.0 1.41.1 1.42.0 Dropping Sass back to 1.39.2 means it works again. > Executing task: yarn run build < yarn run v1.22.11 $ react-scripts build Creating an optimized production build... Failed to compile. ./src/index.scss ParserError: Syntax Error at line: 1, column 25 // index.scss @import "bootstrap/scss/bootstrap"; { "name": "bootstrap-cra-test", "version": "0.1.0", "private": true, "dependencies": { "@testing-library/jest-dom": "^5.11.4", "@testing-library/react": "^11.1.0", "@testing-library/user-event": "^12.1.10", "bootstrap": "5.1.1", "react": "^17.0.2", "react-dom": "^17.0.2", "react-scripts": "4.0.3", "web-vitals": "^1.0.1" }, "devDependencies": { "sass": "1.42.0" }, "scripts": { "start": "react-scripts start", "build": "react-scripts build", "test": "react-scripts test", "eject": "react-scripts eject" }, "eslintConfig": { "extends": [ "react-app", "react-app/jest" ] }, "browserslist": { "production": [ ">0.2%", "not dead", "not op_mini all" ], "development": [ "last 1 chrome version", "last 1 firefox version", "last 1 safari version" ] } } Were I to delve into node_modules/bootstrap/bootstrap.scss, and comment out everything under "Layout & Components", the project builds successfully. Narrowed it down to here: https://github.com/twbs/bootstrap/blob/60d3eb34f8a1aabe5d2a2308bb841f30cf16c9c4/scss/_grid.scss#L5-L13 Narrowed it down to here: https://github.com/twbs/bootstrap/blob/60d3eb34f8a1aabe5d2a2308bb841f30cf16c9c4/scss/_grid.scss#L5-L13 Hello, I have created a ticket in sass (I don't know if it is correct) because I have identified the problem https://github.com/sass/dart-sass/issues/1496 I think what makes this gnarlier to suss out is that a basic set-up compiles with no errors thrown: { "name": "bootstrap-test", "version": "1.0.0", "description": "", "scripts": { "start": "npx sass style.scss style.css --load-path=node_modules" }, "author": "", "license": "ISC", "devDependencies": { "sass": "1.42.0" }, "dependencies": { "bootstrap": "5.1.1" } } @import "bootstrap/scss/bootstrap"; Still researching, but it could be because of sass-loader that create-react-app uses It seems like SASS fixed it in the latest version (1.42.0) It seems like SASS fixed it in the latest version (1.42.0) Checked locally, it really works now For me it does not work with 1.42.0, 1.39.2 is the last version I can use. A temporary fix is to wrap the negative values in parenthesis. margin-top: calc(var(--#{$variable-prefix}gutter-y) * (-1)); // stylelint-disable-line function-disallowed-list margin-right: calc(var(--#{$variable-prefix}gutter-x) * (-.5)); // stylelint-disable-line function-disallowed-list margin-left: calc(var(--#{$variable-prefix}gutter-x) * (-.5)); // stylelint-disable-line function-disallowed-list @kris2kris so great . nice for the success Uninstall sass, install node-sass. Uninstall sass, install node-sass. On the GitHub page of node-sass it's written than node-sass is deprecated and to use dart-sass instead This is an issue with postcss-values-parser, see https://github.com/shellscape/postcss-values-parser/issues/138. We have a #35033 which is supposed to work around the issue.
2025-04-01T04:35:47.409052
2022-11-13T17:28:08
1447012234
{ "authors": [ "julien-deramond", "mpaglia0" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11696", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/issues/37489" }
gharchive/issue
Deviantart iconSuggest a new feature Prerequisites [X] I have searched for duplicate or closed feature requests [X] I have read the contributing guidelines Proposal Can you please draw Deviantart icon? Motivation and context Deviantart is a big and active community of artist Thanks for creating this issue @mpaglia0. For the icons, there is a specific repository where such a request is already in the backlog: https://github.com/twbs/icons/issues/1388. I am sorry!! Thank you for the info
2025-04-01T04:35:47.409974
2015-11-13T22:17:56
116864479
{ "authors": [ "cvrebert", "vsn4ik" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11697", "repo": "twbs/bootstrap", "url": "https://github.com/twbs/bootstrap/pull/18246" }
gharchive/pull-request
[v3] Use border-radius mixin in button-groups.less After merge commit ef8bc288902fc3e5f0b8b5745a3a30e0fab22f76 (PR #16684). Thanks!
2025-04-01T04:35:47.413385
2022-07-06T18:13:46
1296201341
{ "authors": [ "mdo", "sunergeo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11698", "repo": "twbs/examples", "url": "https://github.com/twbs/examples/issues/6" }
gharchive/issue
Importing bootstrap into Vue3/Vite app I have submitted a pull request at https://github.com/twbs/examples/pull/5 to illustrate the issue I have encountered. I am having trouble getting bootstrap to import into a basic Vue3/Vite app. I have studied the other examples in this repository, and cannot figure out what I am missing. When the project builds, the following error is generated: ENOENT: no such file or directory, stat '[base path]/vite-vue/node_modules/bootstrap/scss/bootstrap' This is coming from the vite:css plugin, and I can find no information about how to configure that plugin correctly. I have verified my vite.config.js has the correct ~bootstrap alias by hard-coding the path to node_modules/bootstrap/scss/bootstrap, and the same error occurs. If I reference the _bootstrap.scss file directly, another error occurs because the inline comments are unrecognized. Any assistance would be greatly appreciated! Your example isn't in line with the others we have in this a repo, so I've closed the PR for now. We could add a Vue example here, but I'm not sure when we'll get around to it. When we do, there's no guarantee it'll be the setup that you need :). For now, consider asking around more as you've done in the main repo's discussions. Ok, no problem ... I thought this would be helpful. See #13.
2025-04-01T04:35:47.414770
2014-03-07T06:55:36
28941807
{ "authors": [ "XhmikosR", "fat" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11699", "repo": "twbs/ratchet", "url": "https://github.com/twbs/ratchet/issues/447" }
gharchive/issue
moz demos if we want the demos to work on mozilla we'll have to adapt the plugins to work with firefox prefixes… currently all the js uses -webkit prefixes… This should work fine now with the current master.
2025-04-01T04:35:47.421822
2016-06-13T17:49:17
160007215
{ "authors": [ "Aaron1011", "Harmon758", "kmario23" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11700", "repo": "tweepy/tweepy", "url": "https://github.com/tweepy/tweepy/issues/745" }
gharchive/issue
TweepError: failed to send request It was executing perfectly but all of a sudden it started throwing this error, though I did not make any changes to the codebase. searched_tweets = ( status._json for status in tweepy.Cursor(api.search, q=query, count=300, since=from_date, until=to_date, File "build/bdist.linux-x86_64/egg/tweepy/cursor.py", line 197, in next File "build/bdist.linux-x86_64/egg/tweepy/cursor.py", line 108, in next File "build/bdist.linux-x86_64/egg/tweepy/binder.py", line 245, in _call File "build/bdist.linux-x86_64/egg/tweepy/binder.py", line 189, in execute tweepy.error.TweepError: Failed to send request: 'module' object has no attribute 'HTTPMessage' Any help is highly appreciated! Can you reproduce this with the latest Tweepy from Github (run python setup.py install to replace the installed version from PyPi). This seems to have been an issue with the requests dependency's urllib3 dependency having issues when installed along with specific versions of future (see https://github.com/elastic/elasticsearch-py/issues/253 and https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=796356) rather than an issue with Tweepy. However, you seem to have figured this out already.
2025-04-01T04:35:47.441160
2019-10-19T19:45:25
509500928
{ "authors": [ "kwhinnery", "vinc456" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11701", "repo": "twilio-labs/open-pixel-art", "url": "https://github.com/twilio-labs/open-pixel-art/issues/1312" }
gharchive/issue
TwilioQuest AppImage fails to run on Kali Linux I was not able to run TwilioQuest AppImage on Kali Linux. 5.2.0-kali3-amd64 Distributor ID: Kali Description: Kali GNU/Linux Rolling Release: 2019.4 Codename: kali-rolling I'm not very familiar with AppImage but I did change the executable bit and tried to run the file, only to encounter the following error. ./TwilioQuest.AppImage [28482:1019/054935.507976:FATAL:setuid_sandbox_host.cc(157)] The SUID sandbox helper binary was found, but is not configured correctly. Rather than run without sandboxing I'm aborting now. You need to make sure that /tmp/.mount_TwiliozhGxvg/chrome-sandbox is owned by root and has mode 4755. Trace/breakpoint trap $ ./TwilioQuest.AppImage [28889:1019/055402.790344:FATAL:setuid_sandbox_host.cc(157)] The SUID sandbox helper binary was found, but is not configured correctly. Rather than run without sandboxing I'm aborting now. You need to make sure that /tmp/.mount_TwiliorD2I5Z/chrome-sandbox is owned by root and has mode 4755. Didn't investigate further, hopefully PEBKAC. Hm, a quick google search of the error seems to indicate that there are permissions issues at play - check this comment: https://github.com/electron/electron/issues/17972#issuecomment-487369441 A suggestion within is to execute the AppImage with the --no-sandbox option - would be worth a shot I think? Otherwise, changing ownership and permissions of the AppImage file may help. Sorry, TBH I am still a bit of a n00b when it comes to Linux + Electron.
2025-04-01T04:35:47.444717
2021-07-13T11:13:52
943198793
{ "authors": [ "nsmrystr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11702", "repo": "twilio-labs/open-pixel-art", "url": "https://github.com/twilio-labs/open-pixel-art/pull/3906" }
gharchive/pull-request
feat(pixels): add my new pixel Checklist [x] I ran npm test locally and it passed without errors. [x] I only edited the _data/pixels.json file. [x] I entered the username in the pixels.json that I'm also using to create this pull request. [x] I acknowledge that all my contributions will be made under the project's license. close
2025-04-01T04:35:47.494728
2022-01-28T14:11:54
1117452102
{ "authors": [ "LaurenceJJones", "vasjan04" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11703", "repo": "twintproject/twint", "url": "https://github.com/twintproject/twint/issues/1340" }
gharchive/issue
Any idea how can I solve this If you used the search and search for keywords like CeilTimeout you find the fix https://github.com/twintproject/twint/issues/1297#issuecomment-958105596
2025-04-01T04:35:47.499765
2022-01-24T20:26:49
1113106424
{ "authors": [ "craigpastro", "wmatveyenko" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11704", "repo": "twitchtv/twirp", "url": "https://github.com/twitchtv/twirp/issues/348" }
gharchive/issue
Broken link in docs 👋 In the docs, on the "Usage Example: Haberdasher" page, the next link at the bottom, which should go to "Best Practices", is broken and leads to a 404. I had a look at how this could be fixed, but I am not very familiar with docusaurus, so I ended up giving up after some minutes. Sorry about that! Thanks. We regenerated the documentation site and this link should be fixed now.
2025-04-01T04:35:47.502600
2012-02-21T07:33:17
3307583
{ "authors": [ "MechanisM", "markdotto", "mneuhaus" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11705", "repo": "twitter/bootstrap", "url": "https://github.com/twitter/bootstrap/issues/2126" }
gharchive/issue
Carousel in jcarousel mode? [feature request] Look at this http://jsfiddle.net/MechanisM/jjNUk/show/ it's looks ugly coz I need to add 4 items inside one carousel item. maybe you'll improve bootstrap-carousel.js? to let it work like jcarousel and show more than one item in one time. also will be great to have vertical mode. It looks like you do have it working with 4 items within 1. What would you suggest we change to improve this for everyone? Well, just wanted grid based items inside carousel. let's say if width of carousel = span12, then I can see 2 * span6 blocks inside it, 3 * span4 blocks, 4 * span3 blocks etc. currently the carousel only allows 1 item to be active. for me a that is a slider. a carousel can show multiple items and animate through them. Like in the jsfiddle above. You have 4 items visible. a click on the right button should move the hole thing 1 item to the left. +1 :)
2025-04-01T04:35:47.505857
2011-11-24T09:33:57
2339138
{ "authors": [ "fat", "jenswaern" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11706", "repo": "twitter/bootstrap", "url": "https://github.com/twitter/bootstrap/issues/685" }
gharchive/issue
Twipsy on pages with horizontal scrollbar In chrome (only browser I tested) when an item is to the right of shown content when page id loaded the 'stem' is not centered properly, see here: http://cl.ly/1e1p1N3O1Z3k45330U2E hm.. interesting - could you reproduce that in a jsfiddle? the issue is because the text is wrapping and the sizing is getting off - if you just add a white-space: nowrap rule to the twipsy definition it seems to work: http://jsfiddle.net/QXQTa/14/ Right you are! Cheers!
2025-04-01T04:35:47.554758
2018-01-29T12:10:56
292384129
{ "authors": [ "n8downs", "paullaffitte", "pleerock", "sapkra", "tanakahisateru" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:11707", "repo": "twitter/twemoji", "url": "https://github.com/twitter/twemoji/issues/272" }
gharchive/issue
Keycap emoji sequence not detected Below characters are not detected by twemoji.parse() because only their sequences start with 7bit ASCII characters + U+FE0F, I guess. #️⃣ U+0023 U+FE0F U+20E3 keycap: # *️⃣ U+002A U+FE0F U+20E3 keycap: * 0️⃣ U+0030 U+FE0F U+20E3 keycap: 0 1️⃣ U+0031 U+FE0F U+20E3 keycap: 1 2️⃣ U+0032 U+FE0F U+20E3 keycap: 2 3️⃣ U+0033 U+FE0F U+20E3 keycap: 3 4️⃣ U+0034 U+FE0F U+20E3 keycap: 4 5️⃣ U+0035 U+FE0F U+20E3 keycap: 5 6️⃣ U+0036 U+FE0F U+20E3 keycap: 6 7️⃣ U+0037 U+FE0F U+20E3 keycap: 7 8️⃣ U+0038 U+FE0F U+20E3 keycap: 8 9️⃣ U+0039 U+FE0F U+20E3 keycap: 9 https://unicode.org/emoji/charts-5.0/emoji-list.html#keycap Additionally, also eye-with-balloon is so. 👁‍🗨 (U+1F441 U+FE0F U+200D U+1F5E8 U+FE0F) converted to 👁 and 💬 https://unicode.org/emoji/charts-5.0/emoji-list.html#1f441_fe0f_200d_1f5e8_fe0f Issue persists for the two years... You can easily check there are no mentioned emojis in https://github.com/twitter/twemoji/tree/gh-pages/v/13.0.0/svg release a few days ago. The naming of the files are wrong: For example: #️⃣ Current naming: 23-20e3.svg Correct naming: 0023-fe0f-20e3.svg The same problem is occurring for: ©️ (copyright) ®️ (registered) 👁️‍🗨️ (eye in speech bubble https://github.com/twitter/twemoji/issues/405) and those named above The list for the right names can be found here: https://unicode.org/Public/emoji/13.0/emoji-test.txt I have (I think) the same issue with 🖌️, which is parsed as 1f58c-fe0f, but the corresponding image is available under the name 1f58c.png @n8downs Is there any plan to fix this soon? Because our parser is generous in its matching of unqualified emoji sequences, we strip out VS16s (fe0f) for the purpose of creating asset URLs. We do not have any plans to change this design, sorry! Actually, it looks that twemoji.parse is working as expected, and twemoji.convert.toCodePoint too, but only produce the valid codepoint, which is not the valid filename. You can also use twemoji-parser. Here is an example: import { parse } from 'twemoji-parser'; import twemoji from 'twemoji' const emoji = '1️⃣'; console.log( parse(emoji), // Outputs [{"url":"https://twemoji.maxcdn.com/v/latest/svg/31-20e3.svg","indices":[0,3],"text":"1️⃣","type":"emoji"}] twemoji.parse(emoji), // Outputs <img class="emoji" draggable="false" alt="1️⃣" src="https://twemoji.maxcdn.com/v/12.1.6/72x72/31-20e3.png"/> twemoji.convert.toCodePoint(emoji) // Outputs 31-fe0f-20e3 ); I was using twemoji.convert.toCodePoint, that's why it wasn't working. But I still wonder why it would not work for you too @tanakahisateru.