id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1718099066
🛑 CHAT Website is down In 4777ea1, CHAT Website (https://chat.devf5r.com) was down: HTTP code: 0 Response time: 0 ms Resolved: CHAT Website is back up in 3509cad.
gharchive/issue
2023-05-20T08:27:30
2025-04-01T06:44:31.208382
{ "authors": [ "info-devf5r" ], "repo": "info-devf5r/VPN", "url": "https://github.com/info-devf5r/VPN/issues/196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
453632616
Add rid template button for md editor This issue in ermrestJS adds support for an rid template in markdown. A button should be added to our markdown editors that inputs [[<rid>]] into the form. The corresponding changes for this issue have been merged into master. A button was added to the editor input. Information was also added to the mdHelp page explaining how this new template can be used. Reassigning to @hongsudt.
gharchive/issue
2019-06-07T18:07:25
2025-04-01T06:44:31.243687
{ "authors": [ "jrchudy" ], "repo": "informatics-isi-edu/chaise", "url": "https://github.com/informatics-isi-edu/chaise/issues/1788", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
174637731
Reference.read default page size This PR, adds default value for page size in read function. If the limit is not defined, it will look for page_size in Table Display annotation, using the current context. Otherwise, it will use "1" as a default page size. Related Issue: #177 These changes are not relevant anymore. We decided to not change the default value explicitly, and just expose the default page size as reference.display.default_page_size .
gharchive/pull-request
2016-09-01T22:08:45
2025-04-01T06:44:31.245274
{ "authors": [ "RFSH" ], "repo": "informatics-isi-edu/ermrestjs", "url": "https://github.com/informatics-isi-edu/ermrestjs/pull/180", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
99971242
--mount-root option broken in v >= 0.2.45 The --mount-root option no longer works for me in version 0.2.45 and up. If I downgrade to 0.2.44 then the following command runs fine: $ forklift 273695c4789e sshd --mount-root ~/docker/fs/ mount: /dev/mapper/docker-252:0-4468428-0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf already mounted or /var/lib/docker/devicemapper/mnt/0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf/ busy Traceback (most recent call last): File "/usr/local/bin/forklift", line 9, in <module> load_entry_point('docker-forklift==0.2.45', 'console_scripts', 'forklift')() File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 338, in main return Forklift(sys.argv).main() File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 319, in main return driver.run(*command) File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 240, in run return self.run_sshd() File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 446, in run_sshd self.mount_root(container) File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 365, in mount_root container_path]) File "/usr/lib/python3.4/subprocess.py", line 557, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['sudo', 'mount', '-o', 'context=""', '/dev/mapper/docker-252:0-4468428-0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf', '/var/lib/docker/devicemapper/mnt/0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf/']' returned non-zero exit status 32 Ubuntu 14.04.3 LTS Docker version 1.7.1, build 786b29d Forklift Version: 0.2.45 Hmm. I'm guessing you don't have SELinux and thus we need to omit the 'context' parameter. I do not SELinux. I gave this a quick shot locally by removing the context argument but I'm still getting the error: $ forklift 3985400cef6c sshd --mount-root ~/docker/fs/ mount: /dev/mapper/docker-252:0-4468428-56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e already mounted or /var/lib/docker/devicemapper/mnt/56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e/ busy Traceback (most recent call last): File "/usr/local/bin/forklift", line 9, in <module> load_entry_point('docker-forklift==0.2.47', 'console_scripts', 'forklift')() File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 338, in main return Forklift(sys.argv).main() File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 319, in main return driver.run(*command) File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 240, in run return self.run_sshd() File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 450, in run_sshd self.mount_root(container) File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 369, in mount_root subprocess.check_call(args) File "/usr/lib/python3.4/subprocess.py", line 557, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['sudo', 'mount', '/dev/mapper/docker-252:0-4468428-56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e', '/var/lib/docker/devicemapper/mnt/56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e/']' returned non-zero exit status 32
gharchive/issue
2015-08-10T04:52:42
2025-04-01T06:44:31.263522
{ "authors": [ "danni", "jscn" ], "repo": "infoxchange/docker-forklift", "url": "https://github.com/infoxchange/docker-forklift/issues/97", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2210316478
feat(azure): add support for azurerm_app_configuration Adds support for Azure App Configuration, a centralized service for managing application settings and feature flags, enabling secure and dynamic configuration updates without redeployment. This service is supported by Infracost through the azurerm_app_configuration Terraform resource. Which contains the following cost components: Instance & replica costs - there is a cost per day for each store and replica that is enabled in Azure App Configuration Additional Requests - Users are billed for every additional 10k requests (outside of the 200k allowance included with the standard instance/replica) to the service Additionally, Azure App Configuration has both the Free and Standard skus. The former is free to run, so we simply return a free resource if a user has specified sku = "free" in the azurerm_app_configuration. I'm not sure why the TestNetworkDdosProtectionPlan are failing now with the usage string, seems odd as these are in the usage example file 🤷 @tim775 digging around here https://azure.microsoft.com/en-gb/pricing/calculator/ I realised that the 200k number is just the number of requests included in the base cost of the instance/replica. It is not a daily number but a monthly one. @tim775 digging around here https://azure.microsoft.com/en-gb/pricing/calculator/ I realised that the 200k number is just the number of requests included in the base cost of the instance/replica. It is not a daily number but a monthly one. Are you sure? I know it's not definitive, but this page says: Cost: Standard tier stores have a daily usage charge. The first 200,000 requests each day are included in the daily charge. There's also an overage charge for requests past the daily allocation. There's no cost to use a free tier store. And the the $1.20/day per replica matches up with the 200,000 @ $0.06 per 10k. So it seems like the 200,000 is kind of like a "minimum" requests per day. Are you sure? I know it's not definitive, but this page Hmm i am not sure.. but that page looks semi-definitive, i'll revert the last commit and merge with the replica/day messaging
gharchive/pull-request
2024-03-27T09:42:16
2025-04-01T06:44:31.273765
{ "authors": [ "hugorut", "tim775" ], "repo": "infracost/infracost", "url": "https://github.com/infracost/infracost/pull/2970", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
513693942
Update joda-time to 2.10.5 Updates joda-time:joda-time from 2.9.9 to 2.10.5. Release Notes/Changelog I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "joda-time", artifactId = "joda-time" } ] labels: semver-minor Codecov Report Merging #316 into master will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #316 +/- ## ====================================== Coverage 80.5% 80.5% ====================================== Files 128 128 Lines 3334 3334 Branches 136 136 ====================================== Hits 2684 2684 Misses 650 650 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update f5d3309...0d752af. Read the comment docs.
gharchive/pull-request
2019-10-29T04:58:28
2025-04-01T06:44:31.281236
{ "authors": [ "codecov-io", "scala-steward" ], "repo": "ing-bank/baker", "url": "https://github.com/ing-bank/baker/pull/316", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1714554735
Circular dependencies It just so happened to me that CPAN.pm died when it encountered these circular dependencies: Recursive dependency detected: Lingy (have: N/A; want: 0.001004) => INGY/Lingy-0.1.4.tar.gz => YAMLTest (have: N/A; requires: 0) => INGY/YAMLTest-0.0.4.tar.gz => Test::More::YAMLScript (have: N/A; requires: v0.0.4) => INGY/Test-More-YAMLScript-0.1.0.tar.gz => YAMLScript (have: N/A; requires: v0.1.0) => INGY/YAMLScript-0.1.0.tar.gz => Lingy (have: N/A; requires: v0.1.4). Cannot resolve. INGY/YAMLScript-0.1.0.tar.gz [depend] -- NOT OK Thanks for reporting this. It seems to be resolved in 0.1.11. Closing for now.
gharchive/issue
2023-05-17T19:45:37
2025-04-01T06:44:31.292920
{ "authors": [ "andk", "ingydotnet" ], "repo": "ingydotnet/lingy", "url": "https://github.com/ingydotnet/lingy/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2438817095
Fix y2k (Y2K-1162) Updated with today's date Well... yeah.
gharchive/pull-request
2024-07-30T23:10:25
2025-04-01T06:44:31.340938
{ "authors": [ "billlumbergh-initech", "petergibbons2" ], "repo": "initech-eng/y2k", "url": "https://github.com/initech-eng/y2k/pull/973", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1242833513
Support for Ctrl-[ I understand it's not feasible to have Inkdrop parse my .vimrc, but supporting the standard Ctrl-[ (does the same thing as Esc) would be awesome. Thanks! You can customize keymaps as you like. Please read the documentation: https://docs.inkdrop.app/manual/customizing-keybindings https://github.com/inkdropapp/inkdrop-vim#key-customizations
gharchive/issue
2022-05-20T08:47:39
2025-04-01T06:44:31.366080
{ "authors": [ "craftzdog", "jcredmond" ], "repo": "inkdropapp/inkdrop-vim", "url": "https://github.com/inkdropapp/inkdrop-vim/issues/38", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
262098169
React 16 support Update peer dependencies to include React 16. Update Enzyme to version that supports React 16. Included in 0.7.0.
gharchive/pull-request
2017-10-02T14:10:52
2025-04-01T06:44:31.380994
{ "authors": [ "bebraw", "sapegin" ], "repo": "inlinestyle/remark-react-lowlight", "url": "https://github.com/inlinestyle/remark-react-lowlight/pull/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1655890201
Expose core logging setup through the stable api As discussed here: https://inmanta.slack.com/archives/CKRF0C8R3/p1680678625072959 It would be nice to expose (at least some part) of the logging setup done by core through the stable api, to allow other python tools developed internally to reuse it. Simplifying their development and making them more consistent with core. The method which would be handy are: inmanta.app._get_default_stream_handler() inmanta.app._get_log_formatter_for_stream_handler(timed=True) inmanta.app._convert_to_log_level(verbose) Special attention should be paid to naming and docstrings. https://code.inmanta.com/solutions/tools/yang-module-generator/-/blob/master/src/yang_module_generator/cli.py#L77-97 https://code.inmanta.com/solutions/modules/restbase/-/blob/master/inmanta_plugins/restbase/generator/app.py#L24-41
gharchive/issue
2023-04-05T15:51:34
2025-04-01T06:44:31.389101
{ "authors": [ "edvgui", "sanderr" ], "repo": "inmanta/inmanta-core", "url": "https://github.com/inmanta/inmanta-core/issues/5815", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
952818096
Add resource history endpoint (#3126) Pull request opened by the merge tool on behalf of #3126 Pull request rejected by merge tool. The tests for this branch did not succeed. Merged into branches iso4 in eb6cd4c028892866396d6f2f2592df687c9f530f
gharchive/pull-request
2021-07-26T11:49:05
2025-04-01T06:44:31.390268
{ "authors": [ "inmantaci" ], "repo": "inmanta/inmanta-core", "url": "https://github.com/inmanta/inmanta-core/pull/3138", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2265952387
Fix core build failures Description Revert the tox.ini changes until the irt changes are in again https://github.com/inmanta/irt/pull/2022 Self Check: Strike through any lines that are not applicable (~~line~~) then check the box [ ] Attached issue to pull request [ ] Changelog entry [ ] Type annotations are present [ ] Code is clear and sufficiently documented [ ] No (preventable) type errors (check using make mypy or make mypy-diff) [ ] Sufficient test cases (reproduces the bug/tests the requested feature) [ ] Correct, in line with design [ ] End user documentation is included or an issue is created for end-user documentation (add ref to issue here: ) [ ] If this PR fixes a race condition in the test suite, also push the fix to the relevant stable branche(s) (see test-fixes for more info) Processing this pull request Merged into branches master in 280804f1e574ffc258957a847f912286ee45f4c9
gharchive/pull-request
2024-04-26T14:50:11
2025-04-01T06:44:31.394377
{ "authors": [ "Hugo-Inmanta", "inmantaci" ], "repo": "inmanta/inmanta-core", "url": "https://github.com/inmanta/inmanta-core/pull/7569", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2046960648
Bump inmanta-dev-dependencies[async,extension,pytest] from 2.96.0 to 2.97.0 (#493) Pull request opened by the merge tool on behalf of #493 Merged in f5ab68bd7bd1ac36ec4fc98419d8199967ec6767
gharchive/pull-request
2023-12-18T15:52:04
2025-04-01T06:44:31.395546
{ "authors": [ "inmantaci" ], "repo": "inmanta/inmanta-ui", "url": "https://github.com/inmanta/inmanta-ui/pull/495", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1075492134
Build(deps-dev): Bump @storybook/addon-links from 6.4.8 to 6.4.9 (#2138) Pull request opened by the merge tool on behalf of #2138 Merged in c75b4352c15be70792932699540e9007c9246317
gharchive/pull-request
2021-12-09T12:00:14
2025-04-01T06:44:31.396563
{ "authors": [ "inmantaci" ], "repo": "inmanta/web-console", "url": "https://github.com/inmanta/web-console/pull/2139", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1150296670
Issue/2632 url state filter hook (#2859) Pull request opened by the merge tool on behalf of #2859 Merged in 77fb58c3466876d5035a3bcafc623caf39f7165a
gharchive/pull-request
2022-02-25T10:39:08
2025-04-01T06:44:31.397347
{ "authors": [ "inmantaci" ], "repo": "inmanta/web-console", "url": "https://github.com/inmanta/web-console/pull/2885", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1441948433
remove link to dashboard Description The link to the dashboard has been removed from the sidebar. closes #4108 Self Check: Strike through any lines that are not applicable (~~line~~) then check the box [x] Attached issue to pull request [x] Changelog entry [x] Code is clear and sufficiently documented [x] Sufficient test cases (reproduces the bug/tests the requested feature) [x] Correct, in line with design [ ] End user documentation is included or an issue is created for end-user documentation (add ref to issue here: ) Reviewer Checklist: [ ] Sufficient test cases (reproduces the bug/tests the requested feature) [ ] Code is clear and sufficiently documented [ ] Correct, in line with design Processing this pull request
gharchive/pull-request
2022-11-09T11:47:14
2025-04-01T06:44:31.401284
{ "authors": [ "LauraStordeur", "inmantaci" ], "repo": "inmanta/web-console", "url": "https://github.com/inmanta/web-console/pull/4303", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2178839272
One Tap for Android will be deprecated soon Caution: One Tap for Android will be deprecated soon. To ensure the continued security and usability of your app, migrate to Credential Manager today. Credential Manager supports passkey, password, and federated identity authentication (such as Sign-in with Google), stronger security, and a more consistent user experience. This is a warning from the developer platform. Android's one-click login will be deprecated soon. Unfortunately, I have not found a better library than this. I hope the author can update it. Thank you. Hi @naonao-a , thank you for the report. I am the author. You are right. The code file GoogleOneTapAuth.java uses classes in the com.google.android.gms.auth.api.signin namespace as GoogleSignIn, GoogleSignInAccount and GoogleSignInClient which are depreciated. However the code also already uses Sign in with Google. You can see in the GoogleOneTapAuth.java file that the classes Identity and SignInClient from the com.google.android.gms.auth.api.identity namespace are used to auto-sign-in the user. Sign-in using a button is using the depreciated way and needs to be updated as described in the link you mentioned. I will try to release an updated version in the next 7 days. OH!! that's very kind of you .Thank you. @naonao-a I also am interested in this update. @naonao-a, @jamesgroat, the fix is now merged and there is a new npm package with version 6.0.0. The new package uses capacitor version 6. The android code was migrated from Java to Kotlin. Further the return type of tryAutoOrOneTapSignIn, tryOneTapSignIn and tryAutoSignIn was simplified to return just one promise instead of a promise containing promises. The documentation in the README is adjusted. @naonao-a, @jamesgroat, the fix is now merged and there is a new npm package with version 6.0.0. The new package uses capacitor version 6. The android code was migrated from Java to Kotlin. Further the return type of tryAutoOrOneTapSignIn, tryOneTapSignIn and tryAutoSignIn was simplified to return just one promise instead of a promise containing promises. The documentation in the README is adjusted. Thank you so much.
gharchive/issue
2024-03-11T11:00:39
2025-04-01T06:44:31.447865
{ "authors": [ "inouiw", "jamesgroat", "naonao-a" ], "repo": "inouiw/capacitor-native-google-one-tap-signin", "url": "https://github.com/inouiw/capacitor-native-google-one-tap-signin/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1598776333
Integrate ruff Check whether scraping flake and flakeheaven for ruff would not be the best idea to proceed. Difficulty will emerge when choosing rules as most of that has to be hand verified. Most of the current checks are implemented (and there are some additional ones which seem to fit well), namely: pyupgrade (UP) flake8-2020 (YTT) flake8-bandit (S) flake8-blind-except (BLE) flake8-boolean-trap (FBT) flake8-comprehensions (C4) flake8-datetimez (DTZ) flake8-debugger (T10) flake8-import-conventions (ICN) flake8-no-pep420 (INP) flake8-print (T20) flake8-raise (RSE) flake8-return (RET) flake8-self (SLF) flake8-type-checking (TCH) flake8-unused-arguments (ARG) eradicate (ERA) Currently missing for ruff which we would like to support: flake8-noqa flake8-annotations-coverage flake8-annotations-complexity (open issue here) flake8-cognitive-complexity is better supported via settings (see here)
gharchive/issue
2023-02-24T14:50:24
2025-04-01T06:44:31.456516
{ "authors": [ "szymonmaszke" ], "repo": "inovintell/py-template", "url": "https://github.com/inovintell/py-template/issues/142", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1855748229
Feat: OpenAPI version in URL path Context API-breaking changes will be a challenge to manage without introducing a version scheme in the URL. Since we're building the client in the same scope as the server, it's possible to abstract the versioning concerns within. Proposed Solution Include the OpenAPI version within the URL scheme Embed the expected version within the client HTTP requests by reading the version from a file generated at build time The tests could use refactoring, as clearly there's some repetition evidenced by the number of changes this PR contains, but I've left that out of scope. Important Changes Introduced While completing this task I found the service level OpenAPI docs were missing the health check paths. I've fixed this in a preliminary commit. Pushed deployment fixes, now it should be good and ready for merge. I want to test this out before doing that, @blaggacao could you help us with that? could you help us with that [assert the changed to the deployment manifests]? The suggestive next steps are: wait for an update to the diff from CI carefully audit the diff: in a CD scenario, that's what a final review on this PR signs off on, as well merge and wait for deployment (side note: I've reasons to believe that this step turned out not working as intended, I'll need to revise this. I have this backlogged as LW-8219) ... iterate if the changes turned out not working as intended. For an even shorter feedback loop to polish the deployment manifests, I can think of three options: arion / docker-compose playground kind playground some of the existing operational playgrounds We may need to find and discuss our preferred option, still. I can not consider this 'testing', also alternative options arent satisfactory since we are dealing with k8s and AWS integration testing
gharchive/pull-request
2023-08-17T21:43:58
2025-04-01T06:44:31.463740
{ "authors": [ "blaggacao", "gytis-ivaskevicius", "rhyslbw" ], "repo": "input-output-hk/cardano-js-sdk", "url": "https://github.com/input-output-hk/cardano-js-sdk/pull/872", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
679608883
Optimise DeserializeShort.substring In my microbenchmark, this brings down the cost of deserialising a Shelley address in viewCompactTxOut from ~360ns to ~60ns. The cloneByteArray function was introduced in version 0.7.1.0 of the primitive package, so update the lower bound accordingly. I believe the old primitive < 0.7 constraint in cabal.project was due to older versions of the cborg package having >=0.5 && <0.7.1.0 as bounds for primitive, but cborg-0.2.4.0 bumped the upper bound to 0.8. This will likely require similar changes to downstream repos. Using cloneByteArray is exactly what was needed. This should speed up things considerably. I cannot speak to the version numbers the projects. Thanks for doing this.
gharchive/pull-request
2020-08-15T16:58:40
2025-04-01T06:44:31.466643
{ "authors": [ "TimSheard", "mrBliss" ], "repo": "input-output-hk/cardano-ledger-specs", "url": "https://github.com/input-output-hk/cardano-ledger-specs/pull/1786", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1070413899
[BUG] - Plutus script cannot unlock funds with no attached datum Internal/External External otherwise. Area Plutus Related to Plutus Scripts (Alonzo). Summary A delegator sent some Astor tokens to our Plutus contract directly in Tx 64718…02513. The normal/documented procedure is to send these tokens to a proxy service, which then constructs/submits the non-trivial Plutus Tx AFAIK, these 80 ADA are now lost, because Plutus cannot release funds that have no datum hash associated with the given UTxO [03-12-2021 10:41:43] INFO Query UTxO for Script 64718d583eefea766dda6787a844eee52d9c3e4301c39fd6eeef3ddb73402513#0 value {lovelace 1444443 3f997b68b1f491c7c2f10af4e2bf9566c5d25bd61df0343065d4fe1c.Astor302 80} datum TxOutDatumNone ba35d7e12f80988f6f2dd36bdb7ced58501711c5b84f8e0968d7d1b2e0be686d#1 value {lovelace 168000000 3f997b68b1f491c7c2f10af4e2bf9566c5d25bd61df0343065d4fe1c.Astor296 804} datum d8a95d4a0b1cd3185fc571eb4ad1c293cdee44fbdf45400616058662b817b827 312d07fbe8290cc541feaed2e4b030237092a55fd570c2eecaa1969e281e6e96#1 value {lovelace 901700000 3f997b68b1f491c7c2f10af4e2bf9566c5d25bd61df0343065d4fe1c.Astor302 716} datum e31b52bb6cf0017e24cfc13ede621e913565e68b364254c009335b470cfff870 I would assume that this is a common enough user error, which may cause lots of frustration. Can we please fix this, for example by providing a known datum value that maps to TxOutDatumNone - (possibly empty byte string)? Closing this. If this is still relevant please reopen.
gharchive/issue
2021-12-03T09:57:11
2025-04-01T06:44:31.470169
{ "authors": [ "Jimbo4350", "tdiesler" ], "repo": "input-output-hk/cardano-node", "url": "https://github.com/input-output-hk/cardano-node/issues/3414", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1293445398
[BUG] - VASIL cardano-cli query leadership-schedule Error (Next Epoch) External Area Other: cardano-node/cardano-cli Summary Error happens when the leadership schedule command in Vasil Testnet for the next epoch is runned on a block producer. Steps to reproduce On a block producing node in testnet running cardano-cli 1.35.0 run the cardano-cli query leadership-schedule --testnet-magic 1097911063 ... --next Error appears after few seconds Expected behavior Leadership schedule for block production for the next Epoch System info: OS Name: Ubuntu OS Version 20.04.4 LTS Node version: cardano-node 1.35.0 - linux-x86_64 - ghc-8.10 git rev 9f1d7dc163ee66410d912e48509d6a2300cfa68a CLI version: cardano-cli 1.35.0 - linux-x86_64 - ghc-8.10 git rev 9f1d7dc163ee66410d912e48509d6a2300cfa68a Error log : Error: Error while calculating the slot range: PastHorizon {...} 75% of epoch wasn't completed..
gharchive/issue
2022-07-04T17:49:03
2025-04-01T06:44:31.475451
{ "authors": [ "Techs2Help" ], "repo": "input-output-hk/cardano-node", "url": "https://github.com/input-output-hk/cardano-node/issues/4131", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
464533939
LiveViewState. TUI draw simplification. Node's data. LiveViewState. TUI draw simplification. Node's data. bors r+
gharchive/pull-request
2019-07-05T08:50:18
2025-04-01T06:44:31.476425
{ "authors": [ "CodiePP", "denisshevchenko" ], "repo": "input-output-hk/cardano-node", "url": "https://github.com/input-output-hk/cardano-node/pull/40", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1341826584
cardano-node: sync percent fixed. Closes #4345 bors r+
gharchive/pull-request
2022-08-17T14:09:33
2025-04-01T06:44:31.477186
{ "authors": [ "denisshevchenko" ], "repo": "input-output-hk/cardano-node", "url": "https://github.com/input-output-hk/cardano-node/pull/4350", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
232115282
[CSL-154] Improve storing of global data in GtLocalData @gromak Currently we store all global data there and fully update it when main blockchain is updated. However, we don't need to store all this data. For example, we don't need to store openings, we only need to store keys after 3k-th slot. We don't need to store full commitments after first k slots. And so on. Another thing is how we update it. Current approach is to copy global payload into GtLocalData every time main blockchain is updated. I think it's not bad, but maybe there are bad options. @iperesadin We don't need to store full commitments after first k slots. But we use commitments in checkShare (Pos.Ssc.GodTossing.Functions), in processShares (Pos.Ssc.GodTossing.LocalData.LocalData) @gromak @iperesadin yes, but we don't need commitments from nodes who sent us openings.
gharchive/issue
2017-05-30T00:13:34
2025-04-01T06:44:31.480123
{ "authors": [ "jagajaga" ], "repo": "input-output-hk/cardano-sl", "url": "https://github.com/input-output-hk/cardano-sl/issues/532", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
636570405
Tx fee estimate edge case when Tx value approach wallet value Context When calculating a transaction fee, if the amount to pay is set to the same value as the wallet holds (in a single address in this test case) or close to it, the TX fee is about as large as the wallets value. This likely has no functional impact as it looks like these TXs should fail anyway due to insufficient fee ADA. But in case it impacts something somewhere else, just FYI. Network discrimination for the send to address is not tested for when calculating fees. This might be by design and not an issue. Just FYI Information - Version 2020.6.5 (git revision: 5290acf902da32e1e541b5dc227629c9a300f1be) Platform Linux Installation Nix build Steps to Reproduce Check the wallet balance $ cardano-wallet-shelley wallet get $WALLETID ... "total": { "quantity": 99627083, "unit": "lovelace" }, "available": { "quantity": 99627083, "unit": "lovelace" } ... Estimate fees by CLI or API: # Set the payment about ~0.13 ADA lower than the wallet value; get a large estimated fee # $ cardano-wallet-shelley transaction fees $WALLETID --payment 99500000@$SEND_TO Ok. { "estimated_min": { "quantity": 99665281, "unit": "lovelace" }, "estimated_max": { "quantity": 99665281, "unit": "lovelace" } } # Set the payment about ~1.13 ADA lower than the wallet value; get an accurate fee # $ cardano-wallet-shelley transaction fees $WALLETID --payment 98500000@$SEND_TO Ok. { "estimated_min": { "quantity": 168141, "unit": "lovelace" }, "estimated_max": { "quantity": 168141, "unit": "lovelace" } } Perhaps not an issue, the network discriminator of the send to address doesn't appear to be checked during fee estimates. Repeat one of the above fee estimates, this time flipping the second digit of the address from a 1 to a 0 or visa-versa. The command completes the same way, successfully, despite one of those addresses not being valid on the network. $ cardano-wallet-shelley transaction fees $WALLETID \ --payment 99500000@$SEND_TO_WITH_MODIFIED_SECOND_DIGIT Ok. { "estimated_min": { "quantity": 99665281, "unit": "lovelace" }, "estimated_max": { "quantity": 99665281, "unit": "lovelace" } } Indeed, I am able to reproduce it for both: shelley and byron wallet. The boundary value (on Shelley testnet) seems to be 165281 lovelace. I.e. Where value is the amount for which you want to estimate fee: If (value <= wallet_balance - 165281) then :+1: if (value > wallet_balance - 165281) then :-1: For instance: wallet balance = 5 000 000 Estimate fee for 5000000−165281 = 4834719 result: 165281 lovelace (estimated_min and estimated_max) Estimate fee for 5000000−165280 = 4834720 result: 5000001 lovelace (estimated_min and estimated_max) lgtm
gharchive/issue
2020-06-10T21:46:43
2025-04-01T06:44:31.488164
{ "authors": [ "johnalotoski", "piotr-iohk" ], "repo": "input-output-hk/cardano-wallet", "url": "https://github.com/input-output-hk/cardano-wallet/issues/1740", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
722459598
Add delistPools to Pool DBLayer This is the first step for garbage collecting stake pools based on SMASH delisting. X-JIRA-Ticket: ADP-478 Questions / Considerations After looking at the logic, I believe we don't have to adjust any other functions of the DBLayer such as listRegisteredPools or readPoolProduction. These can still consider all pools. Only the API layer will consider delisted pools and adjust ApiStakePool accordingly. There's still the open question whether we want to really delete GCed pools via removePools at some point. TODO Feature [x] Add delisted column to metadata table and populate Pool DBLayer with functions [x] Create new internal_state table and have GC thread store last sync time [x] Add garbage collection thread querying the SMASH server for delisted pools and calling delistPools from DBLayer [x] have GC thread store last sync time [x] return last GC sync time [x] Add POST '{ "maintenance_action": "gc_stake_pools" }' /stake-pools endpoint QA [x] database tests [x] check json roundtripping works [ ] Add integration tests? We don't have SMASH server integration tests yet. bors try bors try bors try bors try bors try Add integration tests? We don't have SMASH server integration tests yet. Perhaps we could have a mock SMASH server in the integration tests? Maybe it would be enough as it just mock the situation that there is one stake pool blacklisted. This way we could make a system level test. Currently in our integration suite there are 4 pools being registered. One of them is retired at the very beginning (before the tests start). Two out of remaining three are a scheduled to be retired in the far future (and they are like that for the span of the suite). We could extend this set up to have yet another pool, but this one would served as "blacklisted" in the mock SMASH server and we would just make sure that this pool is not visible when we have metadata fetching strategy pointing to a mock SMASH server. We used to have a mock server for jormungandr metadata registry, that was serving a zip file with metadata https://github.com/input-output-hk/cardano-wallet/blob/master/lib/jormungandr/test/integration/Main.hs#L230-L238... Maybe similar concept could be used. I've tested delisting with the last commit manually with a mock smash server. It works. bors try In particular, I think we should define (with properties) how we expect this flag to behave in the presence of rollback. It looks as though rolling back won't cancel out the de-listing of a pool. If that is the desired behaviour, then it would be good to encode this as a property. That sounds like an integration test. For integration test we want a proper SMASH process imo. Everything else is a half-baked unit test in disguise. I've done quite some work to try to integrate with it, but ultimately this boils down to: https://github.com/input-output-hk/smash/issues/97 https://github.com/input-output-hk/smash/pull/99 https://github.com/input-output-hk/smash/pull/102 https://github.com/input-output-hk/cardano-db-sync/pull/372 The cardano-db-sync maintainer seems to have the opinion that we shouldn't use cardano-db-sync, which I find a bit confusing. So there are still a couple of options with sqlite and in-memory database. That'll still take some time to sort out. But what we can do is write some basic testing of delistPools, such that it actually sets the flag. I have some reservations about the naming of the poolFlag field, the PoolFlag type, and the poolRegistrationFlag DB field. I don't have a strong opinion about the naming, but I have a strong opinion about changing naming back and forth in the process of a PR, because it actually takes more time than it seems (certainly 30 minutes just to make it compile and pushable, then possible another iteration, because I missed something and CI fails). The naming was suggested in the review comments and there was time to discuss it. I think both forms are fine, so I'll keep the current one. @jonathanknowles wrote: In particular, I think we should define (with properties) how we expect this flag to behave in the presence of rollback. It looks as though rolling back won't cancel out the de-listing of a pool. If that is the desired behaviour, then it would be good to encode this as a property. @hasufell wrote: That sounds like an integration test. By rollback, I specifically mean the rollbackTo DB operation, rather than the entire rollback machinery (which is large). We currently have properties covering all of the pool DB operations, and I think it would be desirable to also test delistPools in this way. For example: the basic behaviour of delistPools; the interaction between delistPools and other key operations such as rollbackTo. Even if these end up being very simple properties, I think they are still worth having. Advantages: properties in Cardano.Pool.DB.Properties are tested against both the SQLite implementation and the Model implementation. In future, we test both of these implementations together using QuickCheck state machine, so we really do want to make sure these remain consistent. properties can often generate many more edge cases than we're likely to think of by hand. properties provide a statement of how these (essentially pure) functions are intended to work. @jonathanknowles makes sense, I'll get on it @hasufell wrote: I don't have a strong opinion about the naming, but I have a strong opinion about changing naming back and forth in the process of a PR, because it actually takes more time than it seems I completely sympathize. It's not easy to gain consensus in a PR review, especially when people are in multiple time-zones. Though I also think the time we spend on a PR has to be weighed up against the time we spend on future work. Some names (such as DB columns) once merged to master are more awkward to change (if we need to) than internal variable names, as they require migrations, which we wish to avoid if at all possible. In the case of poolRegistrationFlag, which really is very ambiguous to an uninitiated reader (IMO), I still think this would be worth changing to something more specific. (See proposal https://github.com/input-output-hk/cardano-wallet/pull/2249#pullrequestreview-516440993) Just my 2 ada. :+1: In the case of poolRegistrationFlag, which really is very ambiguous to an uninitiated reader (IMO), I still think this would be worth changing to something more specific. The current naming was suggested by @KtorZ. Maybe you two can sort that out :smile: Otherwise I'll change it again tomorrow to something else. the basic behaviour of delistPools; It just sets a flag in pool_registrations. the interaction between delistPools and other key operations such as rollbackTo. There really is none. The interaction doesn't happen on DBLayer level. It's higher up in the program stack, so only an integration test can touch that. The only property we really have is checking whether the Delisted flag shows up. Hi Jonathan, as you might know, @KtorZ is my co-pilot and so certain decisions have been discussed with him or have been proposed by him. To that end, the existing solution goes to some lengths to ensure that all rows in the pool_registration table have the same delisted status for a particular pool. But this seems unnecessary. At the end of the day, the set of delisted pools is just a set, which should be possible to represent with a single-column SQL table. So I think it would make things a lot simpler to just store the set of delisted pools in its own table, rather than storing each pool's delisted status multiple times, once for each registration record: We discussed this already and the proposed solution was to put the flag into the pool registration table. And I'd like to stick with that now, otherwise this PR will become a chaotic back and forth of ideas. Just an update: @hasufell and I had an out-of-band discussion about this PR, and we came to the conclusion that it would make sense to adopt the DB schema design proposed in #2277, which stores the delisted status just once per pool. bors try bors try bors r+ bors r+
gharchive/pull-request
2020-10-15T15:52:14
2025-04-01T06:44:31.511887
{ "authors": [ "KtorZ", "hasufell", "jonathanknowles", "piotr-iohk" ], "repo": "input-output-hk/cardano-wallet", "url": "https://github.com/input-output-hk/cardano-wallet/pull/2249", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
833543372
Nightly restore bench: Make buildkite red if it fails Issue Number ADP-804 Overview If the nightly restore bench fails to produce a result, then make sure Buildkite has a red status. bors r+ bors r+ bors r+
gharchive/pull-request
2021-03-17T08:52:04
2025-04-01T06:44:31.514545
{ "authors": [ "Anviking", "piotr-iohk", "rvl" ], "repo": "input-output-hk/cardano-wallet", "url": "https://github.com/input-output-hk/cardano-wallet/pull/2566", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2604628705
feat(rust/catalyst-voting): Vote protocol benchmarks Description Added benchmarks for all of the steps of the catalyst vote protocol. Refactored some stuff. @minikin not sure that its actually what we want here, because I will post a results made on my machine, so they will be not the same if some other will run it. Maybe we need to define a general strategy how we want to run our benchmarks on some regular basis, maybe something the same as we have codecov reports. Seems we need to raise this topic on our tech discussions and decide what to do. @Mr-Leshiy, can you please create a task about publishing benchmark results? Let's discuss this topic in our next technical discussion.
gharchive/pull-request
2024-10-22T08:08:06
2025-04-01T06:44:31.516774
{ "authors": [ "Mr-Leshiy", "minikin" ], "repo": "input-output-hk/catalyst-libs", "url": "https://github.com/input-output-hk/catalyst-libs/pull/65", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1705169054
CI bits Split up Cicero into independent jobs such that the logs of different platforms are no longer interleaved, easing log inspection We temporarily bumped the GHC on GHA to 9.2 in #46 as it started to randomly fail with weird out-of-memory errors on 8.10. We still have to support compiling for 8.10 for ~one month, so we are reverting that change here. Will cicero build with docs enabled? i.e. an error in the haddocks will be caught somewhere? Will cicero build with docs enabled? i.e. an error in the haddocks will be caught somewhere? Not yet, will change that :+1: Actually, adding -haddock no longer makes the GHC parser more strict since GHC 9.0 (see https://gitlab.haskell.org/ghc/ghc/-/issues/23379). We could enable -Winvalid-haddock, which currently generates lots of warnings (e.g. all Haddock-ish comments on non-toplevel bindings), but they are not blocking Haddock generation. I will try to enable Haddock again for GHA, maybe it works again :shrug:
gharchive/pull-request
2023-05-11T06:41:09
2025-04-01T06:44:31.539529
{ "authors": [ "amesgen", "jasagredo" ], "repo": "input-output-hk/ouroboros-consensus", "url": "https://github.com/input-output-hk/ouroboros-consensus/pull/76", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
924978453
SCP-2232: Chain index types Starting to separate the chain index from the PAB. See https://hackmd.io/ji8Hs_88TlqMOK0VeX9Row for the original specification. This only defines the types and query interface, implementation still TBD. Pre-submit checklist: Branch [ ] Commit sequence broadly makes sense [ ] Key commits have useful messages [ ] Relevant tickets are mentioned in commit messages [ ] Formatting, materialized Nix files, PNG optimization, etc. are updated PR [ ] Self-reviewed the diff [ ] Useful pull request description [ ] Reviewer requested Pre-merge checklist: [ ] Someone approved it [ ] Commits have useful messages [ ] Review clarifications made it into the code [ ] History is moderately tidy; or going to squash-merge @sjoerdvisscher I fixed the definitions and added a property test. This wasn't straightforward because I had to write the generators to produce the right data. But it seems to be good now. Some of these generators needed to be written anyway, right? I really like the code overall. 💯 @michaelpj this is just hanging on a VM test, could please give it a little nudge? Thank you!
gharchive/pull-request
2021-06-18T14:50:24
2025-04-01T06:44:31.545073
{ "authors": [ "j-mueller", "sjoerdvisscher" ], "repo": "input-output-hk/plutus", "url": "https://github.com/input-output-hk/plutus/pull/3399", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
673802370
export getFileWithAcl alongside getFile This PR exports getFileWithAcl alongside getFile, following the pattern with getSolidDataset and getSolidDatasetWithAcl. I didn't add a unit test because I don't see a pattern of testing exports. [ ] I've added a unit test to test for potential regressions of this bug. [ ] The changelog has been updated, if applicable. [ x] Commits in this PR are minimal and have descriptive commit messages. We're a bit caught up in some other work, but I'll try to look at this after the weekend - hope that's OK. no problem at all! I'm tracking this repository at https://github.com/itme/solid-client-js and just contributing back patches as I have them, so y'all aren't blocking me at all. thanks, btw, for this work, this library is an order of magnitude easier to use than anything else out there already. Yeah, WhiteSource is failing due to this PR coming in as an external contribution (we're aware of this problem, and are working to fix it). In the meantime, I'll create a separate PR to get this fix in - thanks again @travis ! yw, thanks!
gharchive/pull-request
2020-08-05T19:44:58
2025-04-01T06:44:31.564068
{ "authors": [ "Vinnl", "pmcb55", "travis" ], "repo": "inrupt/solid-client-js", "url": "https://github.com/inrupt/solid-client-js/pull/319", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
811588924
Chore/move datafactory to rdfjs Internal refactor - just moved all DataFactory imports to pull from rdfjs.js rather than from N3 directly. The Turtle parsing code still relies explicitly on N3, but that's fine (test needed to be updated to support quad-equality between RDF/JS implementations, but that's a test improvement). This change makes it really easy to flip our DataFactory implementation dependency now - and so I did try that out by flipping it from N3 to @rdfjs/data-model (that flip is not included in this PR). Only 4 tests failed, which is pretty good - with the reasons documented in rdfjs.ts. Those reasons need to be verified, but it would be nice to resolve them (not urgent though - we're still relying on the RDF spec compliance of N3, as we did before). Merged in #837
gharchive/pull-request
2021-02-19T00:46:43
2025-04-01T06:44:31.566178
{ "authors": [ "ajacksified", "pmcb55" ], "repo": "inrupt/solid-client-js", "url": "https://github.com/inrupt/solid-client-js/pull/840", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
691411583
Fix "KeyError: 'creator'" bug with private events Adds missing check on nested dictionary access @jcrowgey, can you take a look now? Updated to fix merge conflicts with master. @jcrowgey LMK if you want to review and sign off, otherwise I can take over for you and get it merged. It would take me some time to recover the context. I'm happy to hand it off to you. On August 22, 2024 9:47:25 PM PDT, David Barnett @.***> wrote: Updated to fix merge conflicts with master. @jcrowgey LMK if you want to review and sign off, otherwise I can take over for you and get it merged. -- Reply to this email directly or view it on GitHub: https://github.com/insanum/gcalcli/pull/562#issuecomment-2306284875 You are receiving this because you were mentioned. Message ID: @.***>
gharchive/pull-request
2020-09-02T20:50:46
2025-04-01T06:44:31.569595
{ "authors": [ "dbarnett", "jcrowgey", "shaicoleman" ], "repo": "insanum/gcalcli", "url": "https://github.com/insanum/gcalcli/pull/562", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
159568691
ported To Cuda: https://sourceforge.net/projects/cudatext/files/addons/plugins/plugin.JS_Sort_Imports.zip/download :+1:
gharchive/issue
2016-06-10T06:34:50
2025-04-01T06:44:31.577090
{ "authors": [ "Alexey-T", "insin" ], "repo": "insin/sublime-sort-javascript-imports", "url": "https://github.com/insin/sublime-sort-javascript-imports/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
370207197
Line break not working Hi, thanks for this great javascript plugin. I notice a small issue I'm having. I'd like to have the text on several lines including line breaks, and it seems they are not working/parsed. window.addEventListener("load", function(){ window.cookieconsent.initialise({ "palette": { "popup": { "background": "#000" }, "button": { "background": "#f1d600" } }, "content": { "message": "\nCe site internet utilise des cookies pour ameliorer votre experience.\n\nEn naviguant sur les page du site internet vous acceptez l'utilisation des cookies.", "dismiss": "OK, J'AI COMPRIS", "link": "En savoir plus" } })}); Hi Tristof, You can use HTML tags instead. For your case, you can use Thanks
gharchive/issue
2018-10-15T15:04:13
2025-04-01T06:44:31.579307
{ "authors": [ "Tristof", "dolav" ], "repo": "insites/cookieconsent", "url": "https://github.com/insites/cookieconsent/issues/473", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1635139325
Feature Request - Add option to no use debrid service in my country there is no need to use a debrid service, why i have to use one with this application? :/ sorry i'm going to wait for an actual solution, thank you for the answer, this project have future! Initial level changes for streaming torrents are in. It takes some time to start streaming. If it's lagging, pause for some time for the download to get ahead. (maybe it's my download speed being slow, idk) On settings there is torrenting section For nya you'll have to give a mirror that works for you. Change your download server here https://premiumize.me/speedtest i recommend not using automatic, cloudflare, or direct add those tend to not play well with players. Personally digital ocean Toronto is my best server Very Initial level changes for streaming torrents are in. It takes some time to start streaming. If it's lagging, pause for some time for the download to get ahead. (maybe it's my download speed being slow, idk) On settings there is torrenting section For nya you'll have to give a mirror that works for you. Change your download server here https://premiumize.me/speedtest i recommend not using automatic, cloudflare, or direct add those tend to not play well with players. Personally digital ocean Toronto is my best server Added option to use VLC media player instead of windows media player. VLC is better than WMP at streaming MKV files. So if you're using torrents, use Vlc media player, otherwise use the default. (Vlc package is still in pre-release) Playing with debrid service works perfectly now. If you don't have a debrid service, I'm downloading the torrent while streaming, so it might some time to start playing. I've added, download information to know if something is happening. Video will start playing after first section is downloaded. After this i didn't notice any buffering (unless you seek). Download will go to. \AppData\Local\Totoro\Torrents Contents of this folder will get cleared at startup. so, if you want to keep the downloaded files, you might want to move them somewhere else before starting.
gharchive/issue
2023-03-22T06:26:36
2025-04-01T06:44:31.586889
{ "authors": [ "DamianArielLerner", "insomniachi", "reddit-reaper" ], "repo": "insomniachi/Totoro", "url": "https://github.com/insomniachi/Totoro/issues/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1352153417
Feat: Launchpad integration test What? Added Multi-process JAX integration test (matches ACME style testing which run in lp_launch_type TEST_MULTI_THREADING) Why? Due to issues with Launchpad, current tests are single process How? Implement simple test that calls each process once. Issue we had previously with not being able to access environment loop run_episode() has been fixed by exposing the method in the wrapper. Extra Refactor jax tests folder (move mock_systems.py into tests/jax/systems folder to house all test systems, rename variables for consistent integration tests) Closes https://github.com/instadeepai/Mava/issues/702 Thank you @AsadJeewa! A general request/comment from my side. Can we please only name actual tests with names that start with test_ otherwise it seems that pytest runs unintended tests by accident. For example if I run pytest tests/jax/systems/ippo_system_test.py it says that 2 tests completed succesfully when there is only one due to the naming of the imported systems. I updated the name of the systems to not be called test_ to solve this. I also named the files with the test systems system_test_data.py to avoid a similar issue Thank you :)
gharchive/pull-request
2022-08-26T11:41:39
2025-04-01T06:44:31.726705
{ "authors": [ "AsadJeewa", "RuanJohn" ], "repo": "instadeepai/Mava", "url": "https://github.com/instadeepai/Mava/pull/703", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
812445246
How to share to facebook Hello, how to share to facebook de content published in instagram using instagram4j this is my code IGClient client = IGlogin(); client.actions() .timeline() .uploadPhoto(getFile(post.getContent()), post.getCaption()) .thenAccept(response -> {}) .join(); Sharing to other social media is not supported in this library currently.
gharchive/issue
2021-02-20T00:12:56
2025-04-01T06:44:31.730267
{ "authors": [ "diegocallealvarez", "jvogit" ], "repo": "instagram4j/instagram4j", "url": "https://github.com/instagram4j/instagram4j/issues/565", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
227971566
Unfollow less the white list I need to give unfollow in followers but I want to preserve about 100 followers how to do is on the whitelist? Is this programmed already? @zsvendo try to use get_user_followers to get all followers and truncate this list kind a arr = arr[:100] after that use this list as "white list". Hello, I have the whitelist of people not to unfollow, but when I execute Unfollow everyone, with (whitelist=True), it still unfollows whitelisted users. Any idea ? Thanks How can i active whitelist for ultimate_schedule unfollow_non_follower? def unfollow_non_followers(self): self.logger.info("Unfollowing non-followers") followings = set([item["pk"] for item in self.getTotalSelfFollowings()]) self.logger.info("You follow %d users." % len(followings)) followers = set([item["pk"] for item in self.getTotalSelfFollowers()]) self.logger.info("You are followed by %d users." % len(followers)) diff = followings - followers - whitelist('/friends.txt') self.logger.info("%d users don't follow you back." % len(diff)) self.unfollow_users(list(diff)) Would this work : diff = followings - followers - whitelist('/friends.txt') ? The whitelist.txt is not working anyone can have some idea: I run the 4.unfollow option and then the 2. unfollow everyone option. But it unfollow all does not respect the whitelist.txt Should be fixed in current version.
gharchive/issue
2017-05-11T12:17:04
2025-04-01T06:44:31.735541
{ "authors": [ "alcaporal", "alexsonnay", "maxkrivich", "ohld", "zsvendo" ], "repo": "instagrambot/instabot", "url": "https://github.com/instagrambot/instabot/issues/195", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1091616424
How to set last post timestamp via LatestStamps("stampsfile").set_last_post_timestamp("myaccount", datetime) I'm trying to write a Python bot that fetches the Instagram posts from an account and posts them on Twitter. I want the bot to post just the newest post, therefore I thought I should make use of the LatestStamps Class. I now have the problem that I don't fully understand how to get the post-timestamp of the post and how to set it in the stampsfile. What kind of value does the datetime parameter need? Nevermind... for post in profile.get_post(): LatestStamps("stampsfile").set_last_post_timestamp("profile", post.date) Nevermind... for post in profile.get_post(): LatestStamps("stampsfile").set_last_post_timestamp("profile", post.date)
gharchive/issue
2021-12-31T16:13:36
2025-04-01T06:44:31.738230
{ "authors": [ "LucEast" ], "repo": "instaloader/instaloader", "url": "https://github.com/instaloader/instaloader/issues/1367", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2330684637
Instaloader Redirected to login page. Use --login Issue I ran into the issue of Redirected to login page. Use --login It seems Instagram changed their API, Is there any new fix to this or is this a new issue? Was working fine until I pulled the same 3 profiles 10 times within 5 minutes, then I got a similar message An error occurred while processing profile 'some_profile': Redirected to login page. Use --login.
gharchive/issue
2024-06-03T09:50:10
2025-04-01T06:44:31.739665
{ "authors": [ "frozencap", "pdbarma" ], "repo": "instaloader/instaloader", "url": "https://github.com/instaloader/instaloader/issues/2275", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1851088439
Underlined featured Hi dev team, could you please support the underlined featured? Thank you Standard markdown syntax does not support underscores. If you want to using this, maybe you can using html tag or some style to support.
gharchive/issue
2023-08-15T08:35:33
2025-04-01T06:44:31.741979
{ "authors": [ "hoangbtmrk", "instance-oom" ], "repo": "instance-oom/ngx-markdown-editor", "url": "https://github.com/instance-oom/ngx-markdown-editor/issues/138", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2487688657
Feature Request: Stripe Connect but for Instant DB Feature request: Stripe Connect but for Instant DB BLUF We’d like to provision Instant DB accounts and apps on behalf of our users, similar to how Squarespace can provision Stripe accounts and payments on behalf of their users. Need GPT Engineer lets you build and deploy a complete frontend app using only prompts (powered by LLMs). For adding persistence we today support a Supabase integration. This requires the user to go into supabase, create an account, and configure the schema and then use the API keys to Connect to GPT Engineer. This is great for users that already use Supabase or who want a lot of control. But many users don’t care about what a persistence layer looks like as long as it gets the job done and would prefer minimal friction. We believe that Instant DB could unlock this low friction use case if there was a way for us to programmatically create Instant DB apps on behalf of our users. We don't want our users to have to go to instantdb.com and sign up, we'd want to sign up for them, ideally without them needing to do anything. Proposed solution A couple of API endpoints (not exactly how they should look but something to start the discussion): POST /users curl https://api.instantdb.com/v1/users \ -u "our_secret_api_key" \ -d email=test@example.com -d password=long_difficult_password Returns { id: string; email: string; } GET /users/<id> curl https://api.instantdb.com/v1/users/<id> \ -u "our_secret_api_key" \ -d email=test@example.com Returns { id: string; email: string; } POST /app curl https://api.instantdb.com/v1/app \ -u "our_secret_api_key" \ -d user_id="user_id" \ -d name="cool_app" Returns { name: string; id: string; public_api_key: string; // The user's client side public API key } GET /app/<id> curl https://api.instantdb.com/v1/app/<id> \ -u "our_secret_api_key" Returns { name: string; id: string; public_api_key: string; // The user's client side public API key } Concerns Validating users One possible concern here would be the validation of users and their emails. We naturally have our users already validated so either we establish trust that way or we device some other way of establishing trust. Ideally we'd like to avoid the user having to re-validate their email again for example. Giving users access to their Instant DB dashboards We naturally would like the user to have access to their Instant DB dashboards just as a normal Instant DB user. This is straightforward if they are using SSO since then that login can just be reused. But if the user is using GPT Engineer with email and password we will likely need to ask the user to add an Instant DB password when creating the user. Proposal #2: Enable claimable projects. The UUID of claimable projects is secret. A person that knows the UUID can claim it. Once it is claimed it is editable by both the creator of the project and the person who claimed it. Thanks for writing these up! It sounds like the most important functionality is: You should be able to spin up an app easily, and associate it with a user (i.e. an email address) You should be able to programmatically give that user access to the app (either by transferring ownership, or by adding that user as a collaborator, or allowing apps to be "claimable") Let me know if I'm understanding things correctly @vikeri @AntonOsika :) Correct! I think Viktor’s solution here is easiest to implement. You would just need to store: user record field: api secret it was created with (optional) project record field: api secret it was created with (optional) and verify api secret, on create project and edit project, if attempted by someone else than user api secrets are given out after appropriate trust+contract is signed with third party Let me know if I'm understanding things correctly Yep!
gharchive/issue
2024-08-26T20:13:15
2025-04-01T06:44:31.750971
{ "authors": [ "AntonOsika", "reichert621", "vikeri" ], "repo": "instantdb/instant", "url": "https://github.com/instantdb/instant/issues/45", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2347057380
Create interface for accessing the training code This PR turns this into a library. View the following PR for an example: https://github.com/instructlab/instructlab/pull/1329 The main idea is that we provide a function like: def train_torchrun(torchargs: TorchrunArguments, training_args: FullTrainingArguments): pass Each class TorchrunArguments and FullTrainingArguments provides training-specific arguments. And then from another library, you would simply provide these arguments to the train_torchrun function. For any other training method that we define, we could provide a similar interface depending on which arguments are needed. We separate the arguments here because TorchrunArguments are the ones passed to torchrun and then the full training arguments are the ones that we actually train with. It's not crucial that these are different, but it makes our lives a lot easier from a maintenance standpoint. https://github.com/instructlab/eval/pull/1 Note: Current requirements will overwrite existing nvidia pytorch installs. Need to ensure that if those exist, we are not installing our own torch. Thank you for the review @JamesKunstle, I've created an issue about your quantization comment here: https://github.com/instructlab/training/issues/29 just to be super clear: squash the first two commits and fix up the commit message create a follow-up issue for cleaning up requirements.txt to specify versions for all dependencies + making sure the versions are aligned with instructlab/instructlab if it's a shared dependency create a follow-up issue for cleaning up requirements.txt to specify versions for all dependencies + making sure the versions are aligned with instructlab/instructlab if it's a shared dependency We talked about this, so I know your intent was to create this issue, but I can't find it anywhere. I went ahead and filed this one: https://github.com/instructlab/training/issues/34
gharchive/pull-request
2024-06-11T18:38:49
2025-04-01T06:44:31.767620
{ "authors": [ "Maxusmusti", "RobotSail", "russellb" ], "repo": "instructlab/training", "url": "https://github.com/instructlab/training/pull/12", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
173308853
Update umadbro count=$(($count+1)) results in a loop that never ends. As $RANDOM gives a number between 0 - 32767 this could result in a quite small fake volume schweeeetSiOn 26 August 2016 at 16:18 int0x80 notifications@github.com wrote:Merged #1.—You are receiving this because you authored the thread.Reply to this email directly, view it on GitHub, or mute the thread. Thanks for the diff :) On Aug 26, 2016 11:55, "SecuritySense" notifications@github.com wrote: schweeeetSiOn 26 August 2016 at 16:18 int0x80 notifications@github.com wrote:Merged #1.—You are receiving this because you authored the thread.Reply to this email directly, view it on GitHub, or mute the thread. — You are receiving this because you modified the open/close state. Reply to this email directly, view it on GitHub https://github.com/int0x80/anti-forensics/pull/1#issuecomment-242790223, or mute the thread https://github.com/notifications/unsubscribe-auth/ABAwqHSVs8qOrAQ2k40g0REaeB8jalL0ks5qjxqYgaJpZM4Jtf24 .
gharchive/pull-request
2016-08-25T20:59:37
2025-04-01T06:44:31.774274
{ "authors": [ "SecuritySense", "int0x80" ], "repo": "int0x80/anti-forensics", "url": "https://github.com/int0x80/anti-forensics/pull/1", "license": "WTFPL", "license_type": "permissive", "license_source": "github-api" }
935161593
🛑 香港|CN2 is down In 91aa550, 香港|CN2 ($HKCN2_SITE_URL) was down: HTTP code: 0 Response time: 0 ms Resolved: 香港|CN2 is back up in a5ddd59.
gharchive/issue
2021-07-01T19:46:47
2025-04-01T06:44:31.776635
{ "authors": [ "Kevincible" ], "repo": "intSailers/uptime", "url": "https://github.com/intSailers/uptime/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1707929079
Support for psr/log ^2.0 || ^3.0 Would it be possible to get the requirement for psr/log updated to a more recent version? Yeah this is starting to conflict with packages... Looks like this will also be fixed in https://github.com/intacct/intacct-sdk-php/pull/187
gharchive/issue
2023-05-12T16:50:37
2025-04-01T06:44:31.778368
{ "authors": [ "rijnhard", "stephenmichael" ], "repo": "intacct/intacct-sdk-php", "url": "https://github.com/intacct/intacct-sdk-php/issues/189", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
523792023
"This app isn't verified" warning from google oauth When I sign in with token='browser', google shows me this I can eventually make it in, but it isn't very confidence-inspiring. I can't figure out how to verify my app. Should be using a different API key maybe? Interesting... This hasn't been reported for gcsfs, which has the same mechanism available, but I don't think it gets used that much. I suppose the link would tell us how to get the app verified, but there may bee a good chance that we don't actually meet google's guidelines... Would you be interested in following the link to verify this package as an app? I couldn't figure out how to do it. Also, today we got this same message with gcsfs proper. Tried starting the process for gcsfs (which is more important). Will repeat or dual-purpose that client if successful. @rabernat, see this comment, seems like there is a way to get credentials objects now, pandas/pydata blessed. https://pandas-gbq.readthedocs.io/en/latest/howto/authentication.html#authenticating-with-a-user-account I am thinking, than gdrive might not be great for throughput, but it's be a great way to share intake catalogues
gharchive/issue
2019-11-16T04:52:14
2025-04-01T06:44:31.782495
{ "authors": [ "martindurant", "rabernat" ], "repo": "intake/gdrivefs", "url": "https://github.com/intake/gdrivefs/issues/7", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2187448243
🛑 Shopintake is down In 84478c0, Shopintake (https://www.shopintake.com) was down: HTTP code: 403 Response time: 446 ms Resolved: Shopintake is back up in cddb4f6 after 16 minutes.
gharchive/issue
2024-03-14T23:30:33
2025-04-01T06:44:31.785014
{ "authors": [ "intakefoods" ], "repo": "intakefoods/status.intakefoods.kr", "url": "https://github.com/intakefoods/status.intakefoods.kr/issues/1985", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2287266253
[PyTorch Upstream] Stock pytorch fp16 e2e test got segmentfault from triton. We got segmentfault when runing stock pytorch fp16 end2end test. we've narrow down a mini reporducer as follow: import faulthandler; faulthandler.enable() import torch from torch import tensor, device import torch.fx as fx from torch._dynamo.testing import rand_strided from math import inf import torch._inductor.inductor_prims import torch._dynamo.config import torch._inductor.config import torch._functorch.config import torch.fx.experimental._config torch._inductor.config.fallback_random = True torch._inductor.config.freezing = True torch._inductor.config.triton.cudagraphs = True torch._functorch.config.unlift_effect_tokens = True torch._functorch.config.debug_partitioner = True isolate_fails_code_str = None from torch.nn import * class Repro(torch.nn.Module): def __init__(self): super().__init__() def forward(self, arg0_1): isnan = torch.ops.aten.isnan.default(arg0_1); arg0_1 = None any_1 = torch.ops.aten.any.default(isnan); isnan = None return (any_1,) def load_args(reader): buf0 = reader.storage(None, 2097152, device=device(type='xpu', index=0), dtype_hint=torch.float16) reader.tensor(buf0, (1, 1024, 1024), dtype=torch.float16, is_leaf=True) # arg0_1 load_args._version = 0 mod = Repro() if __name__ == '__main__': from torch._dynamo.repro.after_aot import run_repro with torch.no_grad(): run_repro(mod, load_args, accuracy=False, command='run', save_dir=None, tracing_mode='real', check_str=None) We got the callstack as follow: Current thread 0x00007ffb1fd0c740 (most recent call first): File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/site-packages/triton/compiler/compiler.py", line 369 in _init_handles File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/runtime/triton_heuristics.py", line 406 in _precompile_config File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/runtime/triton_heuristics.py", line 203 in precompile File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/codecache.py", line 2932 in result File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/codecache.py", line 3128 in wait File "/tmp/torchinductor_xinanlin/lq/clqdbzfwkffulbfq265wdizbwhp2hb3vvnzqekc75wzbkcniuxem.py", line 77 in <module> File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/runtime/compile_tasks.py", line 44 in _reload_python_module File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/codecache.py", line 2567 in load_by_key_path File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/graph.py", line 1657 in compile_to_module File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/utils.py", line 273 in time_wrapper File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/graph.py", line 1710 in compile_to_fn File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/compile_fx.py", line 803 in fx_codegen_and_compile File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/contextlib.py", line 79 in inner File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/compile_fx.py", line 507 in compile_fx_inner File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/utils.py", line 273 in time_wrapper File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/contextlib.py", line 79 in inner File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/contextlib.py", line 79 in inner File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/debug.py", line 304 in inner File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/repro/after_aot.py", line 83 in debug_wrapper File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/repro/after_aot.py", line 708 in repro_run File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/repro/after_aot.py", line 957 in run_repro File "/home/pt-gpu/4T-4652/xinanlin/pytorch/../test_any.py", line 41 in <module> To reproduce, please build stock pytorch with env "USE_XPU=1", and run the above script. @riverliuintel @vlad-penkin this issue blocked Inductor upstream process, please priotize, thanks. I will investigate and report back. The corresponding triton kernel: from ctypes import c_void_p, c_long import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _align as align from torch import device, empty_strided from torch._inductor.codecache import AsyncCompile from torch._inductor.select_algorithm import extern_kernels from torch._inductor.codegen.multi_kernel import MultiKernelCall aten = torch.ops.aten inductor_ops = torch.ops.inductor assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cpu = torch._C._dynamo.guards._empty_strided_cpu empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda alloc_from_pool = torch.ops.inductor._alloc_from_pool reinterpret_tensor = torch.ops.inductor._reinterpret_tensor async_compile = AsyncCompile() # kernel path: /tmp/torchinductor_xinanlin/rb/crb7vl4jnylgbdeqtxzpnde4vgafzac4akzqlpwhnozqn3sna272.py # Source Nodes: [any_1, isnan], Original ATen: [aten.any, aten.isnan] # any_1 => any_1 # isnan => isnan triton_red_fused_any_isnan_0 = async_compile.triton('triton_', ''' import triton import triton.language as tl from triton.compiler.compiler import AttrsDescriptor from torch._inductor import triton_helpers, triton_heuristics from torch._inductor.ir import ReductionHint, TileHint from torch._inductor.triton_helpers import libdevice, math as tl_math from torch._inductor.triton_heuristics import AutotuneHint from torch._inductor.utils import instance_descriptor @triton_heuristics.reduction( size_hints=[1, 1048576], reduction_hint=ReductionHint.DEFAULT, filename=__file__, triton_meta={'signature': {0: '*fp16', 1: '*i1', 2: 'i32', 3: 'i32'}, 'device': 0, 'device_type': 'xpu', 'constants': {2: 1}, 'configs': [AttrsDescriptor( divisible_by_16=(0, 1, 3), equal_to_1=(2,))]}, inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_any_isnan_0', 'mutated_arg_names': [], 'no_x_dim': False, 'backend_hash': '196aef a0a85bda494694e78ae7765ffc5ac720f56091b59c964159c6f79c4a45'} ) @triton.jit [19/92120] def triton_(in_ptr0, out_ptr0, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): xnumel = 1 rnumel = 1048576 xoffset = tl.program_id(0) * XBLOCK xindex = xoffset + tl.arange(0, XBLOCK)[:, None] xmask = xindex < xnumel rbase = tl.arange(0, RBLOCK)[None, :] _tmp3 = tl.full([XBLOCK, RBLOCK], 0, tl.int1) for roffset in range(0, rnumel, RBLOCK): rindex = roffset + rbase rmask = rindex < rnumel r0 = rindex tmp0 = tl.load(in_ptr0 + (r0), None, eviction_policy='evict_first').to(tl.float32) tmp1 = libdevice.isnan(tmp0).to(tl.int1) tmp2 = tl.broadcast_to(tmp1, [XBLOCK, RBLOCK]) tmp4 = _tmp3 | tmp2 _tmp3 = tmp4 tmp3 = triton_helpers.any(_tmp3.to(tl.int8), 1)[:, None].to(tl.int1) tl.store(out_ptr0 + (tl.full([XBLOCK, 1], 0, tl.int32)), tmp3, None) ''', device_str='xpu') import triton import triton.language as tl from torch._inductor.triton_heuristics import grid, split_scan_grid, start_graph, end_graph from torch._C import _xpu_getCurrentRawStream as get_raw_stream async_compile.wait(globals()) del async_compile def call(args): arg0_1, = args args.clear() assert_size_stride(arg0_1, (1, 1024, 1024), (1048576, 1024, 1)) with torch.xpu._DeviceGuard(0): torch.xpu.set_device(0) buf0 = empty_strided((), (), device='xpu', dtype=torch.bool) # Source Nodes: [any_1, isnan], Original ATen: [aten.any, aten.isnan] stream0 = get_raw_stream(0) triton_red_fused_any_isnan_0.run(arg0_1, buf0, 1, 1048576, grid=grid(1), stream=stream0) del arg0_1 return (buf0, ) def benchmark_compiled_module(times=10, repeat=10): from torch._dynamo.testing import rand_strided from torch._inductor.utils import print_performance arg0_1 = rand_strided((1, 1024, 1024), (1048576, 1024, 1), device='xpu:0', dtype=torch.float16) fn = lambda: call([arg0_1]) return print_performance(fn, times=times, repeat=repeat) if __name__ == "__main__": from torch._inductor.wrapper_benchmark import compiled_module_main compiled_module_main('BartForConditionalGeneration', benchmark_compiled_module) Looks like the crash is in IGC: Program received signal SIGSEGV, Segmentation fault. 0x00007ffff319d574 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 (gdb) bt #0 0x00007ffff319d574 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #1 0x00007ffff31a7403 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #2 0x00007ffff31a7ccc in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #3 0x00007ffff3be0aee in llvm::FPPassManager::runOnFunction(llvm::Function&) () from /lib/x86_64-linux-gnu/libigc.so.1 #4 0x00007ffff3be0e14 in llvm::FPPassManager::runOnModule(llvm::Module&) () from /lib/x86_64-linux-gnu/libigc.so.1 #5 0x00007ffff3be1bac in llvm::legacy::PassManagerImpl::run(llvm::Module&) () from /lib/x86_64-linux-gnu/libigc.so.1 #6 0x00007ffff3153b2c in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #7 0x00007ffff2e0c321 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #8 0x00007ffff3050afb in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #9 0x00007ffff2e0e307 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #10 0x00007ffff2e7c4a5 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #11 0x00007ffff2e7da3e in ?? () from /lib/x86_64-linux-gnu/libigc.so.1 #12 0x00007ffff7efac64 in NEO::OfflineCompiler::buildSourceCode() () from /lib/x86_64-linux-gnu/libocloc.so #13 0x00007ffff7efe4e5 in NEO::OfflineCompiler::build() () from /lib/x86_64-linux-gnu/libocloc.so #14 0x00007ffff7f366ce in int SafetyGuardLinux::call<int, NEO::OfflineCompiler, int (NEO::OfflineCompiler::*)()>(NEO::OfflineCompiler*, int (NEO::OfflineCompiler::*)(), int) () from /lib/x86_64-linux-gnu/libocloc.so #15 0x00007ffff7f363ee in buildWithSafetyGuard(NEO::OfflineCompiler*) () from /lib/x86_64-linux-gnu/libocloc.so #16 0x00007ffff7ef19c8 in Ocloc::Commands::compile(OclocArgHelper*, std::vector<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::allocator<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > const&) () from /lib/x86_64-linux-gnu/libocloc.so #17 0x00007ffff7edda35 in oclocInvoke () from /lib/x86_64-linux-gnu/libocloc.so #18 0x0000555555554787 in main () I have the spirv dumps that cause the IGC crash and will submit a ticket to the IGC team. Hi @alexbaden , can you share the IGC ticket link? @etaf you've been added as watcher to the IGC ticket and should've receive an email notification. Hi, @alexbaden @vlad-penkin how can I get the fixed driver?
gharchive/issue
2024-05-09T09:18:10
2025-04-01T06:44:31.882510
{ "authors": [ "alexbaden", "etaf", "vlad-penkin" ], "repo": "intel/intel-xpu-backend-for-triton", "url": "https://github.com/intel/intel-xpu-backend-for-triton/issues/1073", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1499526769
Fix onnxrt numpy round bug Signed-off-by: yuwenzho yuwen.zhou@intel.com Type of Change bug fix Description Defect/regression: TypeError: type numpy.ndarray doesn't define round method Profile: Report: https://inteltf-jenk.sh.intel.com/view/lpot-release/job/intel-lpot-validation-top-release-v2.0/3/artifact/report.html commit id: ae3cf5638ca96052032f11dca2120dce5c8d0316 log: https://inteltf-jenk.sh.intel.com/job/intel-lpot-validation-top-release-v2.0/3/artifact/linux/onnxrt/mobilenetv2-12/onnxrt-mobilenetv2-12-linux-clx8280-tune.log Framework: onnx Framework version: 1.13.1 JIRA ticket: ILITV-2528 How has this PR been tested? extension test: mobilenetv2-12 Dependency Change? NO extension test: https://inteltf-jenk.sh.intel.com/job/intel-lpot-validation-top-mr-extension/3786/
gharchive/pull-request
2022-12-16T03:45:24
2025-04-01T06:44:31.905234
{ "authors": [ "yuwenzho" ], "repo": "intel/neural-compressor", "url": "https://github.com/intel/neural-compressor/pull/287", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
897430483
Support parsing empty records in patterns Issue The following valid Elm code is not being properly parsed emptyRecordFn {} = 1 Pull request status This pull request does two things: Add failing test cases (in the complete suite) Change the parser to allow an optional list of valid fields Note Using the "Add type annotation" intention for a {} = 1 now generates a : { a | } -> number as a type annotation. The compiler itself has a similar problem, which I have pointed at in https://github.com/elm/compiler/issues/2204. I think this is an okay problem to have, especially since there isn't any valid type that could be infered here. LGTM! I'm taking over steward ship of this project and will merge this for the next release. @jfmengels Could you merge master into that branch? I fixed the CI :)
gharchive/pull-request
2021-05-20T20:41:39
2025-04-01T06:44:31.935666
{ "authors": [ "cies", "jfmengels" ], "repo": "intellij-elm/intellij-elm", "url": "https://github.com/intellij-elm/intellij-elm/pull/754", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
243238460
Feature/prepare maven central This PR updates the gradle build file to allow for code signing and artifact deploy to Sonatype, which mirrors out to Maven Central. Additionally, there are minor formatting and Javadoc changes. This PR is a good first step - it allows me to manually publish SNAPSHOTS and releases, but I'd like to get CircleCI at least publishing SNAPSHOTS when the development branch is built. Unfortunately, I am unable to get gradle to read the environment variables from CircleCI, so that's not yet working -- I plan to spend some more time on that this weekend. Just pushed one more change for the naming - let me know what you think @adrianhopebailie
gharchive/pull-request
2017-07-16T13:08:00
2025-04-01T06:44:31.991146
{ "authors": [ "sappenin" ], "repo": "interledger/java-ilp-core", "url": "https://github.com/interledger/java-ilp-core/pull/55", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1780814448
1505/filter grants Changes proposed in this pull request Context Checklist [ ] Related issues linked using fixes #number [ ] Tests added/updated [ ] Documentation added [ ] Make sure that all checks pass Accidentally opened PR
gharchive/pull-request
2023-06-29T13:23:46
2025-04-01T06:44:31.993610
{ "authors": [ "Tymmmy" ], "repo": "interledger/rafiki", "url": "https://github.com/interledger/rafiki/pull/1532", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
505821780
Set up autopep8 as a task and add guidelines to the readme Since we have the pep8speaks bot commenting on our issues, we ought to make it easy for people to format things in a way that makes pep8 happy. This project looks like it would do it: https://pypi.org/project/autopep8/ task add autopep8 to requirements.txt add instructions to the readme file telling people how to run it before making PRs. if you want to pick up this task Take a look at the InterMine contributing guidelines Comment on this issue stating that you intend to work on the task When you're ready, add your work to the repo and create a pull request. What to do if you need help Mention @yochannah, tweet @yoyehudi, pop by to say hi on chat or if needed email yo@intermine.org. Don't forget we're usually only available during uk office hours and will not be able to respond at other times :) Hi I would like to work on this task @barhenkro fantastic - please go ahead, and let us know if you have any questions! :) fixed by #52 - thank you so much @barhenkro !
gharchive/issue
2019-10-11T12:26:11
2025-04-01T06:44:32.007177
{ "authors": [ "barhenkro", "yochannah" ], "repo": "intermine/intermine-ws-python", "url": "https://github.com/intermine/intermine-ws-python/issues/51", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
143210281
Events from 27 to 31 of march Hi, I have an issue with events at the end of march, but not of all years and only in "month" view. Starting from the input below, I had a wrong result where I put a "-" sign, and a good result where I put a "+" sign: - "start":new Date(2011, 2, 28),"end":new Date(2011, 2, 28) - "start":new Date(2012, 2, 28),"end":new Date(2012, 2, 28) + "start":new Date(2013, 2, 28),"end":new Date(2013, 2, 28) + "start":new Date(2014, 2, 28),"end":new Date(2014, 2, 28) + "start":new Date(2015, 2, 28),"end":new Date(2015, 2, 28) - "start":new Date(2016, 2, 28),"end":new Date(2016, 2, 28) - "start":new Date(2017, 2, 28),"end":new Date(2017, 2, 28) - "start":new Date(2018, 2, 28),"end":new Date(2018, 2, 28) + "start":new Date(2019, 2, 28),"end":new Date(2019, 2, 28) + "start":new Date(2020, 2, 28),"end":new Date(2020, 2, 28) + "start":new Date(2021, 2, 28),"end":new Date(2021, 2, 28) What happens? Thanks in advance, Davide Thanks to solve this issue so fast, I tried the fix on my locale machine. When did you think to release this fix officially on NPM? thanks, Davide
gharchive/issue
2016-03-24T10:31:05
2025-04-01T06:44:32.119245
{ "authors": [ "jesty" ], "repo": "intljusticemission/react-big-calendar", "url": "https://github.com/intljusticemission/react-big-calendar/issues/62", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
760507333
Add boundserror when indexing into an invalid column Fix #211 @c42f Is this a good enough error or do I need something better here? I wanted to use BoundsError but it's quite limited for display. I think throwing any exception is a great improvement already, so I'd be happy with this. Having said that, I think this error would be a bit mysterious to users as they're dealing with the column name in using getproperty, but the error emitted here has no mention of the column name: julia> row = execute(conn, "select * from (values (1, 2), (3, 4)) as test (a,b)") |> first LibPQ.Row(PostgreSQL result, 1) julia> row.a 1 julia> row.x ERROR: BoundsError: attempt to access 2-element Array{String,1} at index [0] Stacktrace: [1] column_number at /home/ubuntu/.julia/dev/LibPQ/src/results.jl:481 [inlined] [2] getindex(::LibPQ.Result, ::Int64, ::Int64) at /home/ubuntu/.julia/dev/LibPQ/src/results.jl:509 [3] getproperty(::LibPQ.Row, ::Symbol) at /home/ubuntu/.julia/dev/LibPQ/src/tables.jl:37 [4] top-level scope at REPL[35]:1 A nice extension could be to add some @boundscheck logic to getproperty as well. I'm not sure what error type to throw though. It seems that the error type used for Base when accessing a nonexistent field of a type is ErrorException... which is nice and descriptive but nonspecific in the type.
gharchive/pull-request
2020-12-09T17:05:36
2025-04-01T06:44:32.147614
{ "authors": [ "c42f", "iamed2" ], "repo": "invenia/LibPQ.jl", "url": "https://github.com/invenia/LibPQ.jl/pull/212", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
624947694
Cache artifacts in Travis, AppVeyor, GitHubActions, and Cirrus Closes #166 I've tested Travis, AppVeyor, and GHA, but I couldn't get Cirrus to pick up on my repo for some reason. According to the docs it should work. GitLab doesn't allow caching files outside of the repository directory, so caching can't be used there. DroneCI's caching looks a bit more complicated so maybe I'll look at it another day, it's not used frequently anyways. Awesome, thank you so much! Cirrus is being very unreliable lately, many packages moved to Travis for FreeBSD testing. As far as I know, Cloud Drone doesn't support caching, there are plugins for self-hosted instances. How does FreeBSD on Travis work? And good to know that the Drone stuff will only work on self-hosted, so not much use implementing it here. How does FreeBSD on Travis work? Pretty good, just add - freebsd to the list of operating systems, that's it It seems that it's not mentioned in the Travis docs anywhere... I think it's in beta or something like that, but Alex shared the news on Slack :slightly_smiling_face: Ah, cool. I'll wait until it's GA before I add the option here.
gharchive/pull-request
2020-05-26T14:55:23
2025-04-01T06:44:32.151245
{ "authors": [ "christopher-dG", "giordano" ], "repo": "invenia/PkgTemplates.jl", "url": "https://github.com/invenia/PkgTemplates.jl/pull/173", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
708265063
(Optional?) randomized ports for isolated services Is your feature request related to a problem? Please describe. Imagine the following scenario: I'm running my separate Docker services for some other instance/application (e.g. Zenodo local development). These services are exposing standard ports, i.e. ES on 9200, PostgreSQL on 5432, etc. I take a break to install a separate Invenio module, e.g. invenio-records-rest to make a bugfix. I run invenio-records-rest's tests, but guess what docker-services-cli will have to either: Fail hard and exit, since it will try to bring up services that have already bound ports (the ones from my Zenodo local dev Docker services). This is actually a good/preferred behavior Partially/silently fail and re-use my existing Docker services. This is pretty bad, since as part of tests it might end up modifying/wiping the services, which affect my Zenodo local development setup. Describe the solution you'd like I think checking if the current implementation of docker-services-cli fails hard (and possibly with an informative message) in the above use-case is the bare minimum to make sure things are safe. An alternative, more "sophisticated" solution, would be to bring up each Docker service with a randomized available port (preferably in a higher port range, e.g. 50000+), that can then be exposed to the underlying tests, e.g. via environment variables. A downside of the randomized ports solution is that if one wants mid-test to inspect one of the Docker services state, they have to look up what is the current randomized port and modify their connection parameters accordingly. Another alternative is to standardized these "testing" ports, so that we avoid clashes with other local development ports. E.g. we put ES always on 59200, PostgreSQL always on 55432, etc. To summarize here are the available solutions: [ ] Check that we currently fail hard... [ ] ...with an informative message [ ] Randomized ports, exposed in the output as environment variable [ ] Decide and use alternative "standardized" ports Closed via #39
gharchive/issue
2020-09-24T15:23:20
2025-04-01T06:44:32.156414
{ "authors": [ "slint" ], "repo": "inveniosoftware/docker-services-cli", "url": "https://github.com/inveniosoftware/docker-services-cli/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2076644017
anonymization: fix orders with patron_pid fixes an error when anonymizing an order that does not have the patron_pid in the order lines Closing in favour of: https://github.com/inveniosoftware/invenio-app-ils/pull/1193
gharchive/pull-request
2024-01-11T13:23:05
2025-04-01T06:44:32.160100
{ "authors": [ "ntarocco" ], "repo": "inveniosoftware/invenio-app-ils", "url": "https://github.com/inveniosoftware/invenio-app-ils/pull/1192", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1119786768
membership: invite user to community closes #https://github.com/inveniosoftware/invenio-communities/issues/379 This is a tracer bullet PR. There are a lot of TODOs listed on the board to complete next. I've highlighted areas where feedback would be good. unfortunately, i don't know about the alembic error... i just re-ran the checks on master, and they fail too (with a slightly different error, though), which sounds a bit like a files-related module to me: AssertionError: assert not [('remove_index', Index('ix_uq_partial_files_object_is_head', Column('bucket_id', UUID(), table=<files_object>, nullable=False), Column('key', TEXT(), table=<files_object>, nullable=False)))] The same assert not ext.alembic.compare_metadata() E AssertionError: assert not [('remove_index', Index('ix_uq_partial_files_object_is_head', Column('bucket_id', UUID(), table=<files_object>, nullable=False), Column('key', TEXT(), table=<files_object>, nullable=False)))] error that is affecting the master branch, invenio-rdm-records PRs and basically every repository using invenio-files-rest is affecting this PR too. All the tests pass locally, but I will feel better when this is solved prior to merging. Some of the remarks I couldn't quite address anyway, so it might be worth taking another look. Not further points than the ones already discussed.
gharchive/pull-request
2022-01-31T18:29:31
2025-04-01T06:44:32.163118
{ "authors": [ "fenekku", "lnielsen", "max-moser" ], "repo": "inveniosoftware/invenio-communities", "url": "https://github.com/inveniosoftware/invenio-communities/pull/380", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
942419788
subjects: account for pre-existing vocabularies when loading [+] depends on https://github.com/inveniosoftware/invenio-vocabularies/pull/67 integrate new separate subjects vocabulary (invenio-vocabularies 0.7.8) rely on database content before loading vocabulary closes #686 (and #687 for now) I would actually agree that most of the vocabulary loading would make sense in invenio-vocabularies. It wasn't moved there because: a- Initial loading was in here, so just continued from it b- Dealing with with default vocabularies if the loading logic is placed in invenio-vocabularies is a complex lift in itself best done separately. (more below) Placing the loading responsibility in invenio-vocabularies makes sense, but placing the default vocabularies values in invenio-vocabularies doesn't (at least for me). It's the schema defining package (invenio-rdm-records) that sets out those default values because it knows the kind of records it's dealing with (e.g. InvenioRDM has these default resource types, but other users of invenio-vocabularies may not even have resource types). So then we would need another mechanism for invenio-vocabularies to use things defined in rdm-records without importing it (circular dependency). Entrypoints or a registry-like pattern like seen above could be it. But those are both complex and I already spent my complexity points in the loading code for this PR. Placing the loading responsibility in invenio-vocabularies makes sense, but placing the default vocabularies values in invenio- vocabularies doesn't (at least for me) Agree completely So then we would need another mechanism for invenio-vocabularies to use things defined in rdm-records without importing it (circular dependency) I see... Need to check but if the vocabularies.yaml is made "generic" (sort of as it is now already) it can be used as a point of entry for vocabularies. Then rdm-records would define the defaults in its module. And the mechanics to get from entry points is already coded.
gharchive/pull-request
2021-07-12T20:26:57
2025-04-01T06:44:32.167949
{ "authors": [ "fenekku", "ppanero" ], "repo": "inveniosoftware/invenio-rdm-records", "url": "https://github.com/inveniosoftware/invenio-rdm-records/pull/734", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
407309540
invenio-s3 repeats the bucket name as path the name of the bucket from the S3 configuration gets somehow repeated as part of path, so that files are written to s3://bucket_name/bucket_name/files I am no longer able to reproduce the issue. It seems that the files are now in the correct location. When using the URI s3://default/f7/14/adaa-04f2-48e8-9a6b-de2f57948ce1/data stored in files_files.uri, the following file structure is created: $ tree data data └── default └── f7 └── 14 └── adaa-04f2-48e8-9a6b-de2f57948ce1 └── data @Herrner, please feel free to reopen the issue if you are still able to replicate it.
gharchive/issue
2019-02-06T16:07:57
2025-04-01T06:44:32.169915
{ "authors": [ "Herrner", "egabancho" ], "repo": "inveniosoftware/invenio-s3", "url": "https://github.com/inveniosoftware/invenio-s3/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2098551393
dependency: added tinymce :heart: Thank you for your contribution! Description Please describe briefly your pull request. Checklist Ticks in all boxes and 🟢 on all GitHub actions status checks are required to merge: [ ] I'm aware of the code of conduct. [ ] I've created logical separate commits and followed the commit message format. [ ] I've added relevant test cases. [ ] I've added relevant documentation. [ ] I've marked translation strings (for relevant code). [ ] I've followed the CSS/JS and React guidelines (for relevant code). [ ] I've followed the web accessibility guidelines (for relevant code). [ ] I've followed the user interface guidelines (for relevant code). [ ] I've identified the copyright holder(s) and updated copyright headers for touched files (>15 lines contributions). [ ] I've NOT included third-party code (copy/pasted source code or new dependencies). Third-party code If you've added third-party code (copy/pasted or new dependencies), please reach out to an architect. Reminder By using GitHub, you have already agreed to the GitHub’s Terms of Service including that: You license your contribution under the same terms as the current repository’s license. You agree that you have the right to license your contribution under the current repository’s license. Could you please check that we don't have any other occurrences in other repositories (except react-invenio-deposit which is deprecated)?
gharchive/pull-request
2024-01-24T15:43:14
2025-04-01T06:44:32.177955
{ "authors": [ "0einstein0", "ptamarit" ], "repo": "inveniosoftware/invenio-vocabularies", "url": "https://github.com/inveniosoftware/invenio-vocabularies/pull/289", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2435542475
🛑 Google is down In d912013, Google (https://www.google.com) was down: HTTP code: 429 Response time: 354 ms Resolved: Google is back up in 88fa85e after 10 minutes.
gharchive/issue
2024-07-29T14:10:17
2025-04-01T06:44:32.180595
{ "authors": [ "peschee" ], "repo": "inventage/upptime", "url": "https://github.com/inventage/upptime/issues/392", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1349680732
docs: Update EAS installation instructions Thank you for having great setup instructions! I noticed that this part was slightly out of date and no longer required so I updated it for future readers. Interesting, if the default image now contains it, but the version numbers are going to out of date in the future (r21 will become r22+, ubuntu 20 will go to 22 etc), perhaps there is a more future proof + prescriptive way to word it? something like: if you are using EAS default image you are set (do we need to specify EAS greater than or equal to some version or date? I'm not sure how EAS works, if it is all in cloud and everyone is updated, or if there are versions) if you are not using the EAS image you must specify one that includes JDK11. You may find them (link to Expo docs...) Then we don't have to worry about versions Makes sense! I will rework this to make it more future proof.
gharchive/pull-request
2022-08-24T15:58:55
2025-04-01T06:44:32.200719
{ "authors": [ "thatzacdavis" ], "repo": "invertase/notifee", "url": "https://github.com/invertase/notifee/pull/500", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1472131268
[enhancement]: Write prompt to a standard PNG EXIF field Is there an existing issue for this? [X] I have searched the existing issues Contact Details No response What should this feature add? We could write the prompt (with no other metadata) to a standard PNG EXIF e.g. Description. Then normal image viewers can at least see it. Alternatives No response Aditional Content No response From NemesisFlux @ discord: Maybe just to help out people like me, a note could be put in the comments field, saying something like, "InvokeAI parameters for this image are stored in tEXt fields in the PNG file. You can examine them with the free third-party program, pngcheck." Example implementation: https://gitlab.com/keturn/kikifusion/-/blob/890be52914713d33757c5b900b03ad91c735be21/src/qittnz6Lz/metadata.py#L77 ImageIFD.ImageDescription is probably the most widely-supported, but doesn't cope with multi-byte charsets like UTF-8 or UTF-16, so it's not great. ExifIFD.UserComment does handle unicode and is supported by at least some applications. I've noticed this request most often from Mac users. Is anyone able to point us to docs on which fields OS X displays in Finder and includes in its search index?
gharchive/issue
2022-12-02T00:15:35
2025-04-01T06:44:32.286456
{ "authors": [ "keturn", "psychedelicious" ], "repo": "invoke-ai/InvokeAI", "url": "https://github.com/invoke-ai/InvokeAI/issues/1680", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1481618066
[bug]: Expected all tensors to be on the same device exception Is there an existing issue for this? [X] I have searched the existing issues OS Windows GPU cuda VRAM 4GB What happened? I can make iterations just perfect in TextToImage but once i try to Outpaint in the Unified editor it throws an error Model im using is the: Prompthero/openjourney Once im using the original 1.4 or 1.5 model this does not happen, only happens with certain models from Huggingface. RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking >argument for argument mat1 in method wrapper_addmm) Full error: Traceback (most recent call last): File "d:\ai\invokeai\ldm\generate.py", line 486, in prompt2image results = generator.generate( File "d:\ai\invokeai\ldm\invoke\generator\base.py", line 93, in generate image = make_image(x_T) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "d:\ai\invokeai\ldm\invoke\generator\inpaint.py", line 295, in make_image samples = sampler.decode( File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "d:\ai\invokeai\ldm\models\diffusion\sampler.py", line 365, in decode outs = self.p_sample( File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "d:\ai\invokeai\ldm\models\diffusion\ddim.py", line 58, in p_sample e_t = self.invokeai_diffuser.do_diffusion_step( File "d:\ai\invokeai\ldm\models\diffusion\shared_invokeai_diffusion.py", line 88, in do_diffusion_step unconditioned_next_x, conditioned_next_x = self.apply_standard_conditioning(x, sigma, unconditioning, conditioning) File "d:\ai\invokeai\ldm\models\diffusion\shared_invokeai_diffusion.py", line 104, in apply_standard_conditioning unconditioned_next_x, conditioned_next_x = self.model_forward_callback(x_twice, sigma_twice, File "d:\ai\invokeai\ldm\models\diffusion\ddim.py", line 13, in model_forward_callback = lambda x, sigma, cond: self.model.apply_model(x, sigma, cond)) File "d:\ai\invokeai\ldm\models\diffusion\ddpm.py", line 1441, in apply_model x_recon = self.model(x_noisy, t, **cond) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "d:\ai\invokeai\ldm\models\diffusion\ddpm.py", line 2167, in forward out = self.diffusion_model(x, t, context=cc) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "d:\ai\invokeai\ldm\modules\diffusionmodules\openaimodel.py", line 798, in forward emb = self.time_embed(t_emb) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\container.py", line 139, in forward input = module(input) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\linear.py", line 114, in forward return F.linear(input, self.weight, self.bias) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_addmm) Could not generate image. Screenshots No response Additional context No response Contact Details No response Problem solved If you use conda clean it will break InvokeAI. My bad Do other models work? Do other models work? Seems i was wrong to conclude that, its very weird. when i reinstalled InvokeAI, using the standard 1.5 model i could use all of the features, but when i started on a new second prompt (not unified editor) it gave me this error. I found some that might shed light on this error, the error seems to stay away when you do not use the --free_gpu_mem flag in the start up. Perhaps the problem resides in this flag code. This error was fixed in #1938. This error was fixed in #1938. My bad, i wasn't aware of this fix yet, thank you letting me know!
gharchive/issue
2022-12-07T10:52:16
2025-04-01T06:44:32.304156
{ "authors": [ "TheBarret", "psychedelicious", "rmagur1203" ], "repo": "invoke-ai/InvokeAI", "url": "https://github.com/invoke-ai/InvokeAI/issues/1843", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1419793408
Web UI: Fixes same seed reported when with_variations=[] and iterations > 1 Web UI: Fixes same seed reported when with_variations=[] and iterations > 1 Fixes #1223 @ausbitbank can you please try this change out? I pulled your repository and checked out the fix-incorrect-seed branch , started the api and tested but got the same result still. request (intercepted and edited to add with_variations: [] using burpsuite) 42["generateImage",{"prompt":"test","iterations":2,"steps":15,"cfg_scale":7.5,"threshold":0,"perlin":0,"height":512,"width":512,"sampler_name":"k_lms","seed":3471630626,"seamless":false,"hires_fix":false,"progress_images":false,"variation_amount":0,"with_variations":[]},false,false] both responses have the same seed: 42["generationResult",{"url":"outputs/000001.3471630626.png","mtime":1666535971.5529692,"metadata":{"model":"stable diffusion","model_id":null,"model_hash":"fe4efff1e174c627256e44ec2991ba279b3816e364b49f9be2abc0b3ff3f8556","app_id":"lstein/stable-diffusion","app_version":"v1.15","image":{"prompt":[{"prompt":"test","weight":1.0}],"steps":15,"cfg_scale":7.5,"threshold":0,"perlin":0,"height":512,"width":512,"seed":3471630626,"seamless":false,"hires_fix":false,"postprocessing":null,"sampler":"k_lms","variations":[],"type":"txt2img"}}}] 42["generationResult",{"url":"outputs/000002.3471630626.png","mtime":1666535977.6119702,"metadata":{"model":"stable diffusion","model_id":null,"model_hash":"fe4efff1e174c627256e44ec2991ba279b3816e364b49f9be2abc0b3ff3f8556","app_id":"lstein/stable-diffusion","app_version":"v1.15","image":{"prompt":[{"prompt":"test","weight":1.0}],"steps":15,"cfg_scale":7.5,"threshold":0,"perlin":0,"height":512,"width":512,"seed":3471630626,"seamless":false,"hires_fix":false,"postprocessing":null,"sampler":"k_lms","variations":[],"type":"txt2img"}}}] @psychedelicious should this be merged? @psychedelicious Is this PR still in need of merging? @lstein I'm guessing no, since it doesn't solve the original issue of repeating seeds. Same issue still exists in the latest dev branch, but I would say its very low priority as it does not affect API clients giving the correct parameters as demonstrated by web ui..
gharchive/pull-request
2022-10-23T13:21:48
2025-04-01T06:44:32.309773
{ "authors": [ "ausbitbank", "lstein", "psychedelicious" ], "repo": "invoke-ai/InvokeAI", "url": "https://github.com/invoke-ai/InvokeAI/pull/1224", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2093993635
name generation of generic types Is it possible to customise how the names of generic references are generated? If I have for example the following: type Foo int type Bar int type FooMap[T any] map[Foo]T type struct FooBar { FooBar FooMap[Bar] } The resulting $ref for the FooBar $def is FooMap[github.com/full/path/to/package.Bar]. I'd like to change this to FooMap[Bar]. I wonder if this will be easier once Go exposes the generic child type via reflection, but for now I've been solving this with a recursive namer. I'm not sure what other types like arrays, maps, or slices show up as though. var genericTypeRegex = regexp.MustCompile(`^(.+)\[(.+)\]$`) func getGenericName(name string) string { if matches := genericTypeRegex.FindStringSubmatch(name); matches != nil { parent := matches[1] parts := strings.Split(matches[2], ".") child := parts[len(parts)-1] return fmt.Sprintf("%s[%s]", parent, getGenericName(child)) } return name } func nameGenerics(r reflect.Type) string { return getGenericName(r.Name()) } and then use it like reflector := jsonschema.Reflector{ Namer: nameGenerics } s := reflector.Reflect(myType) If you'd like to add some broken tests for this, it'd be easier to have a look and align expectations. Generics were not available when this library was first built!
gharchive/issue
2024-01-22T14:03:59
2025-04-01T06:44:32.312890
{ "authors": [ "joshstrohminger", "olistrik", "samlown" ], "repo": "invopop/jsonschema", "url": "https://github.com/invopop/jsonschema/issues/131", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2434009595
🛑 Bitwage API (Sandbox) is down In 8c7d3ce, Bitwage API (Sandbox) (https://api.sandbox.bitwage.com) was down: HTTP code: 403 Response time: 62 ms Resolved: Bitwage API (Sandbox) is back up in 1be1cf9 after 36 minutes.
gharchive/issue
2024-07-28T16:12:41
2025-04-01T06:44:32.315602
{ "authors": [ "joelinzy" ], "repo": "inwage/status_page", "url": "https://github.com/inwage/status_page/issues/3490", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1073492931
🛑 bitwage api production is down In e4e12bd, bitwage api production (https://api2.bitwage.com) was down: HTTP code: 500 Response time: 547 ms Resolved: bitwage api production is back up in 7aea99c.
gharchive/issue
2021-12-07T15:34:16
2025-04-01T06:44:32.318183
{ "authors": [ "joelinzy" ], "repo": "inwage/status_page", "url": "https://github.com/inwage/status_page/issues/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2549119731
🛑 Bitwage API (Sandbox) is down In 921278a, Bitwage API (Sandbox) (https://api.sandbox.bitwage.com) was down: HTTP code: 403 Response time: 139 ms Resolved: Bitwage API (Sandbox) is back up in 922e67c after 17 minutes.
gharchive/issue
2024-09-25T22:43:36
2025-04-01T06:44:32.320843
{ "authors": [ "joelinzy" ], "repo": "inwage/status_page", "url": "https://github.com/inwage/status_page/issues/5330", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2634121882
🛑 Bitwage API (Sandbox) is down In bbefb6f, Bitwage API (Sandbox) (https://api.sandbox.bitwage.com) was down: HTTP code: 403 Response time: 117 ms Resolved: Bitwage API (Sandbox) is back up in 4572de8 after 52 minutes.
gharchive/issue
2024-11-04T23:57:23
2025-04-01T06:44:32.323273
{ "authors": [ "joelinzy" ], "repo": "inwage/status_page", "url": "https://github.com/inwage/status_page/issues/6517", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
771537289
Add good-scroll and pixel-scroll to SmoothScrolling EmacsWiki page https://www.emacswiki.org/emacs/SmoothScrolling Fixed https://www.emacswiki.org/emacs?action=browse;diff=2;id=SmoothScrolling;revision=56
gharchive/issue
2020-12-20T08:06:19
2025-04-01T06:44:32.324654
{ "authors": [ "io12" ], "repo": "io12/good-scroll.el", "url": "https://github.com/io12/good-scroll.el/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1357591469
[QA] Check / limit configurable timer values All configurable timer values should be checked / limited to 32 bit or less as setTimeout cannot handle timeouts largen then 32bit (signed). See Telegram posting: @apollon: Mal so aus gegebenem Anlass ein tip: prüft ihr Maximalwerte für setTimeout/setInterval?? Da gibts nen Maximalwert und wenn der überschritten wird ändert nodejs den auf 0 (aka sofort)!!! Siehe auch https://github.com/ioBroker/ioBroker.admin/issues/1664 https://stackoverflow.com/questions/3468607/why-does-settimeout-break-for-large-millisecond-delay-values Timers should be limitedin config json too fixed with upcoming v2.1.9 fixed with v2.1.9
gharchive/issue
2022-08-31T16:37:04
2025-04-01T06:44:32.381426
{ "authors": [ "mcm1957" ], "repo": "iobroker-community-adapters/ioBroker.snmp", "url": "https://github.com/iobroker-community-adapters/ioBroker.snmp/issues/156", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2411577781
Adapter stürzt ab Ich habe in regelmäßigen Abständen folgende Meldung: TypeError: Cannot read properties of undefined (reading 'split') at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) at AdapterClass.emit (node:events:519:28) at AdapterClass.emit (node:domain:488:12) at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:10949:34) at Immediate. (file:///opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:381:37) at processImmediate (node:internal/timers:478:21) Please fill the form provided when creating a new issue of type bug_: Describe the bug A clear and concise description of what the bug is. To Reproduce Steps to reproduce the behavior: Go to '...' Click on '...' Scroll down to '....' See error Expected behavior A clear and concise description of what you expected to happen. Screenshots & Logfiles If applicable, add screenshots and logfiles to help explain your problem. Versions: Adapter version: JS-Controller version: Node version: Operating system: Additional context Add any other context about the problem here. Ich habe ebenfalls die Fehlermeldung durch das Upgrade von 3.7.1 auf 3.8.1 ` telegram.1 | 2024-07-16 21:19:21.031 | info | starting. Version 3.7.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19 -- | -- | -- | -- telegram.0 | 2024-07-16 21:19:20.985 | info | starting. Version 3.7.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19 host.IoBroker | 2024-07-16 21:19:18.635 | info | iobroker exit 0 host.IoBroker | 2024-07-16 21:19:17.790 | info | instance system.adapter.telegram.1 started with pid 9710 host.IoBroker | 2024-07-16 21:19:17.701 | info | instance system.adapter.telegram.0 started with pid 9703 host.IoBroker | 2024-07-16 21:19:17.457 | info | "system.adapter.telegram.1" enabled host.IoBroker | 2024-07-16 21:19:17.320 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.1" is started host.IoBroker | 2024-07-16 21:19:17.294 | info | "system.adapter.telegram.0" enabled host.IoBroker | 2024-07-16 21:19:17.283 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.0" is started host.IoBroker | 2024-07-16 21:19:17.275 | info | iobroker upload [0] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/telegram.png telegram.png image/png host.IoBroker | 2024-07-16 21:19:17.269 | info | iobroker upload [1] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js.map rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.255 | info | iobroker upload [2] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.247 | info | iobroker upload [3] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js.map rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.235 | info | iobroker upload [4] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.230 | info | iobroker upload [5] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_react_index_js.6114a944.chunk.js.map rules/static/js/node_modules_react_index_js.6114a944.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.225 | info | iobroker upload [6] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_react_index_js.6114a944.chunk.js rules/static/js/node_modules_react_index_js.6114a944.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.220 | info | iobroker upload [7] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js.map rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.217 | info | iobroker upload [8] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.214 | info | iobroker upload [9] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js.map rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.212 | info | iobroker upload [10] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.209 | info | iobroker upload [11] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js.map rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.206 | info | iobroker upload [12] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.203 | info | iobroker upload [13] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js.map rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.200 | info | iobroker upload [14] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js application/javascript host.IoBroker | 2024-07-16 21:19:17.198 | info | iobroker upload [15] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/main.44f1f861.js.map rules/static/js/main.44f1f861.js.map application/json host.IoBroker | 2024-07-16 21:19:17.195 | info | iobroker upload [16] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/main.44f1f861.js rules/static/js/main.44f1f861.js application/javascript host.IoBroker | 2024-07-16 21:19:17.192 | info | iobroker upload [17] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/zh-cn.json rules/i18n/zh-cn.json application/json host.IoBroker | 2024-07-16 21:19:17.190 | info | iobroker upload [18] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/uk.json rules/i18n/uk.json application/json host.IoBroker | 2024-07-16 21:19:17.187 | info | iobroker upload [19] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/ru.json rules/i18n/ru.json application/json host.IoBroker | 2024-07-16 21:19:17.184 | info | iobroker upload [20] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/pt.json rules/i18n/pt.json application/json host.IoBroker | 2024-07-16 21:19:17.149 | info | iobroker upload [30] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/asset-manifest.json rules/asset-manifest.json application/json host.IoBroker | 2024-07-16 21:19:17.129 | info | iobroker upload [40] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/i18n/fr/translations.json i18n/fr/translations.json application/json host.IoBroker | 2024-07-16 21:19:17.107 | info | iobroker upload [50] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/custom/static/js/node_modules_prop-types_index_js.6da256b6.chunk.js.map custom/static/js/node_modules_prop-types_index_js.6da256b6.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.085 | info | iobroker upload [60] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/custom/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-de23731.0f3598f3.chunk.js.map custom/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-de23731.0f3598f3.chunk.js.map application/json host.IoBroker | 2024-07-16 21:19:17.063 | info | iobroker upload [70] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/custom/i18n/pl.json custom/i18n/pl.json application/json host.IoBroker | 2024-07-16 21:19:17.044 | info | iobroker upload [79] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/blockly.js blockly.js application/javascript host.IoBroker | 2024-07-16 21:19:16.565 | info | iobroker Update "system.adapter.telegram.1" host.IoBroker | 2024-07-16 21:19:16.526 | info | iobroker Update "system.adapter.telegram.0" host.IoBroker | 2024-07-16 21:19:16.439 | info | iobroker 332 packages are looking for funding run `npm fund` for details host.IoBroker | 2024-07-16 21:19:16.439 | info | iobroker changed 1 package in 5s host.IoBroker | 2024-07-16 21:19:12.243 | info | instance system.adapter.telegram.1 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION) host.IoBroker | 2024-07-16 21:19:12.117 | info | instance system.adapter.telegram.0 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION) host.IoBroker | 2024-07-16 21:19:12.096 | info | stopInstance system.adapter.telegram.1 killing pid 9569 host.IoBroker | 2024-07-16 21:19:12.033 | info | stopInstance system.adapter.telegram.0 killing pid 9586 telegram.1 | 2024-07-16 21:19:11.572 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason telegram.1 | 2024-07-16 21:19:11.572 | info | terminating telegram.0 | 2024-07-16 21:19:11.516 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason telegram.0 | 2024-07-16 21:19:11.516 | info | terminating host.IoBroker | 2024-07-16 21:19:11.166 | info | iobroker NPM version: 10.8.1Installing iobroker.telegram@3.7.1... (System call) host.IoBroker | 2024-07-16 21:19:11.092 | info | stopInstance system.adapter.telegram.1 send kill signal telegram.1 | 2024-07-16 21:19:11.071 | info | Got terminate signal TERMINATE_YOURSELF host.IoBroker | 2024-07-16 21:19:11.032 | info | stopInstance system.adapter.telegram.0 send kill signal host.IoBroker | 2024-07-16 21:19:11.031 | info | stopInstance system.adapter.telegram.1 (force=false, process=true) host.IoBroker | 2024-07-16 21:19:11.031 | info | "system.adapter.telegram.1" disabled telegram.0 | 2024-07-16 21:19:11.015 | info | Got terminate signal TERMINATE_YOURSELF host.IoBroker | 2024-07-16 21:19:11.011 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.1" is stopped. host.IoBroker | 2024-07-16 21:19:10.935 | info | stopInstance system.adapter.telegram.0 (force=false, process=true) host.IoBroker | 2024-07-16 21:19:10.934 | info | "system.adapter.telegram.0" disabled host.IoBroker | 2024-07-16 21:19:10.920 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.0" is stopped. host.IoBroker | 2024-07-16 21:19:10.892 | info | iobroker Update telegram from @3.8.1 to @3.7.1 host.IoBroker | 2024-07-16 21:19:09.261 | info | iobroker upgrade telegram@3.7.1 --debug telegram.0 | 2024-07-16 21:10:43.710 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19 host.IoBroker | 2024-07-16 21:10:41.376 | info | instance system.adapter.telegram.0 started with pid 9586 telegram.1 | 2024-07-16 21:10:14.636 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19 host.IoBroker | 2024-07-16 21:10:12.236 | info | instance system.adapter.telegram.1 started with pid 9569 host.IoBroker | 2024-07-16 21:10:11.279 | info | Restart adapter system.adapter.telegram.0 because enabled host.IoBroker | 2024-07-16 21:10:11.279 | error | instance system.adapter.telegram.0 terminated with code 6 (UNCAUGHT_EXCEPTION) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at processImmediate (node:internal/timers:476:21) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at AdapterClass.emit (node:domain:489:12) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at AdapterClass.emit (node:events:517:28) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: TypeError: Cannot read properties of undefined (reading 'split') host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[0]: This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason: telegram.0 | 2024-07-16 21:10:10.696 | warn | Terminated (UNCAUGHT_EXCEPTION): Without reason telegram.0 | 2024-07-16 21:10:10.695 | info | terminating telegram.0 | 2024-07-16 21:10:10.188 | error | Cannot read properties of undefined (reading 'split') telegram.0 | 2024-07-16 21:10:10.187 | error | TypeError: Cannot read properties of undefined (reading 'split') at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) at AdapterClass.emit (node:events:517:28) at AdapterClass.emit (node:domain:489:12) at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34) at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37) at processImmediate (node:internal/timers:476:21) host.IoBroker | 2024-07-16 21:10:10.129 | info | instance system.adapter.telegram.1 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION) telegram.0 | 2024-07-16 21:10:10.115 | error | unhandled promise rejection: Cannot read properties of undefined (reading 'split') telegram.0 | 2024-07-16 21:10:10.115 | error | Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). telegram.1 | 2024-07-16 21:10:09.554 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason telegram.1 | 2024-07-16 21:10:09.554 | info | terminating host.IoBroker | 2024-07-16 21:10:09.147 | info | stopInstance system.adapter.telegram.1 send kill signal telegram.1 | 2024-07-16 21:10:09.051 | info | Got terminate signal TERMINATE_YOURSELF host.IoBroker | 2024-07-16 21:10:09.027 | info | stopInstance system.adapter.telegram.1 (force=false, process=true) telegram.0 | 2024-07-16 21:08:58.618 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19 host.IoBroker | 2024-07-16 21:08:56.292 | info | instance system.adapter.telegram.0 started with pid 9546 host.IoBroker | 2024-07-16 21:08:54.211 | info | instance system.adapter.telegram.0 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION) telegram.0 | 2024-07-16 21:08:53.636 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason telegram.0 | 2024-07-16 21:08:53.635 | info | terminating host.IoBroker | 2024-07-16 21:08:53.185 | info | stopInstance system.adapter.telegram.0 send kill signal telegram.0 | 2024-07-16 21:08:53.134 | info | Got terminate signal TERMINATE_YOURSELF host.IoBroker | 2024-07-16 21:08:53.056 | info | stopInstance system.adapter.telegram.0 (force=false, process=true) javascript.0 | 2024-07-16 21:08:01.451 | info | script.js.Statusmeldungen.Status_Lichtzählen1: Text: Esstisch javascript.0 | 2024-07-16 21:06:42.869 | info | script.js.Statusmeldungen.Status_Lichtzählen1: Text: Esstisch ,Kueche telegram.0 | 2024-07-16 21:01:01.860 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19 host.IoBroker | 2024-07-16 21:00:59.292 | info | instance system.adapter.telegram.0 started with pid 9483 host.IoBroker | 2024-07-16 21:00:30.487 | info | instance system.adapter.mqtt.0 started with pid 9464 host.IoBroker | 2024-07-16 21:00:29.178 | info | Restart adapter system.adapter.telegram.0 because enabled host.IoBroker | 2024-07-16 21:00:29.178 | error | instance system.adapter.telegram.0 terminated with code 6 (UNCAUGHT_EXCEPTION) host.IoBroker | 2024-07-16 21:00:29.178 | error | Caught by controller[0]: at processImmediate (node:internal/timers:476:21) host.IoBroker | 2024-07-16 21:00:29.178 | error | Caught by controller[0]: at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37) host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34) host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at AdapterClass.emit (node:domain:489:12) host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at AdapterClass.emit (node:events:517:28) host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: TypeError: Cannot read properties of undefined (reading 'split') host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason: host.IoBroker | 2024-07-16 21:00:28.606 | info | iobroker exit 0 telegram.0 | 2024-07-16 21:00:28.600 | warn | Terminated (UNCAUGHT_EXCEPTION): Without reason telegram.0 | 2024-07-16 21:00:28.599 | info | terminating telegram.0 | 2024-07-16 21:00:28.014 | error | Cannot read properties of undefined (reading 'split') telegram.0 | 2024-07-16 21:00:28.014 | error | TypeError: Cannot read properties of undefined (reading 'split') at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) at AdapterClass.emit (node:events:517:28) at AdapterClass.emit (node:domain:489:12) at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34) at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37) at processImmediate (node:internal/timers:476:21) host.IoBroker | 2024-07-16 21:00:27.942 | info | instance system.adapter.mqtt.0 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION) telegram.0 | 2024-07-16 21:00:27.928 | error | unhandled promise rejection: Cannot read properties of undefined (reading 'split') telegram.0 | 2024-07-16 21:00:27.927 | error | Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). ` `CPU: 40 % RAM: 40 % Betriebszeit: 18h Verfügbar: 6.0.8 Installiert: 5.0.19 Ereignisse: ⇥45 / ↦13 Plattform: docker (official image - v9.1.2) Betriebssystem: linux Architektur: x64 CPUs: 16 Geschwindigkeit: 3695 MHz Modell: AMD Ryzen 9 6900HX with Radeon Graphics RAM: 30.6 GB System-Betriebszeit: 9 T. 10:00:11 Node.js: v18.20.4 time: 1721156190271 timeOffset: -120 NPM: 10.8.1 Adapter-Anzahl: 624 Datenträgergröße: 953.9 GB Freier Festplattenspeicher: 642.7 GB Aktive Instanzen: 45 Pfad: /opt/iobroker/ Betriebszeit: 17:54:47` Please retest eith newest release (3.8.2) If problem still exidzs please add complete version infos Adapter Admin Js-controller Node O/S Mit der neuen Version tritt es nicht mehr auf Sounds great that 3.8.2 solves the problem. I got the same problem with 3.8.1 and rolled back to 3.8.0 which workes fine for me. I use the beta repository and there is no 3.8.2. Where may I find this to test if it fixes the problem in my environment? Cheers, Thomas So I will close here. Please open a new issue if problem reoccures
gharchive/issue
2024-07-16T16:21:11
2025-04-01T06:44:32.395564
{ "authors": [ "crunchip77", "fuchs-1978", "mcm1957", "tmaul1968" ], "repo": "iobroker-community-adapters/ioBroker.telegram", "url": "https://github.com/iobroker-community-adapters/ioBroker.telegram/issues/841", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2462876699
Docs incorrect, scripts desctructive, missing options.... Hi guys, I am sorry but what is even going on with this plugin? docs ask us to npm install --save-dev @capacitor/assets but then a few lines later we are using npx so we just added it to our package.json it for no reason? Why does it always delete icons under src/assets/icons? I have never asked it to delete these and there is no option not to? Why be destructive? Just leave them there, they don't hurt nobody, or ask me if I want them deleted. These are in the damn assets folder, a folder that we might use within our app. Why is there no option to generate notification icons? We need a post-generate script that copies an icon to res/drawable because by itself, there is only a splash screen in there. Please guys, get your shit together and release something comprehensive, this is pretty bad. @mchl18 This allows for controlling the version of the tool. Which you should do. The other 2 are valid points, but your tone is not. excuse my tone, there was general frustration in working with capacitor that lead up to my tone, I will try and get a PR up to fix the things I mentioned Will npx not always use whatever it can get from the registry? Or will it prefer the package.json version?
gharchive/issue
2024-08-13T09:34:43
2025-04-01T06:44:32.451537
{ "authors": [ "marklagendijk", "mchl18" ], "repo": "ionic-team/capacitor-assets", "url": "https://github.com/ionic-team/capacitor-assets/issues/618", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
355323777
[4.0.0-beta.5] TypeScript compiler error when running ionic serve Bug Report Ionic Info Ionic: ionic (Ionic CLI) : 4.1.1 (/home/alex/.nvm/versions/node/v10.9.0/lib/node_modules/ionic) Ionic Framework : @ionic/angular 4.0.0-beta.5 @angular-devkit/core : 0.7.5 @angular-devkit/schematics : 0.7.5 @angular/cli : 6.1.5 @ionic/ng-toolkit : 1.0.7 @ionic/schematics-angular : 1.0.5 Cordova: cordova (Cordova CLI) : 8.0.0 Cordova Platforms : none Cordova Plugins : no whitelisted plugins (1 plugins total) System: NodeJS : v10.9.0 (/home/alex/.nvm/versions/node/v10.9.0/bin/node) npm : 6.2.0 OS : Linux 4.18 Describe the Bug Typescript compiler errors when running ionic serve. [ng] ℹ 「wdm」: Compiling... [ng] ERROR in node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,35): error TS1005: ';' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,56): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,60): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,72): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,35): error TS1005: ';' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,55): error TS1005: ';' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,82): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,86): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,104): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,116): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,147): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,151): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,169): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,181): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,193): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,230): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,234): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,252): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,264): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,276): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,288): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,325): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,329): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,360): error TS1005: ',' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,398): error TS1005: '(' expected. [ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,401): error TS1005: ')' expected. [ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,42): error TS1005: ';' expected. [ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,101): error TS1005: ']' expected. [ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,109): error TS1005: ',' expected. [ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,110): error TS1136: Property assignment expected. [ng] Date: 2018-08-29T20:30:04.848Z - Hash: a7760a7286c2a29edc22 - Time: 972ms [ng] 142 unchanged chunks [ng] ℹ 「wdm」: Compiled successfully. Steps to Reproduce My package.json: { "name": "nikkoApp", "version": "0.0.1", "author": "Ionic Framework", "homepage": "http://ionicframework.com/", "scripts": { "ng": "ng", "start": "ng serve", "build": "ng build", "test": "ng test", "lint": "ng lint", "e2e": "ng e2e" }, "private": true, "dependencies": { "@angular/common": "~6.1.1", "@angular/core": "~6.1.1", "@angular/forms": "~6.1.1", "@angular/http": "~6.1.1", "@angular/platform-browser": "~6.1.1", "@angular/platform-browser-dynamic": "~6.1.1", "@angular/router": "~6.1.1", "@ionic-native/core": "^5.0.0-beta.15", "@ionic-native/native-keyboard": "^5.0.0-beta.15", "@ionic-native/splash-screen": "5.0.0-beta.14", "@ionic-native/status-bar": "5.0.0-beta.14", "@ionic/angular": "^4.0.0-beta.5", "@ionic/pro": "2.0.3", "core-js": "^2.5.3", "rxjs": "6.2.2", "zone.js": "^0.8.26" }, "devDependencies": { "@angular/cli": "~6.1.1", "@angular/compiler": "~6.1.1", "@angular/compiler-cli": "~6.1.1", "@angular/language-service": "~6.1.1", "@angular-devkit/architect": "~0.7.2", "@angular-devkit/build-angular": "~0.7.2", "@angular-devkit/core": "~0.7.2", "@angular-devkit/schematics": "~0.7.2", "@ionic/ng-toolkit": "^1.0.0", "@ionic/schematics-angular": "^1.0.0", "@types/jasmine": "~2.8.6", "@types/jasminewd2": "~2.0.3", "@types/node": "~10.7.1", "codelyzer": "~4.4.2", "jasmine-core": "~2.99.1", "jasmine-spec-reporter": "~4.2.1", "karma": "~3.0.0", "karma-chrome-launcher": "~2.2.0", "karma-coverage-istanbul-reporter": "~2.0.0", "karma-jasmine": "~1.1.1", "karma-jasmine-html-reporter": "^0.2.2", "protractor": "~5.4.0", "ts-node": "~7.0.0", "tslint": "~5.11.0", "typescript": "2.9.2" }, "description": "An Ionic project", "cordova": { "plugins": { "cordova-plugin-native-keyboard": {} } } } Additional Context Probably related to #15360 To make it clear, bumping typescript version to 2.9.2 did not help. Nevermind, restarting ionic serve solved the problem. Apparently, it doesn't reload node modules automatically.
gharchive/issue
2018-08-29T20:38:15
2025-04-01T06:44:32.528244
{ "authors": [ "yursha" ], "repo": "ionic-team/ionic", "url": "https://github.com/ionic-team/ionic/issues/15379", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
437714118
bug: ion-datetime not bind as ISO-8601 format when using pickerOptions Bug Report Ionic version: [x] 4.x Current behavior: This bug is basically the same as ISSUE:15408, but my issue applies to when you use custom pickerOptions. I need to be able to clear an ion-datetime because it isn't required, so I would add this option by using pickerOptions. But when I use picker options, the date isn't in the format of ISO-8601 anymore, but rather the weird object form. Expected behavior: What I would expect is that when you use pickerOptions, the format would still be in the ISO-8601 format like when you are not using pickerOptions on a datetime, but this isn't the case for me. Steps to reproduce: ` My pickerOptions: public customStartDate: any = { buttons: [ { text: 'Clear', handler: () => this.eventForm.controls['startDate'].setValue(null) }, { text: 'Done', handler: (data: string) => { console.log("date data:", data); this.eventForm.controls['startDate'].setValue(data); } } ] } public customStartTime: any = { buttons: [ { text: 'Clear', handler: () => this.eventForm.controls['startTime'].setValue(null) }, { text: 'Done', handler: (data: Date) => { this.eventForm.controls['startTime'].setValue(data); } } ] }; ` Replacing the "this.eventForm.controls..." part with a console.log(data) would be enough te be able to see the issue when you somehow trigger it. Hi there, Thanks for the issue. To get around JS Date's implementation of timezones, we use our own object for storing the datetime data, which is what you are seeing. The value that is emitted from the component is valid ISO-8601, but we use that custom object internally. To get the proper ISO-8601 string, you will want to listen for the ionChange event. So you would end up doing something like: <ion-datetime (ionChange)="updateDate($event, formGroup)"></ion-datetime> Does this resolve your issue? Thanks! Hi there, Thanks for the issue. To get around JS Date's implementation of timezones, we use our own object for storing the datetime data, which is what you are seeing. The value that is emitted from the component is valid ISO-8601, but we use that custom object internally. To get the proper ISO-8601 string, you will want to listen for the ionChange event. So you would end up doing something like: <ion-datetime (ionChange)="updateDate($event, formGroup)"></ion-datetime> Does this resolve your issue? Thanks! I don't really see how (ionChange)="updateDate($event, formGroup)" would be able to help me. What is the difference between using (ionChange) and the handler inside the pickerOption? I also don't know what this updateDate($event, formGroup)" needs to contain to be able to work properly. In my opinion these are things that all should be explained inside the documentation of ionic (4). The documentation contains UI component examples, but it lacks the information on how to implement it properly or how to customize it (not only for ion-datetime). But that's somewhat another topic... Hi there, ionChange is the event emitted from the datetime component when the value of the date has been changed (i.e. after the date has been selected and "Done" has been clicked). The result of that is a valid ISO-8601 string. This can be found in the Events section of the ion-datetime docs. The pickerOptions property is something that is passed through to the ion-picker component, which is what ion-datetime is built on. The property allows you to customize the underlying picker component. Details on this can be found in the Properties section of the ion-picker docs. All the updateDate function needs to contain is the code you originally has in the Done button handler. In this case $event is an object containing the ISO-8601 date that was selected. Hopefully that clears things up a bit 🙂 . I am going to close this as it is not an Ionic issue. For further details on how to implement or customize the datetime and picker components, please see the ion-datetime Docs and the ion-picker Docs. For further assistance implementing your code, feel free to post on the forums or on our slack. Thanks! Thank you for your answer. $event indeed contains a valid ISO-8601 object when you work with f.e. picker-format="DD MMMM YYYY". But it doesn't work when you want the time only f.e. picker-format="HH mm ss" (or at least it does not for me): With pickerOptions: Without pickerOptions: Is this really not a bug? Thanks in advance! Hi there, Can you post your code? It's hard to say just seeing the output. HTML with pickerOptions: <ion-item> <ion-label for="startTime" position="stacked">Starttijd</ion-label> <ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)" [pickerOptions]="customStartTime"> </ion-datetime> </ion-item> HTML without pickerOptions: <ion-item> <ion-label for="startTime" position="stacked">Starttijd</ion-label> <ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)"> </ion-datetime> </ion-item> TS: public customStartTime: any = { buttons: [ { text: 'Clear', handler: () => this.eventForm.controls['startTime'].setValue(null) }, { text: 'Bewaar', handler: (data: string) => { this.eventForm.controls['startTime'].setValue(data); } } ] }; eventForm = new FormGroup({ id: new FormControl('', []), eventTypeId: new FormControl('', [Validators.required]), name: new FormControl('', [Validators.required]), description: new FormControl('', [Validators.required]), startDate: new FormControl('', [Validators.required]), startTime: new FormControl({ value: '', disabled: false }), endDate: new FormControl('', [Validators.required]), endTime: new FormControl({ value: '', disabled: false }), isFullDay: new FormControl(false), location: new FormControl('') }); updateStartTime($event, eventForm) { console.log($event); console.log(eventForm); } That is all the code that is relevant. And you can see the results of $event both with and without pickerOptions above. If you need more, just ask me 😃 . Thanks! HTML with pickerOptions: <ion-item> <ion-label for="startTime" position="stacked">Starttijd</ion-label> <ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)" [pickerOptions]="customStartTime"> </ion-datetime> </ion-item> HTML without pickerOptions: <ion-item> <ion-label for="startTime" position="stacked">Starttijd</ion-label> <ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)"> </ion-datetime> </ion-item> TS: public customStartTime: any = { buttons: [ { text: 'Clear', handler: () => this.eventForm.controls['startTime'].setValue(null) }, { text: 'Bewaar', handler: (data: string) => { this.eventForm.controls['startTime'].setValue(data); } } ] }; eventForm = new FormGroup({ id: new FormControl('', []), eventTypeId: new FormControl('', [Validators.required]), name: new FormControl('', [Validators.required]), description: new FormControl('', [Validators.required]), startDate: new FormControl('', [Validators.required]), startTime: new FormControl({ value: '', disabled: false }), endDate: new FormControl('', [Validators.required]), endTime: new FormControl({ value: '', disabled: false }), isFullDay: new FormControl(false), location: new FormControl('') }); updateStartTime($event, eventForm) { console.log($event); console.log(eventForm); } That is all the code that is relevant. And you can see the results of $event both with and without pickerOptions above. If you need more, just ask me 😃 . Thanks! Hi there, The reason why this is happening is that data in button handler is not a string. As I mentioned before, that is the custom object we use internally to parse dates. You are setting the value of the datetime to this object, and so the datetime component is just going to emit that object back. The value property expects a valid ISO-8601 string, as stated in the ion-datetime documentation. ionChange is the event emitted from the datetime component when the value of the date has been changed (i.e. after the date has been selected and "Done" has been clicked). The result of that is a valid ISO-8601 string. This can be found in the Events section of the ion-datetime docs. The reason why this is happening is that data in the button handler is not a string. As I mentioned before, that is the custom object we use internally to parse dates. You are setting the value of the datetime to this object, and so the datetime component is just going to emit that object back. The value property expects a valid ISO-8601 string, as stated in the ion-datetime documentation. This is creating the problem, ionChange doesn't trigger when you don't update the value of the datetime. So I'm not able to get $event when I don't somehow change something of the datetime object... I am completely lost. For a date (year, month, day) it works perfectly, but for time (hour, minute, second) it doesn't when I do exactly the same for both except for the picker-format. And as stated in the documentation, HH:mm:ss is a valid ISO-8601 format. I don't know what else I can say or try to do, would creating a small public GitHub project be interesting for you? Hi there, Do you have an example where the date formatting is working properly? I have the time example, and that has the behavior you described, but the date formatting is also returning the custom object. Thanks! I'm wrong. I'm so sorry. Date has the same behavior as time. Sorry. But still, how would I be able to trigger ionChange without using the handler? How am I able to get an ISO-8601 without setting the value to the custom object? Hi there, When you select a date in the datetime component and then click "Done", ionChange is triggered. If you want to use a custom handler, you will need to parse the date object yourself. Ideally, there would be a "clear" option built into ion-datetime. There is a feature request open for that functionality: https://github.com/ionic-team/ionic/issues/17482. Project where everything can be seen: https://github.com/MatthiasVanMeerbeek-r0670805/Ionic4IonDatetime (Don't know how to properly link a project to an issue.)
gharchive/issue
2019-04-26T14:58:29
2025-04-01T06:44:32.555172
{ "authors": [ "MatthiasVanMeerbeek-r0670805", "liamdebeasi" ], "repo": "ionic-team/ionic", "url": "https://github.com/ionic-team/ionic/issues/18142", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
179801616
nav-decor always being displayed Short description of the problem: The nav-decor is always being displayed with a black background on iOS What behavior are you expecting? That the nav-decor doesn't have a black background This can be re produced using the ionic starter Which Ionic Version? 2 @alan-agius4 @jgw96 sorry, I don't understand what's the problem Justin is aware of the problem On Saturday, 1 October 2016, Manu Mtz.-Almeida notifications@github.com wrote: @alan-agius4 https://github.com/alan-agius4 @jgw96 https://github.com/jgw96 sorry, I don't understand what's the problem — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/driftyco/ionic/issues/8252#issuecomment-250862383, or mute the thread https://github.com/notifications/unsubscribe-auth/AQv-Wn8fHWO7N64EGRmwFBHcLODD1Jf5ks5qvYaAgaJpZM4KI8kt . @jgw96 is fully aware of the problem as I sent him some details over Slack. Hello just to document the issue a little. The nav-decor component seems to always be displayed on iOS. This causes weird thing such as a small black "block" appearing when navigating on iOS sometimes. nav-decor always being displayed, hides GoogleMaps despite the css: ion-app._gmaps_cdv_ .nav-decor{ background: transparent !important; } The css doesn't seem to get applied, but unchecking and checking any of the style property from the safari inspector applies it and the maps are displayed. I faced this issue on the iOS emulator but the code behaves as expected on the device. Can you provide a repo that reproduces this issue? @manucorporat Link to the repo: https://github.com/nakulgulati/map-test Also attaching a gif which demonstrates what I said in the previous comment for clarity. :) Found my issue I had the below ``` html <ion-content [attr.class]="options?.cssClass"> rather than ``` html <ion-content [ngClass]="options?.cssClass"> which resulted in content-ios class not being applied @nakulgulati Unchecking and checking the style property in the web inspector, will cause a re-paint which is most likely why it's working. @alan-agius4 Yep that is exactly why it is working. The gif shows that the styles are loaded but not applied on nav-decor and hence causes it to bug. I tested the map-test: I works for me... ion-nav is and it has to be black, the whole time. What behavior are you expecting? That the nav-decor doesn't have a black background This can be re produced using the ionic starter ^this premise is wrong. nav-decor is using during transitions to have a fade to black effect. I am going to close the issue and please create a new one following the issue template, explaining WHAT is failing in your app, nothing else. Since this is issue is being hard to reproduce please provide a repo and the environment: Desktop? device? ios only? WKWebview? iOS Simulator? @manucorporat Seems fair to close the issue. I was originally facing this issue when using Google Maps in a project and had spun up the test project then. Now the issue seems to have disappeared from the original project and still there in map-test for me. I would look into in more and like you suggested, would open a new issue if required. Thanks. :) This is the author of the cordova-googlemaps-plugin. As of the google maps plugin 2.0.0-beta2-20170705-1715, the plugin inserts the css automatically. ._gmaps_cdv_ .nav-decor { display: none !important; background-color: rgba(0,0,0,0) !important; background: rgba(0,0,0,0) !important; } This thread helps me to solve this issue. Thank you for discussing.
gharchive/issue
2016-09-28T15:10:38
2025-04-01T06:44:32.569231
{ "authors": [ "alan-agius4", "jgw96", "manucorporat", "nakulgulati", "wf9a5m75" ], "repo": "ionic-team/ionic", "url": "https://github.com/ionic-team/ionic/issues/8252", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
772979933
dotnet paket install breaks project parsing After dotnet paket install project system returns project data contains no references at all. It's fixed by manually running dotnet restore after dotnet paket install has finished. We should somehow detect that dotnet paket install/update were run and fire dotnet restore automatically. CC: @forki paket install / update we could touch a file in /obj - but I think it actually already deletes the project.assets.json Shouldn't this be enough? I suspect there's something weird going on with parsing project files for which project.assets.json doesn't exist. I think it's some kind of regression I've introduced when rewriting the project system, not the Paket issue. Do we already have a repro? Krzysztof Cieślak notifications@github.com schrieb am Do., 11. Feb. 2021, 15:36: I suspect there's something weird going on with parsing project files for which project.assets.json doesn't exist. I think it's some kind of regression I've introduced when rewriting the project system, not the Paket issue. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/ionide/ionide-vscode-fsharp/issues/1459#issuecomment-777505196, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAOANDTGAS5JJDAKFT3QHLS6PTNTANCNFSM4VFWEOGA .
gharchive/issue
2020-12-22T14:16:46
2025-04-01T06:44:32.587865
{ "authors": [ "Krzysztof-Cieslak", "forki" ], "repo": "ionide/ionide-vscode-fsharp", "url": "https://github.com/ionide/ionide-vscode-fsharp/issues/1459", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
155134379
Convert from FunScript to Fable Here comes my first attempt to convert the awesome Ionide plugins to Fable! I've used your vscode-fable-demo as a basis. Basically the changes are: For simplicity I've converted the project to a single .fsx file and removed the now unneeded files. Also for simplicity, I just copied the Helper modules into the script. I'll update the ionide-vscode-helpers later. I had to update the Fable Node definitions to fix some mistakes. I'm using the full name to reference a couple of modules that were now conflicting (like Promise). I've added another helper function (within the activate method) to register commands. On a quick test, it seems to work :+1: Can you please have a look and tell me what you think? If you want me to make some changes, just tell me. Cheers! That looks really good! Only thing that bothers me is that you completely removed FAKE build scripts - I think I would want to leave parts responsible for versioning, packaging and publishing as they were. Probably whole script should be almost the same, only difference would be calling npm run build instead of old generation. You're right. I was just trying to find the minimal structure to build the extension with Fable but it seems I went too far ;) I'll put the FAKE script back :+1: Done! Please check if the changes I did in the build.fsx script make sense. Cheers! Before we go with it, I would like to convert Helpers project to Fable so we don't have Helpers module here. I guess we can just create Fable branch on Helpers repo and put this code there for start. Perfect! Please create the branch so I can send the PR there 👍 Also, how do you want to interact with the Helpers from the other packages? Just compiling the code into the final JS or do you want to make a npm package for the Helpers? This could prevent some duplicity of code but for that we will probably have to use F# projects instead of scripts. @alfonsogarciacaro can you pr onto https://github.com/ionide/ionide-vscode-paket/tree/fable instead? And Helpers part to https://github.com/ionide/ionide-vscode-helpers/tree/fable. :) I merged the PR onto this branch https://github.com/ionide/ionide-vscode-paket/tree/fable
gharchive/pull-request
2016-05-16T22:13:04
2025-04-01T06:44:32.594663
{ "authors": [ "Krzysztof-Cieslak", "alfonsogarciacaro", "cloudRoutine" ], "repo": "ionide/ionide-vscode-paket", "url": "https://github.com/ionide/ionide-vscode-paket/pull/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
992908805
Add support for Firefox for Android For now, this extension cannot be installed to the stable version of Firefox for Android. Currently, only a limited number of extensions in our Recommended program are supported on the new Firefox for Android (codename: Fenix) release channel. https://blog.mozilla.org/addons/2020/02/11/faq-for-extension-support-in-new-firefox-for-android/ Advanced users can install this extension to Firefox for Android Nightly according to: https://blog.mozilla.org/addons/2020/09/29/expanded-extension-support-in-firefox-for-android-nightly/ However, "Other search engines" and "Subscription" are not currently available in Firefox for Android Nightly because browser.permissions.request, which is the API to dynamically request permissions, is not yet implemented in GeckoView: mozilla-mobile/fenix#16912 https://bugzilla.mozilla.org/show_bug.cgi?id=1601420 Exceptionally, subscriptions to blocklists hosted on *.githubusercontent.com may be available because permission request for this origin is omitted. re #110 Exceptionally, subscriptions to blocklists hosted on *.githubusercontent.com may be available because permission request for this origin is omitted. re #110 Thanks a lot for this workaround! Would you accept more such domains? I'd like to subscribe to lists on GitLab. Although GitLab itself does not use Access-Control-Allow-Origin: *, some third-party CDN services do: GitHack: using domain glcdn.githack.com for files on GitLab Statically: using domain cdn.statically.io @yan12125 Thank you for your suggestion. You might want to create a pull request. The code to change: https://github.com/iorate/uBlacklist/blob/bbc2d22a24b04cf04ac4caced1d470876da355d8/src/scripts/options/subscription-section.tsx#L53 Thanks! Let me try it out. Sync is not available in Firefox for Android (Fenix) now because browser.identity.launchWebAuthFlow is not implemented. Recently I added support for sync in Safari. Because Safari also lacks browser.identity.launchWebAuthFlow, I implemented an alternative to it. Is that alternative available in Fenix? Currently, no. It requires browser.permissions.request, which is not yet implemented in Fenix. I think this issue should remain open even without activity? I added a label that is exempt from stale According to https://bugzilla.mozilla.org/show_bug.cgi?id=1601420 browser.permissions.request is implemented. I tried uBlacklist 7.6.0 on Firefox Nightly 103, but no permission prompt showed. Maybe we need to wait for an update of Fenix. https://github.com/mozilla-mobile/fenix/issues/16912 any update more? Please support kiwi browser too Is there an update for this? I can't activate UBlacklist on Android from the Firefox Add-On Store in Firefox / Firefox Nightly, nor from the Chrome Web Store in Kiwi. Is there a workaround to use UBlacklist on Android? I can't activate UBlacklist on Android from the Firefox Add-On Store in Firefox / Firefox Nightly You need to use Firefox Beta or Firefox Nightly, and follow these instructions: https://blog.mozilla.org/addons/2020/09/29/expanded-extension-support-in-firefox-for-android-nightly/ However, it only works in the news tab as long as the checkmark for desktop view is checked in Firefox. Works for me on Firefox Beta for Android. (I'm also using Google Search Fixer to bypass the browser discrimination.) I have edited my previous post. The problem occurs only in the news tab. Here it works only in desktop view mode. uBlackList V8.3.1 is working great for Google Search with Firefox Beta 117.0b4. It is not working for other search engines for me. In Options, Other Search Engines, the Enable button does not work if any other search engine is checked. but no permission prompt showed Looks like it should be released in Firefox 120: https://github.com/mozilla-mobile/firefox-android/pull/3917 It works now! https://www.mozilla.org/en-US/firefox/android/120.0/releasenotes/ I was able to install the extension directly from the firefox addon page. Although the mobile search is still limited to officially endorsed addons, accessing the page directly allows to click and install it. Not sure if the rollout is the same for everyone though. I'm closing this issue because we can now install this add-on to Firefox for Android 120.0 (or later). For further problems with Firefox for Android, please create a separate issue for each problem.
gharchive/issue
2021-09-10T05:45:16
2025-04-01T06:44:32.620418
{ "authors": [ "3xploiton3", "Eagle-85", "NotaInutilis", "Semro", "amroncz", "iorate", "item4", "teohhanhui", "yan12125" ], "repo": "iorate/ublacklist", "url": "https://github.com/iorate/ublacklist/issues/117", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1082906606
ClientError: {} Issue description Above error occurs when transferring IOTA to Firefly from an exchange. The transfer has worked previously with different addresses. Firefly version v1.2.0 OS details macOS Monterey v12.0.1 Can the issue be reliably reproduced? No Steps to reproduce the issue Did not reproduce the error as previous IOTA transferred are still unaccounted for. ... Expected behaviour That the IOTA transfer from the exchange to Firefly is successful and auditable, like the 5-6 previous transfers. Actual behaviour The message from the exchange is: Error Message: Withdrawals are currently disabled for this address due to wallet maintenance. Errors IOTA Firefly Error file - 17Dec21.docx The transfer failed on the following address: iota1qzfy3cmlh5umjej2xstffu0mfl6ntvmgn22hhgavxkuwts9p65veuw4g8qk If the exchange disabled withdrawals then this has nothing to do with Firefly. The funds were not yet transferred to your address (probably due to maintenance on their own wallet) as you can see here. The empty ClientError will be fixed in the next release https://github.com/iotaledger/firefly/issues/1433, but as @r-c-k wrote your problem is only on exchange side
gharchive/issue
2021-12-17T05:39:29
2025-04-01T06:44:32.626894
{ "authors": [ "Thoralf-M", "clareiota", "r-c-k" ], "repo": "iotaledger/firefly", "url": "https://github.com/iotaledger/firefly/issues/1772", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2162062026
Add constructor for VerificationMethod in TS Fixes https://github.com/iotaledger/identity.rs/issues/1320 @UMR1352 can you please take a look at this?
gharchive/pull-request
2024-02-29T20:37:25
2025-04-01T06:44:32.628651
{ "authors": [ "abdulmth" ], "repo": "iotaledger/identity.rs", "url": "https://github.com/iotaledger/identity.rs/pull/1321", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1246274391
UI: Reset svelte state when logging out if a user logs out the svelte state shall be resetted otherwise for instance the selected detail view of an identity is still shown to the new logged in user also search is still applied when logging out and in again maybe also contact begona about a good concept to reset state
gharchive/issue
2022-05-24T09:37:42
2025-04-01T06:44:32.635216
{ "authors": [ "dominic22" ], "repo": "iotaledger/is-ui-components", "url": "https://github.com/iotaledger/is-ui-components/issues/33", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2327127612
🛑 备用KMS激活服务器 01 is down In 52a8233, 备用KMS激活服务器 01 (s11.ikms.eu.org) was down: HTTP code: 0 Response time: 0 ms Resolved: 备用KMS激活服务器 01 is back up in 5f99680 after 10 minutes.
gharchive/issue
2024-05-31T07:16:38
2025-04-01T06:44:32.637776
{ "authors": [ "iougemini" ], "repo": "iougemini/ikms-uptime", "url": "https://github.com/iougemini/ikms-uptime/issues/1955", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2738542939
🛑 备用KMS激活服务器 01 is down In 1178e1a, 备用KMS激活服务器 01 (s11.ikms.eu.org) was down: HTTP code: 0 Response time: 0 ms Resolved: 备用KMS激活服务器 01 is back up in aac1a7c after 11 minutes.
gharchive/issue
2024-12-13T14:36:12
2025-04-01T06:44:32.640298
{ "authors": [ "iougemini" ], "repo": "iougemini/ikms-uptime", "url": "https://github.com/iougemini/ikms-uptime/issues/3817", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
733152776
How to get face embedding / encoding? Congrats for this nice project @ipazc! I see the output of your algorithm is something like: [ { 'box': [277, 90, 48, 63], 'keypoints': { 'nose': (303, 131), 'mouth_right': (313, 141), 'right_eye': (314, 114), 'left_eye': (291, 117), 'mouth_left': (296, 143) }, 'confidence': 0.99851983785629272 } ] i.e. it gives the bounding box, the keypoints (nose, mouth, eye, etc.). But how to get an face embedding / face encoding, to be able to do face identification? I may have misunderstood what you are asking, but this project isn't about getting face encodings for recognition purposes. For that, you will need something else, such as https://github.com/ageitgey/face_recognition. @MattyB95 Maybe I misunderstood this project, but I thought MTCNN + FaceNet not only allows face detection, but also recognition / identification, by assigning a 128D vector embedding to each face. https://github.com/davidsandberg/facenet Is it correct @ipazc? @josephernest I'm not completely familiar with FaceNet but wouldn't that be providing the face recognition/identification encodings. This project would be more about cropping the image to the facial region for that purpose, but I will let @ipazc give his verdict :) MTCNN is only used to detect faces in an image, The value in the the "box" are the corners for the face box in an image To get the face encoding you will need to pass the the pixels resulted from cropping the image to a face corners to one of the models that are used for face encoding, as facenet, deepface,.... @imnimn Is there a Python implementation that packs: MTCNN for cropping the face box + facenet or deepface for the encoding? @imnimn Is there a Python implementation that packs together: MTCNN to get the face box + facenet or deepface for the encoding? Please check this article: https://arsfutura.com/magazine/face-recognition-with-facenet-and-mtcnn/, It is implementation is in this repository https://github.com/arsfutura/face-recognition
gharchive/issue
2020-10-30T13:01:29
2025-04-01T06:44:32.661562
{ "authors": [ "MattyB95", "imnimn", "josephernest" ], "repo": "ipazc/mtcnn", "url": "https://github.com/ipazc/mtcnn/issues/102", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1627714379
Where is the changelog? I'm doing some maintenance work to upgrade dependencies. Right now I'm at version 3.4.3. I see the latest version is 4.0.7. When doing updates, I like to go through a changelog to see if there is any breaking changes I should keep a lookout for. I seem to be unable to find the changelog for this library. I assume going from 3 to 4 is potentially a big upgrade. Can you point me to where the changelog is to help with this update? Hi @lightswitch05 this is great idea and am sorry we did not already include one, we will do so going forward.
gharchive/issue
2023-03-16T15:10:39
2025-04-01T06:44:32.663203
{ "authors": [ "jonathan-kosgei", "lightswitch05" ], "repo": "ipdata/python", "url": "https://github.com/ipdata/python/issues/24", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1114180403
p4c-xdp: Build and install p4c-xdp compiler This loads the p4c-xdp compiler extension, builds it, and instals it as a part of the host native install. Useful for experimenting with P4 to XDP. Signed-off-by: Kyle Mestery mestery@mestery.com Closing for now.
gharchive/pull-request
2022-01-25T17:38:34
2025-04-01T06:44:32.664900
{ "authors": [ "mestery" ], "repo": "ipdk-io/ipdk", "url": "https://github.com/ipdk-io/ipdk/pull/68", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2599432266
Introduce ExternManager class architecture This CL implements an architecture that can be used to isolate code that depends on the ES2K-specific IDPF extensions to P4Runtime from the rest of the code. As things stand right now, we cannot upstream our P4InfoManager changes because they break non-IPDK Stratum platforms (barefoot, bcm, fmv2, etc.) unless they are built against our custom P4Runtime. The new architecture adds an interface that allows platform-specific extensions to P4InfoManager to be incorporated at build time. The architecture also provides a mechanism that supports custom extensions to TdiTableManager to be incorporated at build time. The IDPF-dependent code goes into the extensions, which are compiled and linked in as part of the ES2K build. This CL adds the Es2kExternManager class and its relatives, together with a unit test and Bazel updates to support the new components. It does NOT include the P4InfoManager and TdiTableManager changes that remove the existing IDPF support and replace it with the ExternManager code. The new files will be passive additions to the codebase until this is done (in a subsequent commit). See issue https://github.com/ipdk-io/stratum-dev/issues/274 for more information. It includes a UML class diagram. Development notes These changes were developed in Draft PR https://github.com/ipdk-io/stratum-dev/pull/272. The CL consists of 37 files, which is a lot to review. It's also downright scary. To mitigate risk, I reduced the number of changes to TdiTableManager and wrote an extensive unit test for Es2kExternManager. The latter turned out to be a worthwhile effort: it found several bugs, and the feedback led to a number of revisions to the code being tested. (Error handling has been completely overhauled.) The unit test covers 100% of the UUT. I also decided to submit the new code first, including the unit test. Fewer files, very little overlap, and 100% safe because the only thing that compiles or links the new code is the unit test. I copied the new files to a fresh branch, made a few additional edits, and created this PR. Code review notes tdi_table_handler.cc and tdi_table_handlers.h were extracted from tdi_table_manager.cc. I needed to share the functions with the resource handlers, which were also lifted from the TableManager. The originals will be deleted as part of integrating the old and new code. When reviewing the ES2K resource handlers, you can compare the code with the originals in TdiTableManager. It should be essentially the same. The ResourceHandler method is named for the TableManager method from which it was lifted, with the addition of a Do prefix to distinguish between the TableManager and ResourceManager methods. Closed in favor of PR https://github.com/ipdk-io/stratum-dev/pull/276, which implements Stage 1 of the design.
gharchive/pull-request
2024-10-19T17:32:24
2025-04-01T06:44:32.671040
{ "authors": [ "ffoulkes" ], "repo": "ipdk-io/stratum-dev", "url": "https://github.com/ipdk-io/stratum-dev/pull/273", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1454285365
Allow passing arbitrary fetchurl args per registry Motivation Same as #162 but with my own updates pushed to it Closes #159 Checklist [ ] added tests to verify new behavior [ ] added an example template or updated an existing one [ ] updated docs/API.md with changes [ ] updated CHANGELOG.md Note: still need to update the docs and changelog before merging
gharchive/pull-request
2022-11-18T01:30:20
2025-04-01T06:44:32.677630
{ "authors": [ "ipetkov" ], "repo": "ipetkov/crane", "url": "https://github.com/ipetkov/crane/pull/168", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1297190148
Add peergos talk @b5 thanks @ianopolous! merged this content with #104
gharchive/pull-request
2022-07-07T10:49:19
2025-04-01T06:44:32.683243
{ "authors": [ "b5", "ianopolous" ], "repo": "ipfs-shipyard/ipfs-thing-2022", "url": "https://github.com/ipfs-shipyard/ipfs-thing-2022/pull/103", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
119272908
D3 Tree Viz "help wanted" issues @Dignifiedquire I did a bit of thinking and cleanup on this repo after pinging you in #8. I also wrote a readme for this repo, so the whole D3 tree thing should be a bit more scrutable now. Things I would love your help with if you have bandwidth and interest: #15 d3 tree should use js-ipfs-api #16 d3 tree should not load all nodes recursively; should initially load two levels #14 d3 tree should expand downward to fit all nodes #9 d3tree: support paths @RichardLitt perhaps you'd be interested in helping out with #8: add d3 tree dataviz to "examples" on ipfs.io I labeled all of these "help wanted" and have been organizing them in waffle: https://waffle.io/ipfs/dataviz @harlantwood Thanks for doing this! This is pretty awesome. I can help with examples, sure! Awesome, thanks @RichardLitt! BTW I am open to this simple example evolving technically: from CoffeeScript to es2015 from make to webpack/browserify other directions that inspire potential contributors :smile:
gharchive/issue
2015-11-28T05:47:44
2025-04-01T06:44:32.687584
{ "authors": [ "RichardLitt", "harlantwood" ], "repo": "ipfs/dataviz", "url": "https://github.com/ipfs/dataviz/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
241986873
All of POST request's results are 403 Forbidden error: ipfs gateway config: config ~/.ipfs/config "API": { "HTTPHeaders": { "Access-Control-Allow-Methods": [ "PUT", "GET", "POST" ], "Access-Control-Allow-Origin": [ "*" ] } }, @shanelau I try to fix the ~/.ipfs/config as the same you, but it still go into the Web UI of IPFS when i run http://local:5001/ipfs/(PeerId)/#. It show me the 403 error. Same problem ;( I tried to copy @shanelau's snippet but it seems like this might be a bug with go-ipfs v0.5.0-rc1. This Command Works for me ipfs config --json API.HTTPHeaders.Access-Control-Allow-Origin '["'"$origin"'", "http://127.0.0.1:8080","http://localhost:3000"]' you can allow the request from multiple origins
gharchive/issue
2017-07-11T09:40:47
2025-04-01T06:44:33.316154
{ "authors": [ "NatoBoram", "aliawais0007", "hleb-albau", "shanelau", "vutuyetmai130398" ], "repo": "ipfs/ipfs-webui", "url": "https://github.com/ipfs/ipfs-webui/issues/596", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
404962606
Basic auth support For people using a private IPFS cluster with a swarm key, it is also useful to have support for HTTP basic auth. Interaction with the gateway is easy, as there is no library and you can just use URLs like https://user:pass@gatewayurl.com/ipfs/hash. However, for using the IPFS API, this is not as simple. Making py-ipfs-api support basic auth and is a useful and simple change that I'm willing to implement. Instead of vendoring this code I would love to submit a PR. Just want to ask before doing this - would you accept a PR for basic auth support? +1 on this. This worked with ipfsapi but the new mutliaddr scheme doesn't support it and I'm trying to figure out how to connect. @brianmcmichael could you try https://github.com/ldub/py-ipfs-api/ ? I'm having an issue testing it against my own IPFS node because of some SSL issue, but you should be able to do the following as mentioned in the README: >>> import base64 >>> import ipfshttpclient >>> creds = base64.b64encode(b"username:password").decode("utf-8") >>> headers = {"Authorization" : "Basic " + creds} >>> api = ipfshttpclient.connect('/dns/ipfs-api.example.com/tcp/443/https', headers=headers) The TLS issue should be fixed (see #185), I'm still open to adding support for dedicated username and password parameters to the codebase btw. :wink:
gharchive/issue
2019-01-30T21:01:18
2025-04-01T06:44:33.321632
{ "authors": [ "alexander255", "brianmcmichael", "ldub" ], "repo": "ipfs/py-ipfs-api", "url": "https://github.com/ipfs/py-ipfs-api/issues/154", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }