id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1718099066 | 🛑 CHAT Website is down
In 4777ea1, CHAT Website (https://chat.devf5r.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: CHAT Website is back up in 3509cad.
| gharchive/issue | 2023-05-20T08:27:30 | 2025-04-01T06:44:31.208382 | {
"authors": [
"info-devf5r"
],
"repo": "info-devf5r/VPN",
"url": "https://github.com/info-devf5r/VPN/issues/196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
453632616 | Add rid template button for md editor
This issue in ermrestJS adds support for an rid template in markdown. A button should be added to our markdown editors that inputs [[<rid>]] into the form.
The corresponding changes for this issue have been merged into master. A button was added to the editor input. Information was also added to the mdHelp page explaining how this new template can be used. Reassigning to @hongsudt.
| gharchive/issue | 2019-06-07T18:07:25 | 2025-04-01T06:44:31.243687 | {
"authors": [
"jrchudy"
],
"repo": "informatics-isi-edu/chaise",
"url": "https://github.com/informatics-isi-edu/chaise/issues/1788",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
174637731 | Reference.read default page size
This PR, adds default value for page size in read function.
If the limit is not defined, it will look for page_size in Table Display annotation, using the current context. Otherwise, it will use "1" as a default page size.
Related Issue: #177
These changes are not relevant anymore.
We decided to not change the default value explicitly, and just expose the default page size as reference.display.default_page_size .
| gharchive/pull-request | 2016-09-01T22:08:45 | 2025-04-01T06:44:31.245274 | {
"authors": [
"RFSH"
],
"repo": "informatics-isi-edu/ermrestjs",
"url": "https://github.com/informatics-isi-edu/ermrestjs/pull/180",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
99971242 | --mount-root option broken in v >= 0.2.45
The --mount-root option no longer works for me in version 0.2.45 and up. If I downgrade to 0.2.44 then the following command runs fine:
$ forklift 273695c4789e sshd --mount-root ~/docker/fs/
mount: /dev/mapper/docker-252:0-4468428-0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf already mounted or /var/lib/docker/devicemapper/mnt/0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf/ busy
Traceback (most recent call last):
File "/usr/local/bin/forklift", line 9, in <module>
load_entry_point('docker-forklift==0.2.45', 'console_scripts', 'forklift')()
File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 338, in main
return Forklift(sys.argv).main()
File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 319, in main
return driver.run(*command)
File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 240, in run
return self.run_sshd()
File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 446, in run_sshd
self.mount_root(container)
File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 365, in mount_root
container_path])
File "/usr/lib/python3.4/subprocess.py", line 557, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['sudo', 'mount', '-o', 'context=""', '/dev/mapper/docker-252:0-4468428-0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf', '/var/lib/docker/devicemapper/mnt/0de707309f72cd745fd4340ad11852346ee7021cfc585de83ff97bffec093baf/']' returned non-zero exit status 32
Ubuntu 14.04.3 LTS
Docker version 1.7.1, build 786b29d
Forklift Version: 0.2.45
Hmm. I'm guessing you don't have SELinux and thus we need to omit the 'context' parameter.
I do not SELinux. I gave this a quick shot locally by removing the context argument but I'm still getting the error:
$ forklift 3985400cef6c sshd --mount-root ~/docker/fs/
mount: /dev/mapper/docker-252:0-4468428-56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e already mounted or /var/lib/docker/devicemapper/mnt/56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e/ busy
Traceback (most recent call last):
File "/usr/local/bin/forklift", line 9, in <module>
load_entry_point('docker-forklift==0.2.47', 'console_scripts', 'forklift')()
File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 338, in main
return Forklift(sys.argv).main()
File "/usr/local/lib/python3.4/dist-packages/forklift/__init__.py", line 319, in main
return driver.run(*command)
File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 240, in run
return self.run_sshd()
File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 450, in run_sshd
self.mount_root(container)
File "/usr/local/lib/python3.4/dist-packages/forklift/drivers.py", line 369, in mount_root
subprocess.check_call(args)
File "/usr/lib/python3.4/subprocess.py", line 557, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['sudo', 'mount', '/dev/mapper/docker-252:0-4468428-56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e', '/var/lib/docker/devicemapper/mnt/56d47997d635cf7fed30b26f5e98df19cabea5b3215dc234e1c9a1f0a062206e/']' returned non-zero exit status 32
| gharchive/issue | 2015-08-10T04:52:42 | 2025-04-01T06:44:31.263522 | {
"authors": [
"danni",
"jscn"
],
"repo": "infoxchange/docker-forklift",
"url": "https://github.com/infoxchange/docker-forklift/issues/97",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2210316478 | feat(azure): add support for azurerm_app_configuration
Adds support for Azure App Configuration, a centralized service for managing application settings and feature flags, enabling secure and dynamic configuration updates without redeployment.
This service is supported by Infracost through the azurerm_app_configuration Terraform resource. Which contains the following cost components:
Instance & replica costs - there is a cost per day for each store and replica that is enabled in Azure App Configuration
Additional Requests - Users are billed for every additional 10k requests (outside of the 200k allowance included with the standard instance/replica) to the service
Additionally, Azure App Configuration has both the Free and Standard skus. The former is free to run, so we simply return a free resource if a user has specified sku = "free" in the azurerm_app_configuration.
I'm not sure why the TestNetworkDdosProtectionPlan are failing now with the usage string, seems odd as these are in the usage example file 🤷
@tim775 digging around here https://azure.microsoft.com/en-gb/pricing/calculator/ I realised that the 200k number is just the number of requests included in the base cost of the instance/replica. It is not a daily number but a monthly one.
@tim775 digging around here https://azure.microsoft.com/en-gb/pricing/calculator/ I realised that the 200k number is just the number of requests included in the base cost of the instance/replica. It is not a daily number but a monthly one.
Are you sure? I know it's not definitive, but this page says:
Cost: Standard tier stores have a daily usage charge. The first 200,000 requests each day are included in the daily charge. There's also an overage charge for requests past the daily allocation. There's no cost to use a free tier store.
And the the $1.20/day per replica matches up with the 200,000 @ $0.06 per 10k. So it seems like the 200,000 is kind of like a "minimum" requests per day.
Are you sure? I know it's not definitive, but this page
Hmm i am not sure.. but that page looks semi-definitive, i'll revert the last commit and merge with the replica/day messaging
| gharchive/pull-request | 2024-03-27T09:42:16 | 2025-04-01T06:44:31.273765 | {
"authors": [
"hugorut",
"tim775"
],
"repo": "infracost/infracost",
"url": "https://github.com/infracost/infracost/pull/2970",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
513693942 | Update joda-time to 2.10.5
Updates joda-time:joda-time from 2.9.9 to 2.10.5.
Release Notes/Changelog
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "joda-time", artifactId = "joda-time" } ]
labels: semver-minor
Codecov Report
Merging #316 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #316 +/- ##
======================================
Coverage 80.5% 80.5%
======================================
Files 128 128
Lines 3334 3334
Branches 136 136
======================================
Hits 2684 2684
Misses 650 650
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update f5d3309...0d752af. Read the comment docs.
| gharchive/pull-request | 2019-10-29T04:58:28 | 2025-04-01T06:44:31.281236 | {
"authors": [
"codecov-io",
"scala-steward"
],
"repo": "ing-bank/baker",
"url": "https://github.com/ing-bank/baker/pull/316",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1714554735 | Circular dependencies
It just so happened to me that CPAN.pm died when it encountered these circular dependencies:
Recursive dependency detected:
Lingy (have: N/A; want: 0.001004)
=> INGY/Lingy-0.1.4.tar.gz
=> YAMLTest (have: N/A; requires: 0)
=> INGY/YAMLTest-0.0.4.tar.gz
=> Test::More::YAMLScript (have: N/A; requires: v0.0.4)
=> INGY/Test-More-YAMLScript-0.1.0.tar.gz
=> YAMLScript (have: N/A; requires: v0.1.0)
=> INGY/YAMLScript-0.1.0.tar.gz
=> Lingy (have: N/A; requires: v0.1.4).
Cannot resolve.
INGY/YAMLScript-0.1.0.tar.gz
[depend] -- NOT OK
Thanks for reporting this. It seems to be resolved in 0.1.11. Closing for now.
| gharchive/issue | 2023-05-17T19:45:37 | 2025-04-01T06:44:31.292920 | {
"authors": [
"andk",
"ingydotnet"
],
"repo": "ingydotnet/lingy",
"url": "https://github.com/ingydotnet/lingy/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2438817095 | Fix y2k (Y2K-1162)
Updated with today's date
Well... yeah.
| gharchive/pull-request | 2024-07-30T23:10:25 | 2025-04-01T06:44:31.340938 | {
"authors": [
"billlumbergh-initech",
"petergibbons2"
],
"repo": "initech-eng/y2k",
"url": "https://github.com/initech-eng/y2k/pull/973",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1242833513 | Support for Ctrl-[
I understand it's not feasible to have Inkdrop parse my .vimrc, but supporting the standard Ctrl-[ (does the same thing as Esc) would be awesome. Thanks!
You can customize keymaps as you like. Please read the documentation:
https://docs.inkdrop.app/manual/customizing-keybindings
https://github.com/inkdropapp/inkdrop-vim#key-customizations
| gharchive/issue | 2022-05-20T08:47:39 | 2025-04-01T06:44:31.366080 | {
"authors": [
"craftzdog",
"jcredmond"
],
"repo": "inkdropapp/inkdrop-vim",
"url": "https://github.com/inkdropapp/inkdrop-vim/issues/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
262098169 | React 16 support
Update peer dependencies to include React 16.
Update Enzyme to version that supports React 16.
Included in 0.7.0.
| gharchive/pull-request | 2017-10-02T14:10:52 | 2025-04-01T06:44:31.380994 | {
"authors": [
"bebraw",
"sapegin"
],
"repo": "inlinestyle/remark-react-lowlight",
"url": "https://github.com/inlinestyle/remark-react-lowlight/pull/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1655890201 | Expose core logging setup through the stable api
As discussed here: https://inmanta.slack.com/archives/CKRF0C8R3/p1680678625072959
It would be nice to expose (at least some part) of the logging setup done by core through the stable api, to allow other python tools developed internally to reuse it. Simplifying their development and making them more consistent with core.
The method which would be handy are:
inmanta.app._get_default_stream_handler()
inmanta.app._get_log_formatter_for_stream_handler(timed=True)
inmanta.app._convert_to_log_level(verbose)
Special attention should be paid to naming and docstrings.
https://code.inmanta.com/solutions/tools/yang-module-generator/-/blob/master/src/yang_module_generator/cli.py#L77-97
https://code.inmanta.com/solutions/modules/restbase/-/blob/master/inmanta_plugins/restbase/generator/app.py#L24-41
| gharchive/issue | 2023-04-05T15:51:34 | 2025-04-01T06:44:31.389101 | {
"authors": [
"edvgui",
"sanderr"
],
"repo": "inmanta/inmanta-core",
"url": "https://github.com/inmanta/inmanta-core/issues/5815",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
952818096 | Add resource history endpoint (#3126)
Pull request opened by the merge tool on behalf of #3126
Pull request rejected by merge tool. The tests for this branch did not succeed.
Merged into branches iso4 in eb6cd4c028892866396d6f2f2592df687c9f530f
| gharchive/pull-request | 2021-07-26T11:49:05 | 2025-04-01T06:44:31.390268 | {
"authors": [
"inmantaci"
],
"repo": "inmanta/inmanta-core",
"url": "https://github.com/inmanta/inmanta-core/pull/3138",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2265952387 | Fix core build failures
Description
Revert the tox.ini changes until the irt changes are in again https://github.com/inmanta/irt/pull/2022
Self Check:
Strike through any lines that are not applicable (~~line~~) then check the box
[ ] Attached issue to pull request
[ ] Changelog entry
[ ] Type annotations are present
[ ] Code is clear and sufficiently documented
[ ] No (preventable) type errors (check using make mypy or make mypy-diff)
[ ] Sufficient test cases (reproduces the bug/tests the requested feature)
[ ] Correct, in line with design
[ ] End user documentation is included or an issue is created for end-user documentation (add ref to issue here: )
[ ] If this PR fixes a race condition in the test suite, also push the fix to the relevant stable branche(s) (see test-fixes for more info)
Processing this pull request
Merged into branches master in 280804f1e574ffc258957a847f912286ee45f4c9
| gharchive/pull-request | 2024-04-26T14:50:11 | 2025-04-01T06:44:31.394377 | {
"authors": [
"Hugo-Inmanta",
"inmantaci"
],
"repo": "inmanta/inmanta-core",
"url": "https://github.com/inmanta/inmanta-core/pull/7569",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2046960648 | Bump inmanta-dev-dependencies[async,extension,pytest] from 2.96.0 to 2.97.0 (#493)
Pull request opened by the merge tool on behalf of #493
Merged in f5ab68bd7bd1ac36ec4fc98419d8199967ec6767
| gharchive/pull-request | 2023-12-18T15:52:04 | 2025-04-01T06:44:31.395546 | {
"authors": [
"inmantaci"
],
"repo": "inmanta/inmanta-ui",
"url": "https://github.com/inmanta/inmanta-ui/pull/495",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1075492134 | Build(deps-dev): Bump @storybook/addon-links from 6.4.8 to 6.4.9 (#2138)
Pull request opened by the merge tool on behalf of #2138
Merged in c75b4352c15be70792932699540e9007c9246317
| gharchive/pull-request | 2021-12-09T12:00:14 | 2025-04-01T06:44:31.396563 | {
"authors": [
"inmantaci"
],
"repo": "inmanta/web-console",
"url": "https://github.com/inmanta/web-console/pull/2139",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1150296670 | Issue/2632 url state filter hook (#2859)
Pull request opened by the merge tool on behalf of #2859
Merged in 77fb58c3466876d5035a3bcafc623caf39f7165a
| gharchive/pull-request | 2022-02-25T10:39:08 | 2025-04-01T06:44:31.397347 | {
"authors": [
"inmantaci"
],
"repo": "inmanta/web-console",
"url": "https://github.com/inmanta/web-console/pull/2885",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1441948433 | remove link to dashboard
Description
The link to the dashboard has been removed from the sidebar.
closes #4108
Self Check:
Strike through any lines that are not applicable (~~line~~) then check the box
[x] Attached issue to pull request
[x] Changelog entry
[x] Code is clear and sufficiently documented
[x] Sufficient test cases (reproduces the bug/tests the requested feature)
[x] Correct, in line with design
[ ] End user documentation is included or an issue is created for end-user documentation (add ref to issue here: )
Reviewer Checklist:
[ ] Sufficient test cases (reproduces the bug/tests the requested feature)
[ ] Code is clear and sufficiently documented
[ ] Correct, in line with design
Processing this pull request
| gharchive/pull-request | 2022-11-09T11:47:14 | 2025-04-01T06:44:31.401284 | {
"authors": [
"LauraStordeur",
"inmantaci"
],
"repo": "inmanta/web-console",
"url": "https://github.com/inmanta/web-console/pull/4303",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2178839272 | One Tap for Android will be deprecated soon
Caution: One Tap for Android will be deprecated soon. To ensure the continued security and usability of your app, migrate to Credential Manager today. Credential Manager supports passkey, password, and federated identity authentication (such as Sign-in with Google), stronger security, and a more consistent user experience.
This is a warning from the developer platform. Android's one-click login will be deprecated soon. Unfortunately, I have not found a better library than this. I hope the author can update it. Thank you.
Hi @naonao-a , thank you for the report. I am the author.
You are right. The code file GoogleOneTapAuth.java uses classes in the com.google.android.gms.auth.api.signin namespace as GoogleSignIn, GoogleSignInAccount and GoogleSignInClient which are depreciated.
However the code also already uses Sign in with Google. You can see in the GoogleOneTapAuth.java file that the classes Identity and SignInClient from the com.google.android.gms.auth.api.identity namespace are used to auto-sign-in the user. Sign-in using a button is using the depreciated way and needs to be updated as described in the link you mentioned.
I will try to release an updated version in the next 7 days.
OH!! that's very kind of you .Thank you.
@naonao-a I also am interested in this update.
@naonao-a, @jamesgroat,
the fix is now merged and there is a new npm package with version 6.0.0. The new package uses capacitor version 6. The android code was migrated from Java to Kotlin. Further the return type of tryAutoOrOneTapSignIn, tryOneTapSignIn and tryAutoSignIn was simplified to return just one promise instead of a promise containing promises. The documentation in the README is adjusted.
@naonao-a, @jamesgroat, the fix is now merged and there is a new npm package with version 6.0.0. The new package uses capacitor version 6. The android code was migrated from Java to Kotlin. Further the return type of tryAutoOrOneTapSignIn, tryOneTapSignIn and tryAutoSignIn was simplified to return just one promise instead of a promise containing promises. The documentation in the README is adjusted.
Thank you so much.
| gharchive/issue | 2024-03-11T11:00:39 | 2025-04-01T06:44:31.447865 | {
"authors": [
"inouiw",
"jamesgroat",
"naonao-a"
],
"repo": "inouiw/capacitor-native-google-one-tap-signin",
"url": "https://github.com/inouiw/capacitor-native-google-one-tap-signin/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1598776333 | Integrate ruff
Check whether scraping flake and flakeheaven for ruff would not be the best idea to proceed.
Difficulty will emerge when choosing rules as most of that has to be hand verified.
Most of the current checks are implemented (and there are some additional ones which seem to fit well), namely:
pyupgrade (UP)
flake8-2020 (YTT)
flake8-bandit (S)
flake8-blind-except (BLE)
flake8-boolean-trap (FBT)
flake8-comprehensions (C4)
flake8-datetimez (DTZ)
flake8-debugger (T10)
flake8-import-conventions (ICN)
flake8-no-pep420 (INP)
flake8-print (T20)
flake8-raise (RSE)
flake8-return (RET)
flake8-self (SLF)
flake8-type-checking (TCH)
flake8-unused-arguments (ARG)
eradicate (ERA)
Currently missing for ruff which we would like to support:
flake8-noqa
flake8-annotations-coverage
flake8-annotations-complexity (open issue here)
flake8-cognitive-complexity is better supported via settings (see here)
| gharchive/issue | 2023-02-24T14:50:24 | 2025-04-01T06:44:31.456516 | {
"authors": [
"szymonmaszke"
],
"repo": "inovintell/py-template",
"url": "https://github.com/inovintell/py-template/issues/142",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1855748229 | Feat: OpenAPI version in URL path
Context
API-breaking changes will be a challenge to manage without introducing a version scheme in the URL. Since we're building the client in the same scope as the server, it's possible to abstract the versioning concerns within.
Proposed Solution
Include the OpenAPI version within the URL scheme
Embed the expected version within the client HTTP requests by reading the version from a file generated at build time
The tests could use refactoring, as clearly there's some repetition evidenced by the number of changes this PR contains, but I've left that out of scope.
Important Changes Introduced
While completing this task I found the service level OpenAPI docs were missing the health check paths. I've fixed this in a preliminary commit.
Pushed deployment fixes, now it should be good and ready for merge. I want to test this out before doing that, @blaggacao could you help us with that?
could you help us with that [assert the changed to the deployment manifests]?
The suggestive next steps are:
wait for an update to the diff from CI
carefully audit the diff: in a CD scenario, that's what a final review on this PR signs off on, as well
merge and wait for deployment (side note: I've reasons to believe that this step turned out not working as intended, I'll need to revise this. I have this backlogged as LW-8219)
... iterate if the changes turned out not working as intended.
For an even shorter feedback loop to polish the deployment manifests, I can think of three options:
arion / docker-compose playground
kind playground
some of the existing operational playgrounds
We may need to find and discuss our preferred option, still.
I can not consider this 'testing', also alternative options arent satisfactory since we are dealing with k8s and AWS integration testing
| gharchive/pull-request | 2023-08-17T21:43:58 | 2025-04-01T06:44:31.463740 | {
"authors": [
"blaggacao",
"gytis-ivaskevicius",
"rhyslbw"
],
"repo": "input-output-hk/cardano-js-sdk",
"url": "https://github.com/input-output-hk/cardano-js-sdk/pull/872",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
679608883 | Optimise DeserializeShort.substring
In my microbenchmark, this brings down the cost of deserialising a Shelley
address in viewCompactTxOut from ~360ns to ~60ns.
The cloneByteArray function was introduced in version 0.7.1.0 of the
primitive package, so update the lower bound accordingly.
I believe the old primitive < 0.7 constraint in cabal.project was due to
older versions of the cborg package having >=0.5 && <0.7.1.0 as bounds for
primitive, but cborg-0.2.4.0 bumped the upper bound to 0.8.
This will likely require similar changes to downstream repos.
Using cloneByteArray is exactly what was needed. This should speed up things considerably. I cannot speak to the version numbers the projects. Thanks for doing this.
| gharchive/pull-request | 2020-08-15T16:58:40 | 2025-04-01T06:44:31.466643 | {
"authors": [
"TimSheard",
"mrBliss"
],
"repo": "input-output-hk/cardano-ledger-specs",
"url": "https://github.com/input-output-hk/cardano-ledger-specs/pull/1786",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1070413899 | [BUG] - Plutus script cannot unlock funds with no attached datum
Internal/External
External otherwise.
Area
Plutus Related to Plutus Scripts (Alonzo).
Summary
A delegator sent some Astor tokens to our Plutus contract directly in Tx 64718…02513.
The normal/documented procedure is to send these tokens to a proxy service, which then constructs/submits the non-trivial Plutus Tx
AFAIK, these 80 ADA are now lost, because Plutus cannot release funds that have no datum hash associated with the given UTxO
[03-12-2021 10:41:43] INFO Query UTxO for Script
64718d583eefea766dda6787a844eee52d9c3e4301c39fd6eeef3ddb73402513#0 value {lovelace 1444443 3f997b68b1f491c7c2f10af4e2bf9566c5d25bd61df0343065d4fe1c.Astor302 80} datum TxOutDatumNone
ba35d7e12f80988f6f2dd36bdb7ced58501711c5b84f8e0968d7d1b2e0be686d#1 value {lovelace 168000000 3f997b68b1f491c7c2f10af4e2bf9566c5d25bd61df0343065d4fe1c.Astor296 804} datum d8a95d4a0b1cd3185fc571eb4ad1c293cdee44fbdf45400616058662b817b827
312d07fbe8290cc541feaed2e4b030237092a55fd570c2eecaa1969e281e6e96#1 value {lovelace 901700000 3f997b68b1f491c7c2f10af4e2bf9566c5d25bd61df0343065d4fe1c.Astor302 716} datum e31b52bb6cf0017e24cfc13ede621e913565e68b364254c009335b470cfff870
I would assume that this is a common enough user error, which may cause lots of frustration. Can we please fix this, for example by providing a known datum value that maps to TxOutDatumNone - (possibly empty byte string)?
Closing this. If this is still relevant please reopen.
| gharchive/issue | 2021-12-03T09:57:11 | 2025-04-01T06:44:31.470169 | {
"authors": [
"Jimbo4350",
"tdiesler"
],
"repo": "input-output-hk/cardano-node",
"url": "https://github.com/input-output-hk/cardano-node/issues/3414",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1293445398 | [BUG] - VASIL cardano-cli query leadership-schedule Error (Next Epoch)
External
Area
Other: cardano-node/cardano-cli
Summary
Error happens when the leadership schedule command in Vasil Testnet for the next epoch is runned on a block producer.
Steps to reproduce
On a block producing node in testnet running cardano-cli 1.35.0
run the cardano-cli query leadership-schedule --testnet-magic 1097911063 ... --next
Error appears after few seconds
Expected behavior
Leadership schedule for block production for the next Epoch
System info:
OS Name: Ubuntu
OS Version 20.04.4 LTS
Node version:
cardano-node 1.35.0 - linux-x86_64 - ghc-8.10
git rev 9f1d7dc163ee66410d912e48509d6a2300cfa68a
CLI version:
cardano-cli 1.35.0 - linux-x86_64 - ghc-8.10
git rev 9f1d7dc163ee66410d912e48509d6a2300cfa68a
Error log :
Error: Error while calculating the slot range: PastHorizon {...}
75% of epoch wasn't completed..
| gharchive/issue | 2022-07-04T17:49:03 | 2025-04-01T06:44:31.475451 | {
"authors": [
"Techs2Help"
],
"repo": "input-output-hk/cardano-node",
"url": "https://github.com/input-output-hk/cardano-node/issues/4131",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
464533939 | LiveViewState. TUI draw simplification. Node's data.
LiveViewState. TUI draw simplification. Node's data.
bors r+
| gharchive/pull-request | 2019-07-05T08:50:18 | 2025-04-01T06:44:31.476425 | {
"authors": [
"CodiePP",
"denisshevchenko"
],
"repo": "input-output-hk/cardano-node",
"url": "https://github.com/input-output-hk/cardano-node/pull/40",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1341826584 | cardano-node: sync percent fixed.
Closes #4345
bors r+
| gharchive/pull-request | 2022-08-17T14:09:33 | 2025-04-01T06:44:31.477186 | {
"authors": [
"denisshevchenko"
],
"repo": "input-output-hk/cardano-node",
"url": "https://github.com/input-output-hk/cardano-node/pull/4350",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
232115282 | [CSL-154] Improve storing of global data in GtLocalData
@gromak
Currently we store all global data there and fully update it when main blockchain is updated. However, we don't need to store all this data. For example, we don't need to store openings, we only need to store keys after 3k-th slot. We don't need to store full commitments after first k slots. And so on.
Another thing is how we update it. Current approach is to copy global payload into GtLocalData every time main blockchain is updated. I think it's not bad, but maybe there are bad options.
@iperesadin
We don't need to store full commitments after first k slots.
But we use commitments in checkShare (Pos.Ssc.GodTossing.Functions), in processShares (Pos.Ssc.GodTossing.LocalData.LocalData)
@gromak
@iperesadin yes, but we don't need commitments from nodes who sent us openings.
| gharchive/issue | 2017-05-30T00:13:34 | 2025-04-01T06:44:31.480123 | {
"authors": [
"jagajaga"
],
"repo": "input-output-hk/cardano-sl",
"url": "https://github.com/input-output-hk/cardano-sl/issues/532",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
636570405 | Tx fee estimate edge case when Tx value approach wallet value
Context
When calculating a transaction fee, if the amount to pay is set to the same value as the wallet holds (in a single address in this test case) or close to it, the TX fee is about as large as the wallets value. This likely has no functional impact as it looks like these TXs should fail anyway due to insufficient fee ADA. But in case it impacts something somewhere else, just FYI.
Network discrimination for the send to address is not tested for when calculating fees. This might be by design and not an issue. Just FYI
Information
-
Version
2020.6.5 (git revision: 5290acf902da32e1e541b5dc227629c9a300f1be)
Platform
Linux
Installation
Nix build
Steps to Reproduce
Check the wallet balance
$ cardano-wallet-shelley wallet get $WALLETID
...
"total": {
"quantity": 99627083,
"unit": "lovelace"
},
"available": {
"quantity": 99627083,
"unit": "lovelace"
}
...
Estimate fees by CLI or API:
# Set the payment about ~0.13 ADA lower than the wallet value; get a large estimated fee
#
$ cardano-wallet-shelley transaction fees $WALLETID --payment 99500000@$SEND_TO
Ok.
{
"estimated_min": {
"quantity": 99665281,
"unit": "lovelace"
},
"estimated_max": {
"quantity": 99665281,
"unit": "lovelace"
}
}
# Set the payment about ~1.13 ADA lower than the wallet value; get an accurate fee
#
$ cardano-wallet-shelley transaction fees $WALLETID --payment 98500000@$SEND_TO
Ok.
{
"estimated_min": {
"quantity": 168141,
"unit": "lovelace"
},
"estimated_max": {
"quantity": 168141,
"unit": "lovelace"
}
}
Perhaps not an issue, the network discriminator of the send to address doesn't appear to be checked during fee estimates. Repeat one of the above fee estimates, this time flipping the second digit of the address from a 1 to a 0 or visa-versa. The command completes the same way, successfully, despite one of those addresses not being valid on the network.
$ cardano-wallet-shelley transaction fees $WALLETID \
--payment 99500000@$SEND_TO_WITH_MODIFIED_SECOND_DIGIT
Ok.
{
"estimated_min": {
"quantity": 99665281,
"unit": "lovelace"
},
"estimated_max": {
"quantity": 99665281,
"unit": "lovelace"
}
}
Indeed, I am able to reproduce it for both: shelley and byron wallet. The boundary value (on Shelley testnet) seems to be 165281 lovelace.
I.e.
Where value is the amount for which you want to estimate fee:
If (value <= wallet_balance - 165281) then :+1:
if (value > wallet_balance - 165281) then :-1:
For instance:
wallet balance = 5 000 000
Estimate fee for 5000000−165281 = 4834719
result: 165281 lovelace (estimated_min and estimated_max)
Estimate fee for 5000000−165280 = 4834720
result: 5000001 lovelace (estimated_min and estimated_max)
lgtm
| gharchive/issue | 2020-06-10T21:46:43 | 2025-04-01T06:44:31.488164 | {
"authors": [
"johnalotoski",
"piotr-iohk"
],
"repo": "input-output-hk/cardano-wallet",
"url": "https://github.com/input-output-hk/cardano-wallet/issues/1740",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
722459598 | Add delistPools to Pool DBLayer
This is the first step for garbage collecting
stake pools based on SMASH delisting.
X-JIRA-Ticket: ADP-478
Questions / Considerations
After looking at the logic, I believe we don't have to adjust any other functions of the DBLayer such as listRegisteredPools or readPoolProduction. These can still consider all pools. Only the API layer will consider delisted pools and adjust ApiStakePool accordingly.
There's still the open question whether we want to really delete GCed pools via removePools at some point.
TODO
Feature
[x] Add delisted column to metadata table and populate Pool DBLayer with functions
[x] Create new internal_state table and have GC thread store last sync time
[x] Add garbage collection thread querying the SMASH server for delisted pools and calling delistPools from DBLayer
[x] have GC thread store last sync time
[x] return last GC sync time
[x] Add POST '{ "maintenance_action": "gc_stake_pools" }' /stake-pools endpoint
QA
[x] database tests
[x] check json roundtripping works
[ ] Add integration tests? We don't have SMASH server integration tests yet.
bors try
bors try
bors try
bors try
bors try
Add integration tests? We don't have SMASH server integration tests yet.
Perhaps we could have a mock SMASH server in the integration tests? Maybe it would be enough as it just mock the situation that there is one stake pool blacklisted. This way we could make a system level test.
Currently in our integration suite there are 4 pools being registered. One of them is retired at the very beginning (before the tests start). Two out of remaining three are a scheduled to be retired in the far future (and they are like that for the span of the suite). We could extend this set up to have yet another pool, but this one would served as "blacklisted" in the mock SMASH server and we would just make sure that this pool is not visible when we have metadata fetching strategy pointing to a mock SMASH server.
We used to have a mock server for jormungandr metadata registry, that was serving a zip file with metadata https://github.com/input-output-hk/cardano-wallet/blob/master/lib/jormungandr/test/integration/Main.hs#L230-L238... Maybe similar concept could be used.
I've tested delisting with the last commit manually with a mock smash server. It works.
bors try
In particular, I think we should define (with properties) how we expect this flag to behave in the presence of rollback. It looks as though rolling back won't cancel out the de-listing of a pool. If that is the desired behaviour, then it would be good to encode this as a property.
That sounds like an integration test. For integration test we want a proper SMASH process imo. Everything else is a half-baked unit test in disguise. I've done quite some work to try to integrate with it, but ultimately this boils down to:
https://github.com/input-output-hk/smash/issues/97
https://github.com/input-output-hk/smash/pull/99
https://github.com/input-output-hk/smash/pull/102
https://github.com/input-output-hk/cardano-db-sync/pull/372
The cardano-db-sync maintainer seems to have the opinion that we shouldn't use cardano-db-sync, which I find a bit confusing. So there are still a couple of options with sqlite and in-memory database. That'll still take some time to sort out.
But what we can do is write some basic testing of delistPools, such that it actually sets the flag.
I have some reservations about the naming of the poolFlag field, the PoolFlag type, and the poolRegistrationFlag DB field.
I don't have a strong opinion about the naming, but I have a strong opinion about changing naming back and forth in the process of a PR, because it actually takes more time than it seems (certainly 30 minutes just to make it compile and pushable, then possible another iteration, because I missed something and CI fails). The naming was suggested in the review comments and there was time to discuss it. I think both forms are fine, so I'll keep the current one.
@jonathanknowles wrote:
In particular, I think we should define (with properties) how we expect this flag to behave in the presence of rollback. It looks as though rolling back won't cancel out the de-listing of a pool. If that is the desired behaviour, then it would be good to encode this as a property.
@hasufell wrote:
That sounds like an integration test.
By rollback, I specifically mean the rollbackTo DB operation, rather than the entire rollback machinery (which is large).
We currently have properties covering all of the pool DB operations, and I think it would be desirable to also test delistPools in this way. For example:
the basic behaviour of delistPools;
the interaction between delistPools and other key operations such as rollbackTo.
Even if these end up being very simple properties, I think they are still worth having.
Advantages:
properties in Cardano.Pool.DB.Properties are tested against both the SQLite implementation and the Model implementation. In future, we test both of these implementations together using QuickCheck state machine, so we really do want to make sure these remain consistent.
properties can often generate many more edge cases than we're likely to think of by hand.
properties provide a statement of how these (essentially pure) functions are intended to work.
@jonathanknowles makes sense, I'll get on it
@hasufell wrote:
I don't have a strong opinion about the naming, but I have a strong opinion about changing naming back and forth in the process of a PR, because it actually takes more time than it seems
I completely sympathize. It's not easy to gain consensus in a PR review, especially when people are in multiple time-zones.
Though I also think the time we spend on a PR has to be weighed up against the time we spend on future work. Some names (such as DB columns) once merged to master are more awkward to change (if we need to) than internal variable names, as they require migrations, which we wish to avoid if at all possible.
In the case of poolRegistrationFlag, which really is very ambiguous to an uninitiated reader (IMO), I still think this would be worth changing to something more specific. (See proposal https://github.com/input-output-hk/cardano-wallet/pull/2249#pullrequestreview-516440993)
Just my 2 ada. :+1:
In the case of poolRegistrationFlag, which really is very ambiguous to an uninitiated reader (IMO), I still think this would be worth changing to something more specific.
The current naming was suggested by @KtorZ. Maybe you two can sort that out :smile: Otherwise I'll change it again tomorrow to something else.
the basic behaviour of delistPools;
It just sets a flag in pool_registrations.
the interaction between delistPools and other key operations such as rollbackTo.
There really is none. The interaction doesn't happen on DBLayer level. It's higher up in the program stack, so only an integration test can touch that.
The only property we really have is checking whether the Delisted flag shows up.
Hi Jonathan,
as you might know, @KtorZ is my co-pilot and so certain decisions have been discussed with him or have been proposed by him.
To that end, the existing solution goes to some lengths to ensure that all rows in the pool_registration table have the same delisted status for a particular pool.
But this seems unnecessary.
At the end of the day, the set of delisted pools is just a set, which should be possible to represent with a single-column SQL table. So I think it would make things a lot simpler to just store the set of delisted pools in its own table, rather than storing each pool's delisted status multiple times, once for each registration record:
We discussed this already and the proposed solution was to put the flag into the pool registration table. And I'd like to stick with that now, otherwise this PR will become a chaotic back and forth of ideas.
Just an update:
@hasufell and I had an out-of-band discussion about this PR, and we came to the conclusion that it would make sense to adopt the DB schema design proposed in #2277, which stores the delisted status just once per pool.
bors try
bors try
bors r+
bors r+
| gharchive/pull-request | 2020-10-15T15:52:14 | 2025-04-01T06:44:31.511887 | {
"authors": [
"KtorZ",
"hasufell",
"jonathanknowles",
"piotr-iohk"
],
"repo": "input-output-hk/cardano-wallet",
"url": "https://github.com/input-output-hk/cardano-wallet/pull/2249",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
833543372 | Nightly restore bench: Make buildkite red if it fails
Issue Number
ADP-804
Overview
If the nightly restore bench fails to produce a result, then make sure Buildkite has a red status.
bors r+
bors r+
bors r+
| gharchive/pull-request | 2021-03-17T08:52:04 | 2025-04-01T06:44:31.514545 | {
"authors": [
"Anviking",
"piotr-iohk",
"rvl"
],
"repo": "input-output-hk/cardano-wallet",
"url": "https://github.com/input-output-hk/cardano-wallet/pull/2566",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2604628705 | feat(rust/catalyst-voting): Vote protocol benchmarks
Description
Added benchmarks for all of the steps of the catalyst vote protocol.
Refactored some stuff.
@minikin not sure that its actually what we want here, because I will post a results made on my machine, so they will be not the same if some other will run it.
Maybe we need to define a general strategy how we want to run our benchmarks on some regular basis, maybe something the same as we have codecov reports.
Seems we need to raise this topic on our tech discussions and decide what to do.
@Mr-Leshiy, can you please create a task about publishing benchmark results? Let's discuss this topic in our next technical discussion.
| gharchive/pull-request | 2024-10-22T08:08:06 | 2025-04-01T06:44:31.516774 | {
"authors": [
"Mr-Leshiy",
"minikin"
],
"repo": "input-output-hk/catalyst-libs",
"url": "https://github.com/input-output-hk/catalyst-libs/pull/65",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1705169054 | CI bits
Split up Cicero into independent jobs such that the logs of different platforms are no longer interleaved, easing log inspection
We temporarily bumped the GHC on GHA to 9.2 in #46 as it started to randomly fail with weird out-of-memory errors on 8.10. We still have to support compiling for 8.10 for ~one month, so we are reverting that change here.
Will cicero build with docs enabled? i.e. an error in the haddocks will be caught somewhere?
Will cicero build with docs enabled? i.e. an error in the haddocks will be caught somewhere?
Not yet, will change that :+1:
Actually, adding -haddock no longer makes the GHC parser more strict since GHC 9.0 (see https://gitlab.haskell.org/ghc/ghc/-/issues/23379). We could enable -Winvalid-haddock, which currently generates lots of warnings (e.g. all Haddock-ish comments on non-toplevel bindings), but they are not blocking Haddock generation.
I will try to enable Haddock again for GHA, maybe it works again :shrug:
| gharchive/pull-request | 2023-05-11T06:41:09 | 2025-04-01T06:44:31.539529 | {
"authors": [
"amesgen",
"jasagredo"
],
"repo": "input-output-hk/ouroboros-consensus",
"url": "https://github.com/input-output-hk/ouroboros-consensus/pull/76",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
924978453 | SCP-2232: Chain index types
Starting to separate the chain index from the PAB. See https://hackmd.io/ji8Hs_88TlqMOK0VeX9Row for the original specification.
This only defines the types and query interface, implementation still TBD.
Pre-submit checklist:
Branch
[ ] Commit sequence broadly makes sense
[ ] Key commits have useful messages
[ ] Relevant tickets are mentioned in commit messages
[ ] Formatting, materialized Nix files, PNG optimization, etc. are updated
PR
[ ] Self-reviewed the diff
[ ] Useful pull request description
[ ] Reviewer requested
Pre-merge checklist:
[ ] Someone approved it
[ ] Commits have useful messages
[ ] Review clarifications made it into the code
[ ] History is moderately tidy; or going to squash-merge
@sjoerdvisscher I fixed the definitions and added a property test. This wasn't straightforward because I had to write the generators to produce the right data. But it seems to be good now.
Some of these generators needed to be written anyway, right?
I really like the code overall. 💯
@michaelpj this is just hanging on a VM test, could please give it a little nudge? Thank you!
| gharchive/pull-request | 2021-06-18T14:50:24 | 2025-04-01T06:44:31.545073 | {
"authors": [
"j-mueller",
"sjoerdvisscher"
],
"repo": "input-output-hk/plutus",
"url": "https://github.com/input-output-hk/plutus/pull/3399",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
673802370 | export getFileWithAcl alongside getFile
This PR exports getFileWithAcl alongside getFile, following the pattern with getSolidDataset and getSolidDatasetWithAcl. I didn't add a unit test because I don't see a pattern of testing exports.
[ ] I've added a unit test to test for potential regressions of this bug.
[ ] The changelog has been updated, if applicable.
[ x] Commits in this PR are minimal and have descriptive commit messages.
We're a bit caught up in some other work, but I'll try to look at this after the weekend - hope that's OK.
no problem at all! I'm tracking this repository at https://github.com/itme/solid-client-js and just contributing back patches as I have them, so y'all aren't blocking me at all. thanks, btw, for this work, this library is an order of magnitude easier to use than anything else out there already.
Yeah, WhiteSource is failing due to this PR coming in as an external contribution (we're aware of this problem, and are working to fix it). In the meantime, I'll create a separate PR to get this fix in - thanks again @travis !
yw, thanks!
| gharchive/pull-request | 2020-08-05T19:44:58 | 2025-04-01T06:44:31.564068 | {
"authors": [
"Vinnl",
"pmcb55",
"travis"
],
"repo": "inrupt/solid-client-js",
"url": "https://github.com/inrupt/solid-client-js/pull/319",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
811588924 | Chore/move datafactory to rdfjs
Internal refactor - just moved all DataFactory imports to pull from rdfjs.js rather than from N3 directly. The Turtle parsing code still relies explicitly on N3, but that's fine (test needed to be updated to support quad-equality between RDF/JS implementations, but that's a test improvement).
This change makes it really easy to flip our DataFactory implementation dependency now - and so I did try that out by flipping it from N3 to @rdfjs/data-model (that flip is not included in this PR). Only 4 tests failed, which is pretty good - with the reasons documented in rdfjs.ts. Those reasons need to be verified, but it would be nice to resolve them (not urgent though - we're still relying on the RDF spec compliance of N3, as we did before).
Merged in #837
| gharchive/pull-request | 2021-02-19T00:46:43 | 2025-04-01T06:44:31.566178 | {
"authors": [
"ajacksified",
"pmcb55"
],
"repo": "inrupt/solid-client-js",
"url": "https://github.com/inrupt/solid-client-js/pull/840",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
691411583 | Fix "KeyError: 'creator'" bug with private events
Adds missing check on nested dictionary access
@jcrowgey, can you take a look now?
Updated to fix merge conflicts with master. @jcrowgey LMK if you want to review and sign off, otherwise I can take over for you and get it merged.
It would take me some time to recover the context. I'm happy to hand it off to you.
On August 22, 2024 9:47:25 PM PDT, David Barnett @.***> wrote:
Updated to fix merge conflicts with master. @jcrowgey LMK if you want to review and sign off, otherwise I can take over for you and get it merged.
--
Reply to this email directly or view it on GitHub:
https://github.com/insanum/gcalcli/pull/562#issuecomment-2306284875
You are receiving this because you were mentioned.
Message ID: @.***>
| gharchive/pull-request | 2020-09-02T20:50:46 | 2025-04-01T06:44:31.569595 | {
"authors": [
"dbarnett",
"jcrowgey",
"shaicoleman"
],
"repo": "insanum/gcalcli",
"url": "https://github.com/insanum/gcalcli/pull/562",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
159568691 | ported
To Cuda: https://sourceforge.net/projects/cudatext/files/addons/plugins/plugin.JS_Sort_Imports.zip/download
:+1:
| gharchive/issue | 2016-06-10T06:34:50 | 2025-04-01T06:44:31.577090 | {
"authors": [
"Alexey-T",
"insin"
],
"repo": "insin/sublime-sort-javascript-imports",
"url": "https://github.com/insin/sublime-sort-javascript-imports/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
370207197 | Line break not working
Hi, thanks for this great javascript plugin.
I notice a small issue I'm having.
I'd like to have the text on several lines including line breaks, and it seems they are not working/parsed.
window.addEventListener("load", function(){
window.cookieconsent.initialise({
"palette": {
"popup": {
"background": "#000"
},
"button": {
"background": "#f1d600"
}
},
"content": {
"message": "\nCe site internet utilise des cookies pour ameliorer votre experience.\n\nEn naviguant sur les page du site internet vous acceptez l'utilisation des cookies.",
"dismiss": "OK, J'AI COMPRIS",
"link": "En savoir plus"
}
})});
Hi Tristof,
You can use HTML tags instead.
For your case, you can use
Thanks
| gharchive/issue | 2018-10-15T15:04:13 | 2025-04-01T06:44:31.579307 | {
"authors": [
"Tristof",
"dolav"
],
"repo": "insites/cookieconsent",
"url": "https://github.com/insites/cookieconsent/issues/473",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1635139325 | Feature Request - Add option to no use debrid service
in my country there is no need to use a debrid service, why i have to use one with this application?
:/ sorry i'm going to wait for an actual solution, thank you for the answer, this project have future!
Initial level changes for streaming torrents are in.
It takes some time to start streaming. If it's lagging, pause for some time for the download to get ahead. (maybe it's my download speed being slow, idk)
On settings there is torrenting section
For nya you'll have to give a mirror that works for you.
Change your download server here https://premiumize.me/speedtest i recommend not using automatic, cloudflare, or direct add those tend to not play well with players. Personally digital ocean Toronto is my best server
Very
Initial level changes for streaming torrents are in. It takes some time to start streaming. If it's lagging, pause for some time for the download to get ahead. (maybe it's my download speed being slow, idk)
On settings there is torrenting section
For nya you'll have to give a mirror that works for you.
Change your download server here https://premiumize.me/speedtest i recommend not using automatic, cloudflare, or direct add those tend to not play well with players. Personally digital ocean Toronto is my best server
Added option to use VLC media player instead of windows media player.
VLC is better than WMP at streaming MKV files.
So if you're using torrents, use Vlc media player, otherwise use the default. (Vlc package is still in pre-release)
Playing with debrid service works perfectly now.
If you don't have a debrid service, I'm downloading the torrent while streaming, so it might some time to start playing.
I've added, download information to know if something is happening.
Video will start playing after first section is downloaded. After this i didn't notice any buffering (unless you seek).
Download will go to. \AppData\Local\Totoro\Torrents
Contents of this folder will get cleared at startup. so, if you want to keep the downloaded files, you might want to move them somewhere else before starting.
| gharchive/issue | 2023-03-22T06:26:36 | 2025-04-01T06:44:31.586889 | {
"authors": [
"DamianArielLerner",
"insomniachi",
"reddit-reaper"
],
"repo": "insomniachi/Totoro",
"url": "https://github.com/insomniachi/Totoro/issues/22",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1352153417 | Feat: Launchpad integration test
What?
Added Multi-process JAX integration test (matches ACME style testing which run in lp_launch_type TEST_MULTI_THREADING)
Why?
Due to issues with Launchpad, current tests are single process
How?
Implement simple test that calls each process once. Issue we had previously with not being able to access environment loop run_episode() has been fixed by exposing the method in the wrapper.
Extra
Refactor jax tests folder (move mock_systems.py into tests/jax/systems folder to house all test systems, rename variables for consistent integration tests)
Closes https://github.com/instadeepai/Mava/issues/702
Thank you @AsadJeewa! A general request/comment from my side. Can we please only name actual tests with names that start with test_ otherwise it seems that pytest runs unintended tests by accident. For example if I run pytest tests/jax/systems/ippo_system_test.py it says that 2 tests completed succesfully when there is only one due to the naming of the imported systems.
I updated the name of the systems to not be called test_ to solve this. I also named the files with the test systems system_test_data.py to avoid a similar issue
Thank you :)
| gharchive/pull-request | 2022-08-26T11:41:39 | 2025-04-01T06:44:31.726705 | {
"authors": [
"AsadJeewa",
"RuanJohn"
],
"repo": "instadeepai/Mava",
"url": "https://github.com/instadeepai/Mava/pull/703",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
812445246 | How to share to facebook
Hello, how to share to facebook de content published in instagram using instagram4j
this is my code
IGClient client = IGlogin();
client.actions()
.timeline()
.uploadPhoto(getFile(post.getContent()), post.getCaption())
.thenAccept(response -> {})
.join();
Sharing to other social media is not supported in this library currently.
| gharchive/issue | 2021-02-20T00:12:56 | 2025-04-01T06:44:31.730267 | {
"authors": [
"diegocallealvarez",
"jvogit"
],
"repo": "instagram4j/instagram4j",
"url": "https://github.com/instagram4j/instagram4j/issues/565",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
227971566 | Unfollow less the white list
I need to give unfollow in followers but I want to preserve about 100 followers how to do is on the whitelist? Is this programmed already?
@zsvendo try to use get_user_followers to get all followers and truncate this list kind a arr = arr[:100] after that use this list as "white list".
Hello, I have the whitelist of people not to unfollow, but when I execute Unfollow everyone, with (whitelist=True), it still unfollows whitelisted users. Any idea ? Thanks
How can i active whitelist for ultimate_schedule unfollow_non_follower?
def unfollow_non_followers(self):
self.logger.info("Unfollowing non-followers")
followings = set([item["pk"] for item in self.getTotalSelfFollowings()])
self.logger.info("You follow %d users." % len(followings))
followers = set([item["pk"] for item in self.getTotalSelfFollowers()])
self.logger.info("You are followed by %d users." % len(followers))
diff = followings - followers - whitelist('/friends.txt')
self.logger.info("%d users don't follow you back." % len(diff))
self.unfollow_users(list(diff))
Would this work : diff = followings - followers - whitelist('/friends.txt') ?
The whitelist.txt is not working anyone can have some idea:
I run the 4.unfollow option and then the 2. unfollow everyone option.
But it unfollow all does not respect the whitelist.txt
Should be fixed in current version.
| gharchive/issue | 2017-05-11T12:17:04 | 2025-04-01T06:44:31.735541 | {
"authors": [
"alcaporal",
"alexsonnay",
"maxkrivich",
"ohld",
"zsvendo"
],
"repo": "instagrambot/instabot",
"url": "https://github.com/instagrambot/instabot/issues/195",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1091616424 | How to set last post timestamp via LatestStamps("stampsfile").set_last_post_timestamp("myaccount", datetime)
I'm trying to write a Python bot that fetches the Instagram posts from an account and posts them on Twitter.
I want the bot to post just the newest post, therefore I thought I should make use of the LatestStamps Class.
I now have the problem that I don't fully understand how to get the post-timestamp of the post and how to set it in the stampsfile.
What kind of value does the datetime parameter need?
Nevermind...
for post in profile.get_post():
LatestStamps("stampsfile").set_last_post_timestamp("profile", post.date)
Nevermind...
for post in profile.get_post():
LatestStamps("stampsfile").set_last_post_timestamp("profile", post.date)
| gharchive/issue | 2021-12-31T16:13:36 | 2025-04-01T06:44:31.738230 | {
"authors": [
"LucEast"
],
"repo": "instaloader/instaloader",
"url": "https://github.com/instaloader/instaloader/issues/1367",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2330684637 | Instaloader Redirected to login page. Use --login Issue
I ran into the issue of Redirected to login page. Use --login
It seems Instagram changed their API, Is there any new fix to this or is this a new issue?
Was working fine until I pulled the same 3 profiles 10 times within 5 minutes, then I got a similar message
An error occurred while processing profile 'some_profile': Redirected to login page. Use --login.
| gharchive/issue | 2024-06-03T09:50:10 | 2025-04-01T06:44:31.739665 | {
"authors": [
"frozencap",
"pdbarma"
],
"repo": "instaloader/instaloader",
"url": "https://github.com/instaloader/instaloader/issues/2275",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1851088439 | Underlined featured
Hi dev team, could you please support the underlined featured? Thank you
Standard markdown syntax does not support underscores. If you want to using this, maybe you can using html tag or some style to support.
| gharchive/issue | 2023-08-15T08:35:33 | 2025-04-01T06:44:31.741979 | {
"authors": [
"hoangbtmrk",
"instance-oom"
],
"repo": "instance-oom/ngx-markdown-editor",
"url": "https://github.com/instance-oom/ngx-markdown-editor/issues/138",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2487688657 | Feature Request: Stripe Connect but for Instant DB
Feature request: Stripe Connect but for Instant DB
BLUF
We’d like to provision Instant DB accounts and apps on behalf of our users, similar to how Squarespace can provision Stripe accounts and payments on behalf of their users.
Need
GPT Engineer lets you build and deploy a complete frontend app using only prompts (powered by LLMs). For adding persistence we today support a Supabase integration. This requires the user to go into supabase, create an account, and configure the schema and then use the API keys to Connect to GPT Engineer. This is great for users that already use Supabase or who want a lot of control. But many users don’t care about what a persistence layer looks like as long as it gets the job done and would prefer minimal friction. We believe that Instant DB could unlock this low friction use case if there was a way for us to programmatically create Instant DB apps on behalf of our users. We don't want our users to have to go to instantdb.com and sign up, we'd want to sign up for them, ideally without them needing to do anything.
Proposed solution
A couple of API endpoints (not exactly how they should look but something to start the discussion):
POST /users
curl https://api.instantdb.com/v1/users \
-u "our_secret_api_key" \
-d email=test@example.com
-d password=long_difficult_password
Returns
{
id: string;
email: string;
}
GET /users/<id>
curl https://api.instantdb.com/v1/users/<id> \
-u "our_secret_api_key" \
-d email=test@example.com
Returns
{
id: string;
email: string;
}
POST /app
curl https://api.instantdb.com/v1/app \
-u "our_secret_api_key" \
-d user_id="user_id" \
-d name="cool_app"
Returns
{
name: string;
id: string;
public_api_key: string; // The user's client side public API key
}
GET /app/<id>
curl https://api.instantdb.com/v1/app/<id> \
-u "our_secret_api_key"
Returns
{
name: string;
id: string;
public_api_key: string; // The user's client side public API key
}
Concerns
Validating users
One possible concern here would be the validation of users and their emails. We naturally have our users already validated so either we establish trust that way or we device some other way of establishing trust. Ideally we'd like to avoid the user having to re-validate their email again for example.
Giving users access to their Instant DB dashboards
We naturally would like the user to have access to their Instant DB dashboards just as a normal Instant DB user. This is straightforward if they are using SSO since then that login can just be reused. But if the user is using GPT Engineer with email and password we will likely need to ask the user to add an Instant DB password when creating the user.
Proposal #2:
Enable claimable projects.
The UUID of claimable projects is secret. A person that knows the UUID can claim it. Once it is claimed it is editable by both the creator of the project and the person who claimed it.
Thanks for writing these up!
It sounds like the most important functionality is:
You should be able to spin up an app easily, and associate it with a user (i.e. an email address)
You should be able to programmatically give that user access to the app (either by transferring ownership, or by adding that user as a collaborator, or allowing apps to be "claimable")
Let me know if I'm understanding things correctly @vikeri @AntonOsika :)
Correct!
I think Viktor’s solution here is easiest to implement.
You would just need to store:
user record field: api secret it was created with (optional)
project record field: api secret it was created with (optional)
and verify api secret, on create project and edit project, if attempted by someone else than user
api secrets are given out after appropriate trust+contract is signed with third party
Let me know if I'm understanding things correctly
Yep!
| gharchive/issue | 2024-08-26T20:13:15 | 2025-04-01T06:44:31.750971 | {
"authors": [
"AntonOsika",
"reichert621",
"vikeri"
],
"repo": "instantdb/instant",
"url": "https://github.com/instantdb/instant/issues/45",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2347057380 | Create interface for accessing the training code
This PR turns this into a library.
View the following PR for an example: https://github.com/instructlab/instructlab/pull/1329
The main idea is that we provide a function like:
def train_torchrun(torchargs: TorchrunArguments, training_args: FullTrainingArguments):
pass
Each class TorchrunArguments and FullTrainingArguments provides training-specific arguments. And then from another library, you would simply provide these arguments to the train_torchrun function.
For any other training method that we define, we could provide a similar interface depending on which arguments are needed.
We separate the arguments here because TorchrunArguments are the ones passed to torchrun and then the full training arguments are the ones that we actually train with. It's not crucial that these are different, but it makes our lives a lot easier from a maintenance standpoint.
https://github.com/instructlab/eval/pull/1
Note: Current requirements will overwrite existing nvidia pytorch installs. Need to ensure that if those exist, we are not installing our own torch.
Thank you for the review @JamesKunstle, I've created an issue about your quantization comment here: https://github.com/instructlab/training/issues/29
just to be super clear:
squash the first two commits and fix up the commit message
create a follow-up issue for cleaning up requirements.txt to specify versions for all dependencies + making sure the versions are aligned with instructlab/instructlab if it's a shared dependency
create a follow-up issue for cleaning up requirements.txt to specify versions for all dependencies + making sure the versions are aligned with instructlab/instructlab if it's a shared dependency
We talked about this, so I know your intent was to create this issue, but I can't find it anywhere. I went ahead and filed this one: https://github.com/instructlab/training/issues/34
| gharchive/pull-request | 2024-06-11T18:38:49 | 2025-04-01T06:44:31.767620 | {
"authors": [
"Maxusmusti",
"RobotSail",
"russellb"
],
"repo": "instructlab/training",
"url": "https://github.com/instructlab/training/pull/12",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
173308853 | Update umadbro
count=$(($count+1)) results in a loop that never ends.
As $RANDOM gives a number between 0 - 32767 this could result in a quite small fake volume
schweeeetSiOn 26 August 2016 at 16:18 int0x80 notifications@github.com wrote:Merged #1.—You are receiving this because you authored the thread.Reply to this email directly, view it on GitHub, or mute the thread.
Thanks for the diff :)
On Aug 26, 2016 11:55, "SecuritySense" notifications@github.com wrote:
schweeeetSiOn 26 August 2016 at 16:18 int0x80 notifications@github.com
wrote:Merged #1.—You are receiving this because you authored the
thread.Reply to this email directly, view it on GitHub, or mute the
thread.
—
You are receiving this because you modified the open/close state.
Reply to this email directly, view it on GitHub
https://github.com/int0x80/anti-forensics/pull/1#issuecomment-242790223,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABAwqHSVs8qOrAQ2k40g0REaeB8jalL0ks5qjxqYgaJpZM4Jtf24
.
| gharchive/pull-request | 2016-08-25T20:59:37 | 2025-04-01T06:44:31.774274 | {
"authors": [
"SecuritySense",
"int0x80"
],
"repo": "int0x80/anti-forensics",
"url": "https://github.com/int0x80/anti-forensics/pull/1",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
} |
935161593 | 🛑 香港|CN2 is down
In 91aa550, 香港|CN2 ($HKCN2_SITE_URL) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 香港|CN2 is back up in a5ddd59.
| gharchive/issue | 2021-07-01T19:46:47 | 2025-04-01T06:44:31.776635 | {
"authors": [
"Kevincible"
],
"repo": "intSailers/uptime",
"url": "https://github.com/intSailers/uptime/issues/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1707929079 | Support for psr/log ^2.0 || ^3.0
Would it be possible to get the requirement for psr/log updated to a more recent version?
Yeah this is starting to conflict with packages...
Looks like this will also be fixed in https://github.com/intacct/intacct-sdk-php/pull/187
| gharchive/issue | 2023-05-12T16:50:37 | 2025-04-01T06:44:31.778368 | {
"authors": [
"rijnhard",
"stephenmichael"
],
"repo": "intacct/intacct-sdk-php",
"url": "https://github.com/intacct/intacct-sdk-php/issues/189",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
523792023 | "This app isn't verified" warning from google oauth
When I sign in with token='browser', google shows me this
I can eventually make it in, but it isn't very confidence-inspiring. I can't figure out how to verify my app. Should be using a different API key maybe?
Interesting... This hasn't been reported for gcsfs, which has the same mechanism available, but I don't think it gets used that much. I suppose the link would tell us how to get the app verified, but there may bee a good chance that we don't actually meet google's guidelines...
Would you be interested in following the link to verify this package as an app?
I couldn't figure out how to do it.
Also, today we got this same message with gcsfs proper.
Tried starting the process for gcsfs (which is more important). Will repeat or dual-purpose that client if successful.
@rabernat, see this comment, seems like there is a way to get credentials objects now, pandas/pydata blessed.
https://pandas-gbq.readthedocs.io/en/latest/howto/authentication.html#authenticating-with-a-user-account
I am thinking, than gdrive might not be great for throughput, but it's be a great way to share intake catalogues
| gharchive/issue | 2019-11-16T04:52:14 | 2025-04-01T06:44:31.782495 | {
"authors": [
"martindurant",
"rabernat"
],
"repo": "intake/gdrivefs",
"url": "https://github.com/intake/gdrivefs/issues/7",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2187448243 | 🛑 Shopintake is down
In 84478c0, Shopintake (https://www.shopintake.com) was down:
HTTP code: 403
Response time: 446 ms
Resolved: Shopintake is back up in cddb4f6 after 16 minutes.
| gharchive/issue | 2024-03-14T23:30:33 | 2025-04-01T06:44:31.785014 | {
"authors": [
"intakefoods"
],
"repo": "intakefoods/status.intakefoods.kr",
"url": "https://github.com/intakefoods/status.intakefoods.kr/issues/1985",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2287266253 | [PyTorch Upstream] Stock pytorch fp16 e2e test got segmentfault from triton.
We got segmentfault when runing stock pytorch fp16 end2end test. we've narrow down a mini reporducer as follow:
import faulthandler; faulthandler.enable()
import torch
from torch import tensor, device
import torch.fx as fx
from torch._dynamo.testing import rand_strided
from math import inf
import torch._inductor.inductor_prims
import torch._dynamo.config
import torch._inductor.config
import torch._functorch.config
import torch.fx.experimental._config
torch._inductor.config.fallback_random = True
torch._inductor.config.freezing = True
torch._inductor.config.triton.cudagraphs = True
torch._functorch.config.unlift_effect_tokens = True
torch._functorch.config.debug_partitioner = True
isolate_fails_code_str = None
from torch.nn import *
class Repro(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, arg0_1):
isnan = torch.ops.aten.isnan.default(arg0_1); arg0_1 = None
any_1 = torch.ops.aten.any.default(isnan); isnan = None
return (any_1,)
def load_args(reader):
buf0 = reader.storage(None, 2097152, device=device(type='xpu', index=0), dtype_hint=torch.float16)
reader.tensor(buf0, (1, 1024, 1024), dtype=torch.float16, is_leaf=True) # arg0_1
load_args._version = 0
mod = Repro()
if __name__ == '__main__':
from torch._dynamo.repro.after_aot import run_repro
with torch.no_grad():
run_repro(mod, load_args, accuracy=False, command='run', save_dir=None, tracing_mode='real', check_str=None)
We got the callstack as follow:
Current thread 0x00007ffb1fd0c740 (most recent call first):
File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/site-packages/triton/compiler/compiler.py", line 369 in _init_handles
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/runtime/triton_heuristics.py", line 406 in _precompile_config
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/runtime/triton_heuristics.py", line 203 in precompile
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/codecache.py", line 2932 in result
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/codecache.py", line 3128 in wait
File "/tmp/torchinductor_xinanlin/lq/clqdbzfwkffulbfq265wdizbwhp2hb3vvnzqekc75wzbkcniuxem.py", line 77 in <module>
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/runtime/compile_tasks.py", line 44 in _reload_python_module
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/codecache.py", line 2567 in load_by_key_path
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/graph.py", line 1657 in compile_to_module
File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/utils.py", line 273 in time_wrapper
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/graph.py", line 1710 in compile_to_fn
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/compile_fx.py", line 803 in fx_codegen_and_compile
File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/contextlib.py", line 79 in inner
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/compile_fx.py", line 507 in compile_fx_inner
File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/utils.py", line 273 in time_wrapper
File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/contextlib.py", line 79 in inner
File "/home/xinanlin/xinanlin/miniconda3/lib/python3.10/contextlib.py", line 79 in inner
File "/home/xinanlin/xinanlin/pytorch/torch/_inductor/debug.py", line 304 in inner
File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/repro/after_aot.py", line 83 in debug_wrapper
File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/repro/after_aot.py", line 708 in repro_run
File "/home/xinanlin/xinanlin/pytorch/torch/_dynamo/repro/after_aot.py", line 957 in run_repro
File "/home/pt-gpu/4T-4652/xinanlin/pytorch/../test_any.py", line 41 in <module>
To reproduce, please build stock pytorch with env "USE_XPU=1", and run the above script.
@riverliuintel @vlad-penkin this issue blocked Inductor upstream process, please priotize, thanks.
I will investigate and report back.
The corresponding triton kernel:
from ctypes import c_void_p, c_long
import torch
import math
import random
import os
import tempfile
from math import inf, nan
from torch._inductor.hooks import run_intermediate_hooks
from torch._inductor.utils import maybe_profile
from torch._inductor.codegen.memory_planning import _align as align
from torch import device, empty_strided
from torch._inductor.codecache import AsyncCompile
from torch._inductor.select_algorithm import extern_kernels
from torch._inductor.codegen.multi_kernel import MultiKernelCall
aten = torch.ops.aten
inductor_ops = torch.ops.inductor
assert_size_stride = torch._C._dynamo.guards.assert_size_stride
empty_strided_cpu = torch._C._dynamo.guards._empty_strided_cpu
empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda
alloc_from_pool = torch.ops.inductor._alloc_from_pool
reinterpret_tensor = torch.ops.inductor._reinterpret_tensor
async_compile = AsyncCompile()
# kernel path: /tmp/torchinductor_xinanlin/rb/crb7vl4jnylgbdeqtxzpnde4vgafzac4akzqlpwhnozqn3sna272.py
# Source Nodes: [any_1, isnan], Original ATen: [aten.any, aten.isnan]
# any_1 => any_1
# isnan => isnan
triton_red_fused_any_isnan_0 = async_compile.triton('triton_', '''
import triton
import triton.language as tl
from triton.compiler.compiler import AttrsDescriptor
from torch._inductor import triton_helpers, triton_heuristics
from torch._inductor.ir import ReductionHint, TileHint
from torch._inductor.triton_helpers import libdevice, math as tl_math
from torch._inductor.triton_heuristics import AutotuneHint
from torch._inductor.utils import instance_descriptor
@triton_heuristics.reduction(
size_hints=[1, 1048576],
reduction_hint=ReductionHint.DEFAULT,
filename=__file__,
triton_meta={'signature': {0: '*fp16', 1: '*i1', 2: 'i32', 3: 'i32'}, 'device': 0, 'device_type': 'xpu', 'constants': {2: 1}, 'configs': [AttrsDescriptor(
divisible_by_16=(0, 1, 3), equal_to_1=(2,))]},
inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_any_isnan_0', 'mutated_arg_names': [], 'no_x_dim': False, 'backend_hash': '196aef
a0a85bda494694e78ae7765ffc5ac720f56091b59c964159c6f79c4a45'}
)
@triton.jit [19/92120]
def triton_(in_ptr0, out_ptr0, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr):
xnumel = 1
rnumel = 1048576
xoffset = tl.program_id(0) * XBLOCK
xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
xmask = xindex < xnumel
rbase = tl.arange(0, RBLOCK)[None, :]
_tmp3 = tl.full([XBLOCK, RBLOCK], 0, tl.int1)
for roffset in range(0, rnumel, RBLOCK):
rindex = roffset + rbase
rmask = rindex < rnumel
r0 = rindex
tmp0 = tl.load(in_ptr0 + (r0), None, eviction_policy='evict_first').to(tl.float32)
tmp1 = libdevice.isnan(tmp0).to(tl.int1)
tmp2 = tl.broadcast_to(tmp1, [XBLOCK, RBLOCK])
tmp4 = _tmp3 | tmp2
_tmp3 = tmp4
tmp3 = triton_helpers.any(_tmp3.to(tl.int8), 1)[:, None].to(tl.int1)
tl.store(out_ptr0 + (tl.full([XBLOCK, 1], 0, tl.int32)), tmp3, None)
''', device_str='xpu')
import triton
import triton.language as tl
from torch._inductor.triton_heuristics import grid, split_scan_grid, start_graph, end_graph
from torch._C import _xpu_getCurrentRawStream as get_raw_stream
async_compile.wait(globals())
del async_compile
def call(args):
arg0_1, = args
args.clear()
assert_size_stride(arg0_1, (1, 1024, 1024), (1048576, 1024, 1))
with torch.xpu._DeviceGuard(0):
torch.xpu.set_device(0)
buf0 = empty_strided((), (), device='xpu', dtype=torch.bool)
# Source Nodes: [any_1, isnan], Original ATen: [aten.any, aten.isnan]
stream0 = get_raw_stream(0)
triton_red_fused_any_isnan_0.run(arg0_1, buf0, 1, 1048576, grid=grid(1), stream=stream0)
del arg0_1
return (buf0, )
def benchmark_compiled_module(times=10, repeat=10):
from torch._dynamo.testing import rand_strided
from torch._inductor.utils import print_performance
arg0_1 = rand_strided((1, 1024, 1024), (1048576, 1024, 1), device='xpu:0', dtype=torch.float16)
fn = lambda: call([arg0_1])
return print_performance(fn, times=times, repeat=repeat)
if __name__ == "__main__":
from torch._inductor.wrapper_benchmark import compiled_module_main
compiled_module_main('BartForConditionalGeneration', benchmark_compiled_module)
Looks like the crash is in IGC:
Program received signal SIGSEGV, Segmentation fault.
0x00007ffff319d574 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
(gdb) bt
#0 0x00007ffff319d574 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#1 0x00007ffff31a7403 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#2 0x00007ffff31a7ccc in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#3 0x00007ffff3be0aee in llvm::FPPassManager::runOnFunction(llvm::Function&) () from /lib/x86_64-linux-gnu/libigc.so.1
#4 0x00007ffff3be0e14 in llvm::FPPassManager::runOnModule(llvm::Module&) () from /lib/x86_64-linux-gnu/libigc.so.1
#5 0x00007ffff3be1bac in llvm::legacy::PassManagerImpl::run(llvm::Module&) () from /lib/x86_64-linux-gnu/libigc.so.1
#6 0x00007ffff3153b2c in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#7 0x00007ffff2e0c321 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#8 0x00007ffff3050afb in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#9 0x00007ffff2e0e307 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#10 0x00007ffff2e7c4a5 in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#11 0x00007ffff2e7da3e in ?? () from /lib/x86_64-linux-gnu/libigc.so.1
#12 0x00007ffff7efac64 in NEO::OfflineCompiler::buildSourceCode() () from /lib/x86_64-linux-gnu/libocloc.so
#13 0x00007ffff7efe4e5 in NEO::OfflineCompiler::build() () from /lib/x86_64-linux-gnu/libocloc.so
#14 0x00007ffff7f366ce in int SafetyGuardLinux::call<int, NEO::OfflineCompiler, int (NEO::OfflineCompiler::*)()>(NEO::OfflineCompiler*, int (NEO::OfflineCompiler::*)(), int) () from /lib/x86_64-linux-gnu/libocloc.so
#15 0x00007ffff7f363ee in buildWithSafetyGuard(NEO::OfflineCompiler*) () from /lib/x86_64-linux-gnu/libocloc.so
#16 0x00007ffff7ef19c8 in Ocloc::Commands::compile(OclocArgHelper*, std::vector<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::allocator<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > const&) () from /lib/x86_64-linux-gnu/libocloc.so
#17 0x00007ffff7edda35 in oclocInvoke () from /lib/x86_64-linux-gnu/libocloc.so
#18 0x0000555555554787 in main ()
I have the spirv dumps that cause the IGC crash and will submit a ticket to the IGC team.
Hi @alexbaden , can you share the IGC ticket link?
@etaf you've been added as watcher to the IGC ticket and should've receive an email notification.
Hi, @alexbaden @vlad-penkin how can I get the fixed driver?
| gharchive/issue | 2024-05-09T09:18:10 | 2025-04-01T06:44:31.882510 | {
"authors": [
"alexbaden",
"etaf",
"vlad-penkin"
],
"repo": "intel/intel-xpu-backend-for-triton",
"url": "https://github.com/intel/intel-xpu-backend-for-triton/issues/1073",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1499526769 | Fix onnxrt numpy round bug
Signed-off-by: yuwenzho yuwen.zhou@intel.com
Type of Change
bug fix
Description
Defect/regression:
TypeError: type numpy.ndarray doesn't define round method
Profile:
Report: https://inteltf-jenk.sh.intel.com/view/lpot-release/job/intel-lpot-validation-top-release-v2.0/3/artifact/report.html
commit id: ae3cf5638ca96052032f11dca2120dce5c8d0316
log:
https://inteltf-jenk.sh.intel.com/job/intel-lpot-validation-top-release-v2.0/3/artifact/linux/onnxrt/mobilenetv2-12/onnxrt-mobilenetv2-12-linux-clx8280-tune.log
Framework: onnx
Framework version: 1.13.1
JIRA ticket: ILITV-2528
How has this PR been tested?
extension test: mobilenetv2-12
Dependency Change?
NO
extension test: https://inteltf-jenk.sh.intel.com/job/intel-lpot-validation-top-mr-extension/3786/
| gharchive/pull-request | 2022-12-16T03:45:24 | 2025-04-01T06:44:31.905234 | {
"authors": [
"yuwenzho"
],
"repo": "intel/neural-compressor",
"url": "https://github.com/intel/neural-compressor/pull/287",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
897430483 | Support parsing empty records in patterns
Issue
The following valid Elm code is not being properly parsed
emptyRecordFn {} = 1
Pull request status
This pull request does two things:
Add failing test cases (in the complete suite)
Change the parser to allow an optional list of valid fields
Note
Using the "Add type annotation" intention for a {} = 1 now generates a : { a | } -> number as a type annotation. The compiler itself has a similar problem, which I have pointed at in https://github.com/elm/compiler/issues/2204. I think this is an okay problem to have, especially since there isn't any valid type that could be infered here.
LGTM! I'm taking over steward ship of this project and will merge this for the next release.
@jfmengels Could you merge master into that branch? I fixed the CI :)
| gharchive/pull-request | 2021-05-20T20:41:39 | 2025-04-01T06:44:31.935666 | {
"authors": [
"cies",
"jfmengels"
],
"repo": "intellij-elm/intellij-elm",
"url": "https://github.com/intellij-elm/intellij-elm/pull/754",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
243238460 | Feature/prepare maven central
This PR updates the gradle build file to allow for code signing and artifact deploy to Sonatype, which mirrors out to Maven Central.
Additionally, there are minor formatting and Javadoc changes.
This PR is a good first step - it allows me to manually publish SNAPSHOTS and releases, but I'd like to get CircleCI at least publishing SNAPSHOTS when the development branch is built. Unfortunately, I am unable to get gradle to read the environment variables from CircleCI, so that's not yet working -- I plan to spend some more time on that this weekend.
Just pushed one more change for the naming - let me know what you think @adrianhopebailie
| gharchive/pull-request | 2017-07-16T13:08:00 | 2025-04-01T06:44:31.991146 | {
"authors": [
"sappenin"
],
"repo": "interledger/java-ilp-core",
"url": "https://github.com/interledger/java-ilp-core/pull/55",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1780814448 | 1505/filter grants
Changes proposed in this pull request
Context
Checklist
[ ] Related issues linked using fixes #number
[ ] Tests added/updated
[ ] Documentation added
[ ] Make sure that all checks pass
Accidentally opened PR
| gharchive/pull-request | 2023-06-29T13:23:46 | 2025-04-01T06:44:31.993610 | {
"authors": [
"Tymmmy"
],
"repo": "interledger/rafiki",
"url": "https://github.com/interledger/rafiki/pull/1532",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
505821780 | Set up autopep8 as a task and add guidelines to the readme
Since we have the pep8speaks bot commenting on our issues, we ought to make it easy for people to format things in a way that makes pep8 happy. This project looks like it would do it: https://pypi.org/project/autopep8/
task
add autopep8 to requirements.txt
add instructions to the readme file telling people how to run it before making PRs.
if you want to pick up this task
Take a look at the InterMine contributing guidelines
Comment on this issue stating that you intend to work on the task
When you're ready, add your work to the repo and create a pull request.
What to do if you need help
Mention @yochannah, tweet @yoyehudi, pop by to say hi on chat or if needed email yo@intermine.org. Don't forget we're usually only available during uk office hours and will not be able to respond at other times :)
Hi
I would like to work on this task
@barhenkro fantastic - please go ahead, and let us know if you have any questions! :)
fixed by #52 - thank you so much @barhenkro !
| gharchive/issue | 2019-10-11T12:26:11 | 2025-04-01T06:44:32.007177 | {
"authors": [
"barhenkro",
"yochannah"
],
"repo": "intermine/intermine-ws-python",
"url": "https://github.com/intermine/intermine-ws-python/issues/51",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
143210281 | Events from 27 to 31 of march
Hi, I have an issue with events at the end of march, but not of all years and only in "month" view.
Starting from the input below, I had a wrong result where I put a "-" sign, and a good result where I put a "+" sign:
- "start":new Date(2011, 2, 28),"end":new Date(2011, 2, 28)
- "start":new Date(2012, 2, 28),"end":new Date(2012, 2, 28)
+ "start":new Date(2013, 2, 28),"end":new Date(2013, 2, 28)
+ "start":new Date(2014, 2, 28),"end":new Date(2014, 2, 28)
+ "start":new Date(2015, 2, 28),"end":new Date(2015, 2, 28)
- "start":new Date(2016, 2, 28),"end":new Date(2016, 2, 28)
- "start":new Date(2017, 2, 28),"end":new Date(2017, 2, 28)
- "start":new Date(2018, 2, 28),"end":new Date(2018, 2, 28)
+ "start":new Date(2019, 2, 28),"end":new Date(2019, 2, 28)
+ "start":new Date(2020, 2, 28),"end":new Date(2020, 2, 28)
+ "start":new Date(2021, 2, 28),"end":new Date(2021, 2, 28)
What happens?
Thanks in advance,
Davide
Thanks to solve this issue so fast, I tried the fix on my locale machine.
When did you think to release this fix officially on NPM?
thanks, Davide
| gharchive/issue | 2016-03-24T10:31:05 | 2025-04-01T06:44:32.119245 | {
"authors": [
"jesty"
],
"repo": "intljusticemission/react-big-calendar",
"url": "https://github.com/intljusticemission/react-big-calendar/issues/62",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
760507333 | Add boundserror when indexing into an invalid column
Fix #211
@c42f Is this a good enough error or do I need something better here? I wanted to use BoundsError but it's quite limited for display.
I think throwing any exception is a great improvement already, so I'd be happy with this.
Having said that, I think this error would be a bit mysterious to users as they're dealing with the column name in using getproperty, but the error emitted here has no mention of the column name:
julia> row = execute(conn, "select * from (values (1, 2), (3, 4)) as test (a,b)") |> first
LibPQ.Row(PostgreSQL result, 1)
julia> row.a
1
julia> row.x
ERROR: BoundsError: attempt to access 2-element Array{String,1} at index [0]
Stacktrace:
[1] column_number at /home/ubuntu/.julia/dev/LibPQ/src/results.jl:481 [inlined]
[2] getindex(::LibPQ.Result, ::Int64, ::Int64) at /home/ubuntu/.julia/dev/LibPQ/src/results.jl:509
[3] getproperty(::LibPQ.Row, ::Symbol) at /home/ubuntu/.julia/dev/LibPQ/src/tables.jl:37
[4] top-level scope at REPL[35]:1
A nice extension could be to add some @boundscheck logic to getproperty as well. I'm not sure what error type to throw though. It seems that the error type used for Base when accessing a nonexistent field of a type is ErrorException... which is nice and descriptive but nonspecific in the type.
| gharchive/pull-request | 2020-12-09T17:05:36 | 2025-04-01T06:44:32.147614 | {
"authors": [
"c42f",
"iamed2"
],
"repo": "invenia/LibPQ.jl",
"url": "https://github.com/invenia/LibPQ.jl/pull/212",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
624947694 | Cache artifacts in Travis, AppVeyor, GitHubActions, and Cirrus
Closes #166
I've tested Travis, AppVeyor, and GHA, but I couldn't get Cirrus to pick up on my repo for some reason. According to the docs it should work.
GitLab doesn't allow caching files outside of the repository directory, so caching can't be used there. DroneCI's caching looks a bit more complicated so maybe I'll look at it another day, it's not used frequently anyways.
Awesome, thank you so much!
Cirrus is being very unreliable lately, many packages moved to Travis for FreeBSD testing. As far as I know, Cloud Drone doesn't support caching, there are plugins for self-hosted instances.
How does FreeBSD on Travis work? And good to know that the Drone stuff will only work on self-hosted, so not much use implementing it here.
How does FreeBSD on Travis work?
Pretty good, just add
- freebsd
to the list of operating systems, that's it
It seems that it's not mentioned in the Travis docs anywhere...
I think it's in beta or something like that, but Alex shared the news on Slack :slightly_smiling_face:
Ah, cool. I'll wait until it's GA before I add the option here.
| gharchive/pull-request | 2020-05-26T14:55:23 | 2025-04-01T06:44:32.151245 | {
"authors": [
"christopher-dG",
"giordano"
],
"repo": "invenia/PkgTemplates.jl",
"url": "https://github.com/invenia/PkgTemplates.jl/pull/173",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
708265063 | (Optional?) randomized ports for isolated services
Is your feature request related to a problem? Please describe.
Imagine the following scenario:
I'm running my separate Docker services for some other instance/application (e.g. Zenodo local development). These services are exposing standard ports, i.e. ES on 9200, PostgreSQL on 5432, etc.
I take a break to install a separate Invenio module, e.g. invenio-records-rest to make a bugfix.
I run invenio-records-rest's tests, but guess what docker-services-cli will have to either:
Fail hard and exit, since it will try to bring up services that have already bound ports (the ones from my Zenodo local dev Docker services). This is actually a good/preferred behavior
Partially/silently fail and re-use my existing Docker services. This is pretty bad, since as part of tests it might end up modifying/wiping the services, which affect my Zenodo local development setup.
Describe the solution you'd like
I think checking if the current implementation of docker-services-cli fails hard (and possibly with an informative message) in the above use-case is the bare minimum to make sure things are safe.
An alternative, more "sophisticated" solution, would be to bring up each Docker service with a randomized available port (preferably in a higher port range, e.g. 50000+), that can then be exposed to the underlying tests, e.g. via environment variables. A downside of the randomized ports solution is that if one wants mid-test to inspect one of the Docker services state, they have to look up what is the current randomized port and modify their connection parameters accordingly.
Another alternative is to standardized these "testing" ports, so that we avoid clashes with other local development ports. E.g. we put ES always on 59200, PostgreSQL always on 55432, etc.
To summarize here are the available solutions:
[ ] Check that we currently fail hard...
[ ] ...with an informative message
[ ] Randomized ports, exposed in the output as environment variable
[ ] Decide and use alternative "standardized" ports
Closed via #39
| gharchive/issue | 2020-09-24T15:23:20 | 2025-04-01T06:44:32.156414 | {
"authors": [
"slint"
],
"repo": "inveniosoftware/docker-services-cli",
"url": "https://github.com/inveniosoftware/docker-services-cli/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2076644017 | anonymization: fix orders with patron_pid
fixes an error when anonymizing an order that does not have the patron_pid in the order lines
Closing in favour of: https://github.com/inveniosoftware/invenio-app-ils/pull/1193
| gharchive/pull-request | 2024-01-11T13:23:05 | 2025-04-01T06:44:32.160100 | {
"authors": [
"ntarocco"
],
"repo": "inveniosoftware/invenio-app-ils",
"url": "https://github.com/inveniosoftware/invenio-app-ils/pull/1192",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1119786768 | membership: invite user to community
closes #https://github.com/inveniosoftware/invenio-communities/issues/379
This is a tracer bullet PR. There are a lot of TODOs listed on the board to complete next. I've highlighted areas where feedback would be good.
unfortunately, i don't know about the alembic error...
i just re-ran the checks on master, and they fail too (with a slightly different error, though), which sounds a bit like a files-related module to me:
AssertionError: assert not [('remove_index', Index('ix_uq_partial_files_object_is_head', Column('bucket_id', UUID(), table=<files_object>, nullable=False), Column('key', TEXT(), table=<files_object>, nullable=False)))]
The same
assert not ext.alembic.compare_metadata()
E AssertionError: assert not [('remove_index', Index('ix_uq_partial_files_object_is_head', Column('bucket_id', UUID(), table=<files_object>, nullable=False), Column('key', TEXT(), table=<files_object>, nullable=False)))]
error that is affecting the master branch, invenio-rdm-records PRs and basically every repository using invenio-files-rest is affecting this PR too. All the tests pass locally, but I will feel better when this is solved prior to merging. Some of the remarks I couldn't quite address anyway, so it might be worth taking another look.
Not further points than the ones already discussed.
| gharchive/pull-request | 2022-01-31T18:29:31 | 2025-04-01T06:44:32.163118 | {
"authors": [
"fenekku",
"lnielsen",
"max-moser"
],
"repo": "inveniosoftware/invenio-communities",
"url": "https://github.com/inveniosoftware/invenio-communities/pull/380",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
942419788 | subjects: account for pre-existing vocabularies when loading [+]
depends on https://github.com/inveniosoftware/invenio-vocabularies/pull/67
integrate new separate subjects vocabulary (invenio-vocabularies 0.7.8)
rely on database content before loading vocabulary
closes #686 (and #687 for now)
I would actually agree that most of the vocabulary loading would make sense in invenio-vocabularies. It wasn't moved there because:
a- Initial loading was in here, so just continued from it
b- Dealing with with default vocabularies if the loading logic is placed in invenio-vocabularies is a complex lift in itself best done separately. (more below)
Placing the loading responsibility in invenio-vocabularies makes sense, but placing the default vocabularies values in invenio-vocabularies doesn't (at least for me). It's the schema defining package (invenio-rdm-records) that sets out those default values because it knows the kind of records it's dealing with (e.g. InvenioRDM has these default resource types, but other users of invenio-vocabularies may not even have resource types). So then we would need another mechanism for invenio-vocabularies to use things defined in rdm-records without importing it (circular dependency). Entrypoints or a registry-like pattern like seen above could be it. But those are both complex and I already spent my complexity points in the loading code for this PR.
Placing the loading responsibility in invenio-vocabularies makes sense, but placing the default vocabularies values in invenio-
vocabularies doesn't (at least for me)
Agree completely
So then we would need another mechanism for invenio-vocabularies to use things defined in rdm-records without importing it (circular dependency)
I see... Need to check but if the vocabularies.yaml is made "generic" (sort of as it is now already) it can be used as a point of entry for vocabularies. Then rdm-records would define the defaults in its module. And the mechanics to get from entry points is already coded.
| gharchive/pull-request | 2021-07-12T20:26:57 | 2025-04-01T06:44:32.167949 | {
"authors": [
"fenekku",
"ppanero"
],
"repo": "inveniosoftware/invenio-rdm-records",
"url": "https://github.com/inveniosoftware/invenio-rdm-records/pull/734",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
407309540 | invenio-s3 repeats the bucket name as path
the name of the bucket from the S3 configuration gets somehow repeated as part of path, so that files are written to s3://bucket_name/bucket_name/files
I am no longer able to reproduce the issue. It seems that the files are now in the correct location.
When using the URI s3://default/f7/14/adaa-04f2-48e8-9a6b-de2f57948ce1/data stored in files_files.uri, the following file structure is created:
$ tree data
data
└── default
└── f7
└── 14
└── adaa-04f2-48e8-9a6b-de2f57948ce1
└── data
@Herrner, please feel free to reopen the issue if you are still able to replicate it.
| gharchive/issue | 2019-02-06T16:07:57 | 2025-04-01T06:44:32.169915 | {
"authors": [
"Herrner",
"egabancho"
],
"repo": "inveniosoftware/invenio-s3",
"url": "https://github.com/inveniosoftware/invenio-s3/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2098551393 | dependency: added tinymce
:heart: Thank you for your contribution!
Description
Please describe briefly your pull request.
Checklist
Ticks in all boxes and 🟢 on all GitHub actions status checks are required to merge:
[ ] I'm aware of the code of conduct.
[ ] I've created logical separate commits and followed the commit message format.
[ ] I've added relevant test cases.
[ ] I've added relevant documentation.
[ ] I've marked translation strings (for relevant code).
[ ] I've followed the CSS/JS and React guidelines (for relevant code).
[ ] I've followed the web accessibility guidelines (for relevant code).
[ ] I've followed the user interface guidelines (for relevant code).
[ ] I've identified the copyright holder(s) and updated copyright headers for touched files (>15 lines contributions).
[ ] I've NOT included third-party code (copy/pasted source code or new dependencies).
Third-party code
If you've added third-party code (copy/pasted or new dependencies), please reach out to an architect.
Reminder
By using GitHub, you have already agreed to the GitHub’s Terms of Service including that:
You license your contribution under the same terms as the current repository’s license.
You agree that you have the right to license your contribution under the current repository’s license.
Could you please check that we don't have any other occurrences in other repositories (except react-invenio-deposit which is deprecated)?
| gharchive/pull-request | 2024-01-24T15:43:14 | 2025-04-01T06:44:32.177955 | {
"authors": [
"0einstein0",
"ptamarit"
],
"repo": "inveniosoftware/invenio-vocabularies",
"url": "https://github.com/inveniosoftware/invenio-vocabularies/pull/289",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2435542475 | 🛑 Google is down
In d912013, Google (https://www.google.com) was down:
HTTP code: 429
Response time: 354 ms
Resolved: Google is back up in 88fa85e after 10 minutes.
| gharchive/issue | 2024-07-29T14:10:17 | 2025-04-01T06:44:32.180595 | {
"authors": [
"peschee"
],
"repo": "inventage/upptime",
"url": "https://github.com/inventage/upptime/issues/392",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1349680732 | docs: Update EAS installation instructions
Thank you for having great setup instructions! I noticed that this part was slightly out of date and no longer required so I updated it for future readers.
Interesting, if the default image now contains it, but the version numbers are going to out of date in the future (r21 will become r22+, ubuntu 20 will go to 22 etc), perhaps there is a more future proof + prescriptive way to word it?
something like:
if you are using EAS default image you are set (do we need to specify EAS greater than or equal to some version or date? I'm not sure how EAS works, if it is all in cloud and everyone is updated, or if there are versions)
if you are not using the EAS image you must specify one that includes JDK11. You may find them (link to Expo docs...)
Then we don't have to worry about versions
Makes sense! I will rework this to make it more future proof.
| gharchive/pull-request | 2022-08-24T15:58:55 | 2025-04-01T06:44:32.200719 | {
"authors": [
"thatzacdavis"
],
"repo": "invertase/notifee",
"url": "https://github.com/invertase/notifee/pull/500",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1472131268 | [enhancement]: Write prompt to a standard PNG EXIF field
Is there an existing issue for this?
[X] I have searched the existing issues
Contact Details
No response
What should this feature add?
We could write the prompt (with no other metadata) to a standard PNG EXIF e.g. Description. Then normal image viewers can at least see it.
Alternatives
No response
Aditional Content
No response
From NemesisFlux @ discord:
Maybe just to help out people like me, a note could be put in the comments field, saying something like, "InvokeAI parameters for this image are stored in tEXt fields in the PNG file. You can examine them with the free third-party program, pngcheck."
Example implementation: https://gitlab.com/keturn/kikifusion/-/blob/890be52914713d33757c5b900b03ad91c735be21/src/qittnz6Lz/metadata.py#L77
ImageIFD.ImageDescription is probably the most widely-supported, but doesn't cope with multi-byte charsets like UTF-8 or UTF-16, so it's not great.
ExifIFD.UserComment does handle unicode and is supported by at least some applications.
I've noticed this request most often from Mac users. Is anyone able to point us to docs on which fields OS X displays in Finder and includes in its search index?
| gharchive/issue | 2022-12-02T00:15:35 | 2025-04-01T06:44:32.286456 | {
"authors": [
"keturn",
"psychedelicious"
],
"repo": "invoke-ai/InvokeAI",
"url": "https://github.com/invoke-ai/InvokeAI/issues/1680",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1481618066 | [bug]: Expected all tensors to be on the same device exception
Is there an existing issue for this?
[X] I have searched the existing issues
OS
Windows
GPU
cuda
VRAM
4GB
What happened?
I can make iterations just perfect in TextToImage but once i try to Outpaint in the Unified editor it throws an error
Model im using is the: Prompthero/openjourney
Once im using the original 1.4 or 1.5 model this does not happen, only happens with certain models from Huggingface.
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking >argument for argument mat1 in method wrapper_addmm)
Full error:
Traceback (most recent call last):
File "d:\ai\invokeai\ldm\generate.py", line 486, in prompt2image
results = generator.generate(
File "d:\ai\invokeai\ldm\invoke\generator\base.py", line 93, in generate
image = make_image(x_T)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "d:\ai\invokeai\ldm\invoke\generator\inpaint.py", line 295, in make_image
samples = sampler.decode(
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "d:\ai\invokeai\ldm\models\diffusion\sampler.py", line 365, in decode
outs = self.p_sample(
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "d:\ai\invokeai\ldm\models\diffusion\ddim.py", line 58, in p_sample
e_t = self.invokeai_diffuser.do_diffusion_step(
File "d:\ai\invokeai\ldm\models\diffusion\shared_invokeai_diffusion.py", line 88, in do_diffusion_step
unconditioned_next_x, conditioned_next_x = self.apply_standard_conditioning(x, sigma, unconditioning, conditioning)
File "d:\ai\invokeai\ldm\models\diffusion\shared_invokeai_diffusion.py", line 104, in apply_standard_conditioning
unconditioned_next_x, conditioned_next_x = self.model_forward_callback(x_twice, sigma_twice,
File "d:\ai\invokeai\ldm\models\diffusion\ddim.py", line 13, in
model_forward_callback = lambda x, sigma, cond: self.model.apply_model(x, sigma, cond))
File "d:\ai\invokeai\ldm\models\diffusion\ddpm.py", line 1441, in apply_model
x_recon = self.model(x_noisy, t, **cond)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl
return forward_call(*input, **kwargs)
File "d:\ai\invokeai\ldm\models\diffusion\ddpm.py", line 2167, in forward
out = self.diffusion_model(x, t, context=cc)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl
return forward_call(*input, **kwargs)
File "d:\ai\invokeai\ldm\modules\diffusionmodules\openaimodel.py", line 798, in forward
emb = self.time_embed(t_emb)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\container.py", line 139, in forward
input = module(input)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\username\anaconda3\envs\invokeai\lib\site-packages\torch\nn\modules\linear.py", line 114, in forward
return F.linear(input, self.weight, self.bias)
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_addmm)
Could not generate image.
Screenshots
No response
Additional context
No response
Contact Details
No response
Problem solved
If you use conda clean it will break InvokeAI.
My bad
Do other models work?
Do other models work?
Seems i was wrong to conclude that, its very weird. when i reinstalled InvokeAI, using the standard 1.5 model i could use all of the features, but when i started on a new second prompt (not unified editor) it gave me this error.
I found some that might shed light on this error, the error seems to stay away when you do not use the --free_gpu_mem flag in the start up. Perhaps the problem resides in this flag code.
This error was fixed in #1938.
This error was fixed in #1938.
My bad, i wasn't aware of this fix yet, thank you letting me know!
| gharchive/issue | 2022-12-07T10:52:16 | 2025-04-01T06:44:32.304156 | {
"authors": [
"TheBarret",
"psychedelicious",
"rmagur1203"
],
"repo": "invoke-ai/InvokeAI",
"url": "https://github.com/invoke-ai/InvokeAI/issues/1843",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1419793408 | Web UI: Fixes same seed reported when with_variations=[] and iterations > 1
Web UI: Fixes same seed reported when with_variations=[] and iterations > 1
Fixes #1223
@ausbitbank can you please try this change out?
I pulled your repository and checked out the fix-incorrect-seed branch , started the api and tested but got the same result still.
request (intercepted and edited to add with_variations: [] using burpsuite)
42["generateImage",{"prompt":"test","iterations":2,"steps":15,"cfg_scale":7.5,"threshold":0,"perlin":0,"height":512,"width":512,"sampler_name":"k_lms","seed":3471630626,"seamless":false,"hires_fix":false,"progress_images":false,"variation_amount":0,"with_variations":[]},false,false]
both responses have the same seed:
42["generationResult",{"url":"outputs/000001.3471630626.png","mtime":1666535971.5529692,"metadata":{"model":"stable diffusion","model_id":null,"model_hash":"fe4efff1e174c627256e44ec2991ba279b3816e364b49f9be2abc0b3ff3f8556","app_id":"lstein/stable-diffusion","app_version":"v1.15","image":{"prompt":[{"prompt":"test","weight":1.0}],"steps":15,"cfg_scale":7.5,"threshold":0,"perlin":0,"height":512,"width":512,"seed":3471630626,"seamless":false,"hires_fix":false,"postprocessing":null,"sampler":"k_lms","variations":[],"type":"txt2img"}}}]
42["generationResult",{"url":"outputs/000002.3471630626.png","mtime":1666535977.6119702,"metadata":{"model":"stable diffusion","model_id":null,"model_hash":"fe4efff1e174c627256e44ec2991ba279b3816e364b49f9be2abc0b3ff3f8556","app_id":"lstein/stable-diffusion","app_version":"v1.15","image":{"prompt":[{"prompt":"test","weight":1.0}],"steps":15,"cfg_scale":7.5,"threshold":0,"perlin":0,"height":512,"width":512,"seed":3471630626,"seamless":false,"hires_fix":false,"postprocessing":null,"sampler":"k_lms","variations":[],"type":"txt2img"}}}]
@psychedelicious should this be merged?
@psychedelicious Is this PR still in need of merging?
@lstein I'm guessing no, since it doesn't solve the original issue of repeating seeds.
Same issue still exists in the latest dev branch, but I would say its very low priority as it does not affect API clients giving the correct parameters as demonstrated by web ui..
| gharchive/pull-request | 2022-10-23T13:21:48 | 2025-04-01T06:44:32.309773 | {
"authors": [
"ausbitbank",
"lstein",
"psychedelicious"
],
"repo": "invoke-ai/InvokeAI",
"url": "https://github.com/invoke-ai/InvokeAI/pull/1224",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2093993635 | name generation of generic types
Is it possible to customise how the names of generic references are generated?
If I have for example the following:
type Foo int
type Bar int
type FooMap[T any] map[Foo]T
type struct FooBar {
FooBar FooMap[Bar]
}
The resulting $ref for the FooBar $def is FooMap[github.com/full/path/to/package.Bar].
I'd like to change this to FooMap[Bar].
I wonder if this will be easier once Go exposes the generic child type via reflection, but for now I've been solving this with a recursive namer. I'm not sure what other types like arrays, maps, or slices show up as though.
var genericTypeRegex = regexp.MustCompile(`^(.+)\[(.+)\]$`)
func getGenericName(name string) string {
if matches := genericTypeRegex.FindStringSubmatch(name); matches != nil {
parent := matches[1]
parts := strings.Split(matches[2], ".")
child := parts[len(parts)-1]
return fmt.Sprintf("%s[%s]", parent, getGenericName(child))
}
return name
}
func nameGenerics(r reflect.Type) string {
return getGenericName(r.Name())
}
and then use it like
reflector := jsonschema.Reflector{ Namer: nameGenerics }
s := reflector.Reflect(myType)
If you'd like to add some broken tests for this, it'd be easier to have a look and align expectations. Generics were not available when this library was first built!
| gharchive/issue | 2024-01-22T14:03:59 | 2025-04-01T06:44:32.312890 | {
"authors": [
"joshstrohminger",
"olistrik",
"samlown"
],
"repo": "invopop/jsonschema",
"url": "https://github.com/invopop/jsonschema/issues/131",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2434009595 | 🛑 Bitwage API (Sandbox) is down
In 8c7d3ce, Bitwage API (Sandbox) (https://api.sandbox.bitwage.com) was down:
HTTP code: 403
Response time: 62 ms
Resolved: Bitwage API (Sandbox) is back up in 1be1cf9 after 36 minutes.
| gharchive/issue | 2024-07-28T16:12:41 | 2025-04-01T06:44:32.315602 | {
"authors": [
"joelinzy"
],
"repo": "inwage/status_page",
"url": "https://github.com/inwage/status_page/issues/3490",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1073492931 | 🛑 bitwage api production is down
In e4e12bd, bitwage api production (https://api2.bitwage.com) was down:
HTTP code: 500
Response time: 547 ms
Resolved: bitwage api production is back up in 7aea99c.
| gharchive/issue | 2021-12-07T15:34:16 | 2025-04-01T06:44:32.318183 | {
"authors": [
"joelinzy"
],
"repo": "inwage/status_page",
"url": "https://github.com/inwage/status_page/issues/36",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2549119731 | 🛑 Bitwage API (Sandbox) is down
In 921278a, Bitwage API (Sandbox) (https://api.sandbox.bitwage.com) was down:
HTTP code: 403
Response time: 139 ms
Resolved: Bitwage API (Sandbox) is back up in 922e67c after 17 minutes.
| gharchive/issue | 2024-09-25T22:43:36 | 2025-04-01T06:44:32.320843 | {
"authors": [
"joelinzy"
],
"repo": "inwage/status_page",
"url": "https://github.com/inwage/status_page/issues/5330",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2634121882 | 🛑 Bitwage API (Sandbox) is down
In bbefb6f, Bitwage API (Sandbox) (https://api.sandbox.bitwage.com) was down:
HTTP code: 403
Response time: 117 ms
Resolved: Bitwage API (Sandbox) is back up in 4572de8 after 52 minutes.
| gharchive/issue | 2024-11-04T23:57:23 | 2025-04-01T06:44:32.323273 | {
"authors": [
"joelinzy"
],
"repo": "inwage/status_page",
"url": "https://github.com/inwage/status_page/issues/6517",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
771537289 | Add good-scroll and pixel-scroll to SmoothScrolling EmacsWiki page
https://www.emacswiki.org/emacs/SmoothScrolling
Fixed https://www.emacswiki.org/emacs?action=browse;diff=2;id=SmoothScrolling;revision=56
| gharchive/issue | 2020-12-20T08:06:19 | 2025-04-01T06:44:32.324654 | {
"authors": [
"io12"
],
"repo": "io12/good-scroll.el",
"url": "https://github.com/io12/good-scroll.el/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1357591469 | [QA] Check / limit configurable timer values
All configurable timer values should be checked / limited to 32 bit or less as setTimeout cannot handle timeouts largen then 32bit (signed).
See Telegram posting:
@apollon:
Mal so aus gegebenem Anlass ein tip: prüft ihr Maximalwerte für setTimeout/setInterval?? Da gibts nen Maximalwert und wenn der überschritten wird ändert nodejs den auf 0 (aka sofort)!!!
Siehe auch https://github.com/ioBroker/ioBroker.admin/issues/1664
https://stackoverflow.com/questions/3468607/why-does-settimeout-break-for-large-millisecond-delay-values
Timers should be limitedin config json too
fixed with upcoming v2.1.9
fixed with v2.1.9
| gharchive/issue | 2022-08-31T16:37:04 | 2025-04-01T06:44:32.381426 | {
"authors": [
"mcm1957"
],
"repo": "iobroker-community-adapters/ioBroker.snmp",
"url": "https://github.com/iobroker-community-adapters/ioBroker.snmp/issues/156",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2411577781 | Adapter stürzt ab
Ich habe in regelmäßigen Abständen folgende Meldung:
TypeError: Cannot read properties of undefined (reading 'split') at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) at AdapterClass.emit (node:events:519:28) at AdapterClass.emit (node:domain:488:12) at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:10949:34) at Immediate. (file:///opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:381:37) at processImmediate (node:internal/timers:478:21)
Please fill the form provided when creating a new issue of type bug_:
Describe the bug
A clear and concise description of what the bug is.
To Reproduce
Steps to reproduce the behavior:
Go to '...'
Click on '...'
Scroll down to '....'
See error
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots & Logfiles
If applicable, add screenshots and logfiles to help explain your problem.
Versions:
Adapter version:
JS-Controller version:
Node version:
Operating system:
Additional context
Add any other context about the problem here.
Ich habe ebenfalls die Fehlermeldung durch das Upgrade von 3.7.1 auf 3.8.1
`
telegram.1 | 2024-07-16 21:19:21.031 | info | starting. Version 3.7.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19
-- | -- | -- | --
telegram.0 | 2024-07-16 21:19:20.985 | info | starting. Version 3.7.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19
host.IoBroker | 2024-07-16 21:19:18.635 | info | iobroker exit 0
host.IoBroker | 2024-07-16 21:19:17.790 | info | instance system.adapter.telegram.1 started with pid 9710
host.IoBroker | 2024-07-16 21:19:17.701 | info | instance system.adapter.telegram.0 started with pid 9703
host.IoBroker | 2024-07-16 21:19:17.457 | info | "system.adapter.telegram.1" enabled
host.IoBroker | 2024-07-16 21:19:17.320 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.1" is started
host.IoBroker | 2024-07-16 21:19:17.294 | info | "system.adapter.telegram.0" enabled
host.IoBroker | 2024-07-16 21:19:17.283 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.0" is started
host.IoBroker | 2024-07-16 21:19:17.275 | info | iobroker upload [0] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/telegram.png telegram.png image/png
host.IoBroker | 2024-07-16 21:19:17.269 | info | iobroker upload [1] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js.map rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.255 | info | iobroker upload [2] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js rules/static/js/src_bootstrap_jsx.daa4bc6c.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.247 | info | iobroker upload [3] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js.map rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.235 | info | iobroker upload [4] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js rules/static/js/src_ActionTelegram_jsx.606264b1.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.230 | info | iobroker upload [5] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_react_index_js.6114a944.chunk.js.map rules/static/js/node_modules_react_index_js.6114a944.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.225 | info | iobroker upload [6] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_react_index_js.6114a944.chunk.js rules/static/js/node_modules_react_index_js.6114a944.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.220 | info | iobroker upload [7] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js.map rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.217 | info | iobroker upload [8] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js rules/static/js/node_modules_prop-types_index_js.c5b5d86f.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.214 | info | iobroker upload [9] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js.map rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.212 | info | iobroker upload [10] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d271.bbc6ce97.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.209 | info | iobroker upload [11] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js.map rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.206 | info | iobroker upload [12] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js rules/static/js/node_modules_iobroker_adapter-react-v5_node_modules_react_index_js-_9d270.bc9865f0.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.203 | info | iobroker upload [13] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js.map rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.200 | info | iobroker upload [14] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js rules/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-d234ae.c537ddf6.chunk.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.198 | info | iobroker upload [15] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/main.44f1f861.js.map rules/static/js/main.44f1f861.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.195 | info | iobroker upload [16] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/static/js/main.44f1f861.js rules/static/js/main.44f1f861.js application/javascript
host.IoBroker | 2024-07-16 21:19:17.192 | info | iobroker upload [17] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/zh-cn.json rules/i18n/zh-cn.json application/json
host.IoBroker | 2024-07-16 21:19:17.190 | info | iobroker upload [18] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/uk.json rules/i18n/uk.json application/json
host.IoBroker | 2024-07-16 21:19:17.187 | info | iobroker upload [19] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/ru.json rules/i18n/ru.json application/json
host.IoBroker | 2024-07-16 21:19:17.184 | info | iobroker upload [20] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/i18n/pt.json rules/i18n/pt.json application/json
host.IoBroker | 2024-07-16 21:19:17.149 | info | iobroker upload [30] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/rules/asset-manifest.json rules/asset-manifest.json application/json
host.IoBroker | 2024-07-16 21:19:17.129 | info | iobroker upload [40] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/i18n/fr/translations.json i18n/fr/translations.json application/json
host.IoBroker | 2024-07-16 21:19:17.107 | info | iobroker upload [50] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/custom/static/js/node_modules_prop-types_index_js.6da256b6.chunk.js.map custom/static/js/node_modules_prop-types_index_js.6da256b6.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.085 | info | iobroker upload [60] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/custom/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-de23731.0f3598f3.chunk.js.map custom/static/js/node_modules_iobroker_adapter-react-v5_assets_devices_sync_recursive_-node_modules_iobroker_a-de23731.0f3598f3.chunk.js.map application/json
host.IoBroker | 2024-07-16 21:19:17.063 | info | iobroker upload [70] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/custom/i18n/pl.json custom/i18n/pl.json application/json
host.IoBroker | 2024-07-16 21:19:17.044 | info | iobroker upload [79] telegram.admin /opt/iobroker/node_modules/iobroker.telegram/admin/blockly.js blockly.js application/javascript
host.IoBroker | 2024-07-16 21:19:16.565 | info | iobroker Update "system.adapter.telegram.1"
host.IoBroker | 2024-07-16 21:19:16.526 | info | iobroker Update "system.adapter.telegram.0"
host.IoBroker | 2024-07-16 21:19:16.439 | info | iobroker 332 packages are looking for funding run `npm fund` for details
host.IoBroker | 2024-07-16 21:19:16.439 | info | iobroker changed 1 package in 5s
host.IoBroker | 2024-07-16 21:19:12.243 | info | instance system.adapter.telegram.1 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION)
host.IoBroker | 2024-07-16 21:19:12.117 | info | instance system.adapter.telegram.0 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION)
host.IoBroker | 2024-07-16 21:19:12.096 | info | stopInstance system.adapter.telegram.1 killing pid 9569
host.IoBroker | 2024-07-16 21:19:12.033 | info | stopInstance system.adapter.telegram.0 killing pid 9586
telegram.1 | 2024-07-16 21:19:11.572 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason
telegram.1 | 2024-07-16 21:19:11.572 | info | terminating
telegram.0 | 2024-07-16 21:19:11.516 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason
telegram.0 | 2024-07-16 21:19:11.516 | info | terminating
host.IoBroker | 2024-07-16 21:19:11.166 | info | iobroker NPM version: 10.8.1Installing iobroker.telegram@3.7.1... (System call)
host.IoBroker | 2024-07-16 21:19:11.092 | info | stopInstance system.adapter.telegram.1 send kill signal
telegram.1 | 2024-07-16 21:19:11.071 | info | Got terminate signal TERMINATE_YOURSELF
host.IoBroker | 2024-07-16 21:19:11.032 | info | stopInstance system.adapter.telegram.0 send kill signal
host.IoBroker | 2024-07-16 21:19:11.031 | info | stopInstance system.adapter.telegram.1 (force=false, process=true)
host.IoBroker | 2024-07-16 21:19:11.031 | info | "system.adapter.telegram.1" disabled
telegram.0 | 2024-07-16 21:19:11.015 | info | Got terminate signal TERMINATE_YOURSELF
host.IoBroker | 2024-07-16 21:19:11.011 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.1" is stopped.
host.IoBroker | 2024-07-16 21:19:10.935 | info | stopInstance system.adapter.telegram.0 (force=false, process=true)
host.IoBroker | 2024-07-16 21:19:10.934 | info | "system.adapter.telegram.0" disabled
host.IoBroker | 2024-07-16 21:19:10.920 | info | iobroker host.IoBroker Adapter "system.adapter.telegram.0" is stopped.
host.IoBroker | 2024-07-16 21:19:10.892 | info | iobroker Update telegram from @3.8.1 to @3.7.1
host.IoBroker | 2024-07-16 21:19:09.261 | info | iobroker upgrade telegram@3.7.1 --debug
telegram.0 | 2024-07-16 21:10:43.710 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19
host.IoBroker | 2024-07-16 21:10:41.376 | info | instance system.adapter.telegram.0 started with pid 9586
telegram.1 | 2024-07-16 21:10:14.636 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19
host.IoBroker | 2024-07-16 21:10:12.236 | info | instance system.adapter.telegram.1 started with pid 9569
host.IoBroker | 2024-07-16 21:10:11.279 | info | Restart adapter system.adapter.telegram.0 because enabled
host.IoBroker | 2024-07-16 21:10:11.279 | error | instance system.adapter.telegram.0 terminated with code 6 (UNCAUGHT_EXCEPTION)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at processImmediate (node:internal/timers:476:21)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at AdapterClass.emit (node:domain:489:12)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at AdapterClass.emit (node:events:517:28)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52)
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[1]: TypeError: Cannot read properties of undefined (reading 'split')
host.IoBroker | 2024-07-16 21:10:11.279 | error | Caught by controller[0]: This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason:
telegram.0 | 2024-07-16 21:10:10.696 | warn | Terminated (UNCAUGHT_EXCEPTION): Without reason
telegram.0 | 2024-07-16 21:10:10.695 | info | terminating
telegram.0 | 2024-07-16 21:10:10.188 | error | Cannot read properties of undefined (reading 'split')
telegram.0 | 2024-07-16 21:10:10.187 | error | TypeError: Cannot read properties of undefined (reading 'split') at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) at AdapterClass.emit (node:events:517:28) at AdapterClass.emit (node:domain:489:12) at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34) at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37) at processImmediate (node:internal/timers:476:21)
host.IoBroker | 2024-07-16 21:10:10.129 | info | instance system.adapter.telegram.1 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION)
telegram.0 | 2024-07-16 21:10:10.115 | error | unhandled promise rejection: Cannot read properties of undefined (reading 'split')
telegram.0 | 2024-07-16 21:10:10.115 | error | Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch().
telegram.1 | 2024-07-16 21:10:09.554 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason
telegram.1 | 2024-07-16 21:10:09.554 | info | terminating
host.IoBroker | 2024-07-16 21:10:09.147 | info | stopInstance system.adapter.telegram.1 send kill signal
telegram.1 | 2024-07-16 21:10:09.051 | info | Got terminate signal TERMINATE_YOURSELF
host.IoBroker | 2024-07-16 21:10:09.027 | info | stopInstance system.adapter.telegram.1 (force=false, process=true)
telegram.0 | 2024-07-16 21:08:58.618 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19
host.IoBroker | 2024-07-16 21:08:56.292 | info | instance system.adapter.telegram.0 started with pid 9546
host.IoBroker | 2024-07-16 21:08:54.211 | info | instance system.adapter.telegram.0 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION)
telegram.0 | 2024-07-16 21:08:53.636 | info | Terminated (ADAPTER_REQUESTED_TERMINATION): Without reason
telegram.0 | 2024-07-16 21:08:53.635 | info | terminating
host.IoBroker | 2024-07-16 21:08:53.185 | info | stopInstance system.adapter.telegram.0 send kill signal
telegram.0 | 2024-07-16 21:08:53.134 | info | Got terminate signal TERMINATE_YOURSELF
host.IoBroker | 2024-07-16 21:08:53.056 | info | stopInstance system.adapter.telegram.0 (force=false, process=true)
javascript.0 | 2024-07-16 21:08:01.451 | info | script.js.Statusmeldungen.Status_Lichtzählen1: Text: Esstisch
javascript.0 | 2024-07-16 21:06:42.869 | info | script.js.Statusmeldungen.Status_Lichtzählen1: Text: Esstisch ,Kueche
telegram.0 | 2024-07-16 21:01:01.860 | info | starting. Version 3.8.1 in /opt/iobroker/node_modules/iobroker.telegram, node: v18.20.4, js-controller: 5.0.19
host.IoBroker | 2024-07-16 21:00:59.292 | info | instance system.adapter.telegram.0 started with pid 9483
host.IoBroker | 2024-07-16 21:00:30.487 | info | instance system.adapter.mqtt.0 started with pid 9464
host.IoBroker | 2024-07-16 21:00:29.178 | info | Restart adapter system.adapter.telegram.0 because enabled
host.IoBroker | 2024-07-16 21:00:29.178 | error | instance system.adapter.telegram.0 terminated with code 6 (UNCAUGHT_EXCEPTION)
host.IoBroker | 2024-07-16 21:00:29.178 | error | Caught by controller[0]: at processImmediate (node:internal/timers:476:21)
host.IoBroker | 2024-07-16 21:00:29.178 | error | Caught by controller[0]: at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37)
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34)
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at AdapterClass.emit (node:domain:489:12)
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at AdapterClass.emit (node:events:517:28)
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17)
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52)
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: TypeError: Cannot read properties of undefined (reading 'split')
host.IoBroker | 2024-07-16 21:00:29.177 | error | Caught by controller[0]: This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). The promise rejected with the reason:
host.IoBroker | 2024-07-16 21:00:28.606 | info | iobroker exit 0
telegram.0 | 2024-07-16 21:00:28.600 | warn | Terminated (UNCAUGHT_EXCEPTION): Without reason
telegram.0 | 2024-07-16 21:00:28.599 | info | terminating
telegram.0 | 2024-07-16 21:00:28.014 | error | Cannot read properties of undefined (reading 'split')
telegram.0 | 2024-07-16 21:00:28.014 | error | TypeError: Cannot read properties of undefined (reading 'split') at processMessage (/opt/iobroker/node_modules/iobroker.telegram/main.js:1191:52) at AdapterClass. (/opt/iobroker/node_modules/iobroker.telegram/main.js:156:17) at AdapterClass.emit (node:events:517:28) at AdapterClass.emit (node:domain:489:12) at change (/opt/iobroker/node_modules/@iobroker/js-controller-adapter/src/lib/adapter/adapter.ts:11114:34) at Immediate. (/opt/iobroker/node_modules/@iobroker/db-states-redis/src/lib/states/statesInRedisClient.ts:379:37) at processImmediate (node:internal/timers:476:21)
host.IoBroker | 2024-07-16 21:00:27.942 | info | instance system.adapter.mqtt.0 terminated with code 11 (ADAPTER_REQUESTED_TERMINATION)
telegram.0 | 2024-07-16 21:00:27.928 | error | unhandled promise rejection: Cannot read properties of undefined (reading 'split')
telegram.0 | 2024-07-16 21:00:27.927 | error | Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch().
`
`CPU: 40 %
RAM: 40 %
Betriebszeit: 18h
Verfügbar: 6.0.8
Installiert: 5.0.19
Ereignisse: ⇥45 / ↦13
Plattform: docker (official image - v9.1.2)
Betriebssystem: linux
Architektur: x64
CPUs: 16
Geschwindigkeit: 3695 MHz
Modell: AMD Ryzen 9 6900HX with Radeon Graphics
RAM: 30.6 GB
System-Betriebszeit: 9 T. 10:00:11
Node.js: v18.20.4
time: 1721156190271
timeOffset: -120
NPM: 10.8.1
Adapter-Anzahl: 624
Datenträgergröße: 953.9 GB
Freier Festplattenspeicher: 642.7 GB
Aktive Instanzen: 45
Pfad: /opt/iobroker/
Betriebszeit: 17:54:47`
Please retest eith newest release (3.8.2)
If problem still exidzs please add complete version infos
Adapter
Admin
Js-controller
Node
O/S
Mit der neuen Version tritt es nicht mehr auf
Sounds great that 3.8.2 solves the problem.
I got the same problem with 3.8.1 and rolled back to 3.8.0 which workes fine for me.
I use the beta repository and there is no 3.8.2. Where may I find this to test if it fixes the problem in my environment?
Cheers, Thomas
So I will close here.
Please open a new issue if problem reoccures
| gharchive/issue | 2024-07-16T16:21:11 | 2025-04-01T06:44:32.395564 | {
"authors": [
"crunchip77",
"fuchs-1978",
"mcm1957",
"tmaul1968"
],
"repo": "iobroker-community-adapters/ioBroker.telegram",
"url": "https://github.com/iobroker-community-adapters/ioBroker.telegram/issues/841",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2462876699 | Docs incorrect, scripts desctructive, missing options....
Hi guys,
I am sorry but what is even going on with this plugin?
docs ask us to npm install --save-dev @capacitor/assets but then a few lines later we are using npx so we just added it to our package.json it for no reason?
Why does it always delete icons under src/assets/icons? I have never asked it to delete these and there is no option not to? Why be destructive? Just leave them there, they don't hurt nobody, or ask me if I want them deleted. These are in the damn assets folder, a folder that we might use within our app.
Why is there no option to generate notification icons? We need a post-generate script that copies an icon to res/drawable because by itself, there is only a splash screen in there.
Please guys, get your shit together and release something comprehensive, this is pretty bad.
@mchl18
This allows for controlling the version of the tool. Which you should do.
The other 2 are valid points, but your tone is not.
excuse my tone, there was general frustration in working with capacitor that lead up to my tone, I will try and get a PR up to fix the things I mentioned
Will npx not always use whatever it can get from the registry? Or will it prefer the package.json version?
| gharchive/issue | 2024-08-13T09:34:43 | 2025-04-01T06:44:32.451537 | {
"authors": [
"marklagendijk",
"mchl18"
],
"repo": "ionic-team/capacitor-assets",
"url": "https://github.com/ionic-team/capacitor-assets/issues/618",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
355323777 | [4.0.0-beta.5] TypeScript compiler error when running ionic serve
Bug Report
Ionic Info
Ionic:
ionic (Ionic CLI) : 4.1.1 (/home/alex/.nvm/versions/node/v10.9.0/lib/node_modules/ionic)
Ionic Framework : @ionic/angular 4.0.0-beta.5
@angular-devkit/core : 0.7.5
@angular-devkit/schematics : 0.7.5
@angular/cli : 6.1.5
@ionic/ng-toolkit : 1.0.7
@ionic/schematics-angular : 1.0.5
Cordova:
cordova (Cordova CLI) : 8.0.0
Cordova Platforms : none
Cordova Plugins : no whitelisted plugins (1 plugins total)
System:
NodeJS : v10.9.0 (/home/alex/.nvm/versions/node/v10.9.0/bin/node)
npm : 6.2.0
OS : Linux 4.18
Describe the Bug
Typescript compiler errors when running ionic serve.
[ng] ℹ 「wdm」: Compiling...
[ng] ERROR in node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,35): error TS1005: ';' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,56): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,60): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(3,72): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,35): error TS1005: ';' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,55): error TS1005: ';' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,82): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,86): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,104): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,116): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,147): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,151): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,169): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,181): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,193): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,230): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,234): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,252): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,264): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,276): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,288): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,325): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,329): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,360): error TS1005: ',' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,398): error TS1005: '(' expected.
[ng] node_modules/@ionic/angular/dist/directives/proxies.d.ts(4,401): error TS1005: ')' expected.
[ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,42): error TS1005: ';' expected.
[ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,101): error TS1005: ']' expected.
[ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,109): error TS1005: ',' expected.
[ng] node_modules/@ionic/core/dist/types/interface.d.ts(41,110): error TS1136: Property assignment expected.
[ng] Date: 2018-08-29T20:30:04.848Z - Hash: a7760a7286c2a29edc22 - Time: 972ms
[ng] 142 unchanged chunks
[ng] ℹ 「wdm」: Compiled successfully.
Steps to Reproduce
My package.json:
{
"name": "nikkoApp",
"version": "0.0.1",
"author": "Ionic Framework",
"homepage": "http://ionicframework.com/",
"scripts": {
"ng": "ng",
"start": "ng serve",
"build": "ng build",
"test": "ng test",
"lint": "ng lint",
"e2e": "ng e2e"
},
"private": true,
"dependencies": {
"@angular/common": "~6.1.1",
"@angular/core": "~6.1.1",
"@angular/forms": "~6.1.1",
"@angular/http": "~6.1.1",
"@angular/platform-browser": "~6.1.1",
"@angular/platform-browser-dynamic": "~6.1.1",
"@angular/router": "~6.1.1",
"@ionic-native/core": "^5.0.0-beta.15",
"@ionic-native/native-keyboard": "^5.0.0-beta.15",
"@ionic-native/splash-screen": "5.0.0-beta.14",
"@ionic-native/status-bar": "5.0.0-beta.14",
"@ionic/angular": "^4.0.0-beta.5",
"@ionic/pro": "2.0.3",
"core-js": "^2.5.3",
"rxjs": "6.2.2",
"zone.js": "^0.8.26"
},
"devDependencies": {
"@angular/cli": "~6.1.1",
"@angular/compiler": "~6.1.1",
"@angular/compiler-cli": "~6.1.1",
"@angular/language-service": "~6.1.1",
"@angular-devkit/architect": "~0.7.2",
"@angular-devkit/build-angular": "~0.7.2",
"@angular-devkit/core": "~0.7.2",
"@angular-devkit/schematics": "~0.7.2",
"@ionic/ng-toolkit": "^1.0.0",
"@ionic/schematics-angular": "^1.0.0",
"@types/jasmine": "~2.8.6",
"@types/jasminewd2": "~2.0.3",
"@types/node": "~10.7.1",
"codelyzer": "~4.4.2",
"jasmine-core": "~2.99.1",
"jasmine-spec-reporter": "~4.2.1",
"karma": "~3.0.0",
"karma-chrome-launcher": "~2.2.0",
"karma-coverage-istanbul-reporter": "~2.0.0",
"karma-jasmine": "~1.1.1",
"karma-jasmine-html-reporter": "^0.2.2",
"protractor": "~5.4.0",
"ts-node": "~7.0.0",
"tslint": "~5.11.0",
"typescript": "2.9.2"
},
"description": "An Ionic project",
"cordova": {
"plugins": {
"cordova-plugin-native-keyboard": {}
}
}
}
Additional Context
Probably related to #15360
To make it clear, bumping typescript version to 2.9.2 did not help.
Nevermind, restarting ionic serve solved the problem. Apparently, it doesn't reload node modules automatically.
| gharchive/issue | 2018-08-29T20:38:15 | 2025-04-01T06:44:32.528244 | {
"authors": [
"yursha"
],
"repo": "ionic-team/ionic",
"url": "https://github.com/ionic-team/ionic/issues/15379",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
437714118 | bug: ion-datetime not bind as ISO-8601 format when using pickerOptions
Bug Report
Ionic version:
[x] 4.x
Current behavior:
This bug is basically the same as ISSUE:15408, but my issue applies to when you use custom pickerOptions. I need to be able to clear an ion-datetime because it isn't required, so I would add this option by using pickerOptions. But when I use picker options, the date isn't in the format of ISO-8601 anymore, but rather the weird object form.
Expected behavior:
What I would expect is that when you use pickerOptions, the format would still be in the ISO-8601 format like when you are not using pickerOptions on a datetime, but this isn't the case for me.
Steps to reproduce:
`
My pickerOptions:
public customStartDate: any = {
buttons: [
{
text: 'Clear',
handler: () => this.eventForm.controls['startDate'].setValue(null)
},
{
text: 'Done',
handler: (data: string) => {
console.log("date data:", data);
this.eventForm.controls['startDate'].setValue(data);
}
}
]
}
public customStartTime: any = {
buttons: [
{
text: 'Clear',
handler: () => this.eventForm.controls['startTime'].setValue(null)
},
{
text: 'Done',
handler: (data: Date) => {
this.eventForm.controls['startTime'].setValue(data);
}
}
]
};
`
Replacing the "this.eventForm.controls..." part with a console.log(data) would be enough te be able to see the issue when you somehow trigger it.
Hi there,
Thanks for the issue. To get around JS Date's implementation of timezones, we use our own object for storing the datetime data, which is what you are seeing. The value that is emitted from the component is valid ISO-8601, but we use that custom object internally.
To get the proper ISO-8601 string, you will want to listen for the ionChange event.
So you would end up doing something like: <ion-datetime (ionChange)="updateDate($event, formGroup)"></ion-datetime>
Does this resolve your issue?
Thanks!
Hi there,
Thanks for the issue. To get around JS Date's implementation of timezones, we use our own object for storing the datetime data, which is what you are seeing. The value that is emitted from the component is valid ISO-8601, but we use that custom object internally.
To get the proper ISO-8601 string, you will want to listen for the ionChange event.
So you would end up doing something like: <ion-datetime (ionChange)="updateDate($event, formGroup)"></ion-datetime>
Does this resolve your issue?
Thanks!
I don't really see how (ionChange)="updateDate($event, formGroup)" would be able to help me. What is the difference between using (ionChange) and the handler inside the pickerOption? I also don't know what this updateDate($event, formGroup)" needs to contain to be able to work properly.
In my opinion these are things that all should be explained inside the documentation of ionic (4). The documentation contains UI component examples, but it lacks the information on how to implement it properly or how to customize it (not only for ion-datetime). But that's somewhat another topic...
Hi there,
ionChange is the event emitted from the datetime component when the value of the date has been changed (i.e. after the date has been selected and "Done" has been clicked). The result of that is a valid ISO-8601 string. This can be found in the Events section of the ion-datetime docs.
The pickerOptions property is something that is passed through to the ion-picker component, which is what ion-datetime is built on. The property allows you to customize the underlying picker component. Details on this can be found in the Properties section of the ion-picker docs.
All the updateDate function needs to contain is the code you originally has in the Done button handler. In this case $event is an object containing the ISO-8601 date that was selected.
Hopefully that clears things up a bit 🙂 . I am going to close this as it is not an Ionic issue.
For further details on how to implement or customize the datetime and picker components, please see the ion-datetime Docs and the ion-picker Docs.
For further assistance implementing your code, feel free to post on the forums or on our slack.
Thanks!
Thank you for your answer.
$event indeed contains a valid ISO-8601 object when you work with f.e. picker-format="DD MMMM YYYY".
But it doesn't work when you want the time only f.e. picker-format="HH mm ss" (or at least it does not for me):
With pickerOptions:
Without pickerOptions:
Is this really not a bug?
Thanks in advance!
Hi there,
Can you post your code? It's hard to say just seeing the output.
HTML with pickerOptions:
<ion-item>
<ion-label for="startTime" position="stacked">Starttijd</ion-label>
<ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)" [pickerOptions]="customStartTime">
</ion-datetime>
</ion-item>
HTML without pickerOptions:
<ion-item>
<ion-label for="startTime" position="stacked">Starttijd</ion-label>
<ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)">
</ion-datetime>
</ion-item>
TS:
public customStartTime: any = {
buttons: [
{
text: 'Clear',
handler: () => this.eventForm.controls['startTime'].setValue(null)
},
{
text: 'Bewaar',
handler: (data: string) => {
this.eventForm.controls['startTime'].setValue(data);
}
}
]
};
eventForm = new FormGroup({
id: new FormControl('', []),
eventTypeId: new FormControl('', [Validators.required]),
name: new FormControl('', [Validators.required]),
description: new FormControl('', [Validators.required]),
startDate: new FormControl('', [Validators.required]),
startTime: new FormControl({ value: '', disabled: false }),
endDate: new FormControl('', [Validators.required]),
endTime: new FormControl({ value: '', disabled: false }),
isFullDay: new FormControl(false),
location: new FormControl('')
});
updateStartTime($event, eventForm) {
console.log($event);
console.log(eventForm);
}
That is all the code that is relevant.
And you can see the results of $event both with and without pickerOptions above.
If you need more, just ask me 😃 .
Thanks!
HTML with pickerOptions:
<ion-item>
<ion-label for="startTime" position="stacked">Starttijd</ion-label>
<ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)" [pickerOptions]="customStartTime">
</ion-datetime>
</ion-item>
HTML without pickerOptions:
<ion-item>
<ion-label for="startTime" position="stacked">Starttijd</ion-label>
<ion-datetime placeholder="hh:mm:ss" display-format="HH:mm:ss" picker-format="HH mm ss" formControlName="startTime" class="form-control" id="startTime" (ionChange)="updateStartTime($event, eventForm)">
</ion-datetime>
</ion-item>
TS:
public customStartTime: any = {
buttons: [
{
text: 'Clear',
handler: () => this.eventForm.controls['startTime'].setValue(null)
},
{
text: 'Bewaar',
handler: (data: string) => {
this.eventForm.controls['startTime'].setValue(data);
}
}
]
};
eventForm = new FormGroup({
id: new FormControl('', []),
eventTypeId: new FormControl('', [Validators.required]),
name: new FormControl('', [Validators.required]),
description: new FormControl('', [Validators.required]),
startDate: new FormControl('', [Validators.required]),
startTime: new FormControl({ value: '', disabled: false }),
endDate: new FormControl('', [Validators.required]),
endTime: new FormControl({ value: '', disabled: false }),
isFullDay: new FormControl(false),
location: new FormControl('')
});
updateStartTime($event, eventForm) {
console.log($event);
console.log(eventForm);
}
That is all the code that is relevant.
And you can see the results of $event both with and without pickerOptions above.
If you need more, just ask me 😃 .
Thanks!
Hi there,
The reason why this is happening is that data in button handler is not a string. As I mentioned before, that is the custom object we use internally to parse dates.
You are setting the value of the datetime to this object, and so the datetime component is just going to emit that object back. The value property expects a valid ISO-8601 string, as stated in the ion-datetime documentation.
ionChange is the event emitted from the datetime component when the value of the date has been changed (i.e. after the date has been selected and "Done" has been clicked). The result of that is a valid ISO-8601 string. This can be found in the Events section of the ion-datetime docs.
The reason why this is happening is that data in the button handler is not a string. As I mentioned before, that is the custom object we use internally to parse dates.
You are setting the value of the datetime to this object, and so the datetime component is just going to emit that object back. The value property expects a valid ISO-8601 string, as stated in the ion-datetime documentation.
This is creating the problem, ionChange doesn't trigger when you don't update the value of the datetime. So I'm not able to get $event when I don't somehow change something of the datetime object...
I am completely lost. For a date (year, month, day) it works perfectly, but for time (hour, minute, second) it doesn't when I do exactly the same for both except for the picker-format. And as stated in the documentation, HH:mm:ss is a valid ISO-8601 format.
I don't know what else I can say or try to do, would creating a small public GitHub project be interesting for you?
Hi there,
Do you have an example where the date formatting is working properly? I have the time example, and that has the behavior you described, but the date formatting is also returning the custom object.
Thanks!
I'm wrong. I'm so sorry. Date has the same behavior as time. Sorry.
But still, how would I be able to trigger ionChange without using the handler? How am I able to get an ISO-8601 without setting the value to the custom object?
Hi there,
When you select a date in the datetime component and then click "Done", ionChange is triggered. If you want to use a custom handler, you will need to parse the date object yourself.
Ideally, there would be a "clear" option built into ion-datetime. There is a feature request open for that functionality: https://github.com/ionic-team/ionic/issues/17482.
Project where everything can be seen:
https://github.com/MatthiasVanMeerbeek-r0670805/Ionic4IonDatetime
(Don't know how to properly link a project to an issue.)
| gharchive/issue | 2019-04-26T14:58:29 | 2025-04-01T06:44:32.555172 | {
"authors": [
"MatthiasVanMeerbeek-r0670805",
"liamdebeasi"
],
"repo": "ionic-team/ionic",
"url": "https://github.com/ionic-team/ionic/issues/18142",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
179801616 | nav-decor always being displayed
Short description of the problem:
The nav-decor is always being displayed with a black background on iOS
What behavior are you expecting?
That the nav-decor doesn't have a black background
This can be re produced using the ionic starter
Which Ionic Version? 2
@alan-agius4 @jgw96 sorry, I don't understand what's the problem
Justin is aware of the problem
On Saturday, 1 October 2016, Manu Mtz.-Almeida notifications@github.com
wrote:
@alan-agius4 https://github.com/alan-agius4 @jgw96
https://github.com/jgw96 sorry, I don't understand what's the problem
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/driftyco/ionic/issues/8252#issuecomment-250862383,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AQv-Wn8fHWO7N64EGRmwFBHcLODD1Jf5ks5qvYaAgaJpZM4KI8kt
.
@jgw96 is fully aware of the problem as I sent him some details over Slack.
Hello just to document the issue a little. The nav-decor component seems to always be displayed on iOS. This causes weird thing such as a small black "block" appearing when navigating on iOS sometimes.
nav-decor always being displayed, hides GoogleMaps despite the css:
ion-app._gmaps_cdv_ .nav-decor{ background: transparent !important; }
The css doesn't seem to get applied, but unchecking and checking any of the style property from the safari inspector applies it and the maps are displayed.
I faced this issue on the iOS emulator but the code behaves as expected on the device.
Can you provide a repo that reproduces this issue?
@manucorporat Link to the repo: https://github.com/nakulgulati/map-test
Also attaching a gif which demonstrates what I said in the previous comment for clarity. :)
Found my issue
I had the below ``` html
<ion-content [attr.class]="options?.cssClass">
rather than
``` html
<ion-content [ngClass]="options?.cssClass">
which resulted in content-ios class not being applied
@nakulgulati Unchecking and checking the style property in the web inspector, will cause a re-paint which is most likely why it's working.
@alan-agius4 Yep that is exactly why it is working. The gif shows that the styles are loaded but not applied on nav-decor and hence causes it to bug.
I tested the map-test:
I works for me...
ion-nav is and it has to be black, the whole time.
What behavior are you expecting?
That the nav-decor doesn't have a black background
This can be re produced using the ionic starter
^this premise is wrong. nav-decor is using during transitions to have a fade to black effect.
I am going to close the issue and please create a new one following the issue template, explaining WHAT is failing in your app, nothing else.
Since this is issue is being hard to reproduce please provide a repo and the environment:
Desktop? device? ios only? WKWebview? iOS Simulator?
@manucorporat Seems fair to close the issue. I was originally facing this issue when using Google Maps in a project and had spun up the test project then. Now the issue seems to have disappeared from the original project and still there in map-test for me.
I would look into in more and like you suggested, would open a new issue if required.
Thanks. :)
This is the author of the cordova-googlemaps-plugin.
As of the google maps plugin 2.0.0-beta2-20170705-1715, the plugin inserts the css automatically.
._gmaps_cdv_ .nav-decor {
display: none !important;
background-color: rgba(0,0,0,0) !important;
background: rgba(0,0,0,0) !important;
}
This thread helps me to solve this issue. Thank you for discussing.
| gharchive/issue | 2016-09-28T15:10:38 | 2025-04-01T06:44:32.569231 | {
"authors": [
"alan-agius4",
"jgw96",
"manucorporat",
"nakulgulati",
"wf9a5m75"
],
"repo": "ionic-team/ionic",
"url": "https://github.com/ionic-team/ionic/issues/8252",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
772979933 | dotnet paket install breaks project parsing
After dotnet paket install project system returns project data contains no references at all. It's fixed by manually running dotnet restore after dotnet paket install has finished.
We should somehow detect that dotnet paket install/update were run and fire dotnet restore automatically.
CC: @forki
paket install / update we could touch a file in /obj - but I think it actually already deletes the project.assets.json
Shouldn't this be enough?
I suspect there's something weird going on with parsing project files for which project.assets.json doesn't exist. I think it's some kind of regression I've introduced when rewriting the project system, not the Paket issue.
Do we already have a repro?
Krzysztof Cieślak notifications@github.com schrieb am Do., 11. Feb. 2021,
15:36:
I suspect there's something weird going on with parsing project files for
which project.assets.json doesn't exist. I think it's some kind of
regression I've introduced when rewriting the project system, not the Paket
issue.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/ionide/ionide-vscode-fsharp/issues/1459#issuecomment-777505196,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AAAOANDTGAS5JJDAKFT3QHLS6PTNTANCNFSM4VFWEOGA
.
| gharchive/issue | 2020-12-22T14:16:46 | 2025-04-01T06:44:32.587865 | {
"authors": [
"Krzysztof-Cieslak",
"forki"
],
"repo": "ionide/ionide-vscode-fsharp",
"url": "https://github.com/ionide/ionide-vscode-fsharp/issues/1459",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
155134379 | Convert from FunScript to Fable
Here comes my first attempt to convert the awesome Ionide plugins to Fable!
I've used your vscode-fable-demo as a basis. Basically the changes are:
For simplicity I've converted the project to a single .fsx file and removed the now unneeded files.
Also for simplicity, I just copied the Helper modules into the script. I'll update the ionide-vscode-helpers later.
I had to update the Fable Node definitions to fix some mistakes.
I'm using the full name to reference a couple of modules that were now conflicting (like Promise).
I've added another helper function (within the activate method) to register commands.
On a quick test, it seems to work :+1: Can you please have a look and tell me what you think? If you want me to make some changes, just tell me.
Cheers!
That looks really good!
Only thing that bothers me is that you completely removed FAKE build scripts - I think I would want to leave parts responsible for versioning, packaging and publishing as they were. Probably whole script should be almost the same, only difference would be calling npm run build instead of old generation.
You're right. I was just trying to find the minimal structure to build the extension with Fable but it seems I went too far ;) I'll put the FAKE script back :+1:
Done! Please check if the changes I did in the build.fsx script make sense. Cheers!
Before we go with it, I would like to convert Helpers project to Fable so we don't have Helpers module here. I guess we can just create Fable branch on Helpers repo and put this code there for start.
Perfect! Please create the branch so I can send the PR there 👍 Also, how do you want to interact with the Helpers from the other packages? Just compiling the code into the final JS or do you want to make a npm package for the Helpers? This could prevent some duplicity of code but for that we will probably have to use F# projects instead of scripts.
@alfonsogarciacaro can you pr onto https://github.com/ionide/ionide-vscode-paket/tree/fable instead?
And Helpers part to https://github.com/ionide/ionide-vscode-helpers/tree/fable. :)
I merged the PR onto this branch https://github.com/ionide/ionide-vscode-paket/tree/fable
| gharchive/pull-request | 2016-05-16T22:13:04 | 2025-04-01T06:44:32.594663 | {
"authors": [
"Krzysztof-Cieslak",
"alfonsogarciacaro",
"cloudRoutine"
],
"repo": "ionide/ionide-vscode-paket",
"url": "https://github.com/ionide/ionide-vscode-paket/pull/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
992908805 | Add support for Firefox for Android
For now, this extension cannot be installed to the stable version of Firefox for Android.
Currently, only a limited number of extensions in our Recommended program are supported on the new Firefox for Android (codename: Fenix) release channel.
https://blog.mozilla.org/addons/2020/02/11/faq-for-extension-support-in-new-firefox-for-android/
Advanced users can install this extension to Firefox for Android Nightly according to:
https://blog.mozilla.org/addons/2020/09/29/expanded-extension-support-in-firefox-for-android-nightly/
However, "Other search engines" and "Subscription" are not currently available in Firefox for Android Nightly because browser.permissions.request, which is the API to dynamically request permissions, is not yet implemented in GeckoView:
mozilla-mobile/fenix#16912
https://bugzilla.mozilla.org/show_bug.cgi?id=1601420
Exceptionally, subscriptions to blocklists hosted on *.githubusercontent.com may be available because permission request for this origin is omitted. re #110
Exceptionally, subscriptions to blocklists hosted on *.githubusercontent.com may be available because permission request for this origin is omitted. re #110
Thanks a lot for this workaround! Would you accept more such domains? I'd like to subscribe to lists on GitLab. Although GitLab itself does not use Access-Control-Allow-Origin: *, some third-party CDN services do:
GitHack: using domain glcdn.githack.com for files on GitLab
Statically: using domain cdn.statically.io
@yan12125
Thank you for your suggestion. You might want to create a pull request. The code to change:
https://github.com/iorate/uBlacklist/blob/bbc2d22a24b04cf04ac4caced1d470876da355d8/src/scripts/options/subscription-section.tsx#L53
Thanks! Let me try it out.
Sync is not available in Firefox for Android (Fenix) now because browser.identity.launchWebAuthFlow is not implemented.
Recently I added support for sync in Safari. Because Safari also lacks browser.identity.launchWebAuthFlow, I implemented an alternative to it. Is that alternative available in Fenix? Currently, no. It requires browser.permissions.request, which is not yet implemented in Fenix.
I think this issue should remain open even without activity?
I added a label that is exempt from stale
According to https://bugzilla.mozilla.org/show_bug.cgi?id=1601420 browser.permissions.request is implemented.
I tried uBlacklist 7.6.0 on Firefox Nightly 103, but no permission prompt showed.
Maybe we need to wait for an update of Fenix.
https://github.com/mozilla-mobile/fenix/issues/16912
any update more?
Please support kiwi browser too
Is there an update for this?
I can't activate UBlacklist on Android from the Firefox Add-On Store in Firefox / Firefox Nightly, nor from the Chrome Web Store in Kiwi.
Is there a workaround to use UBlacklist on Android?
I can't activate UBlacklist on Android from the Firefox Add-On Store in Firefox / Firefox Nightly
You need to use Firefox Beta or Firefox Nightly, and follow these instructions: https://blog.mozilla.org/addons/2020/09/29/expanded-extension-support-in-firefox-for-android-nightly/
However, it only works in the news tab as long as the checkmark for desktop view is checked in Firefox.
Works for me on Firefox Beta for Android. (I'm also using Google Search Fixer to bypass the browser discrimination.)
I have edited my previous post. The problem occurs only in the news tab. Here it works only in desktop view mode.
uBlackList V8.3.1 is working great for Google Search with Firefox Beta 117.0b4.
It is not working for other search engines for me.
In Options, Other Search Engines, the Enable button does not work if any other search engine is checked.
but no permission prompt showed
Looks like it should be released in Firefox 120: https://github.com/mozilla-mobile/firefox-android/pull/3917
It works now! https://www.mozilla.org/en-US/firefox/android/120.0/releasenotes/
I was able to install the extension directly from the firefox addon page. Although the mobile search is still limited to officially endorsed addons, accessing the page directly allows to click and install it.
Not sure if the rollout is the same for everyone though.
I'm closing this issue because we can now install this add-on to Firefox for Android 120.0 (or later).
For further problems with Firefox for Android, please create a separate issue for each problem.
| gharchive/issue | 2021-09-10T05:45:16 | 2025-04-01T06:44:32.620418 | {
"authors": [
"3xploiton3",
"Eagle-85",
"NotaInutilis",
"Semro",
"amroncz",
"iorate",
"item4",
"teohhanhui",
"yan12125"
],
"repo": "iorate/ublacklist",
"url": "https://github.com/iorate/ublacklist/issues/117",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1082906606 | ClientError: {}
Issue description
Above error occurs when transferring IOTA to Firefly from an exchange.
The transfer has worked previously with different addresses.
Firefly version
v1.2.0
OS details
macOS Monterey v12.0.1
Can the issue be reliably reproduced?
No
Steps to reproduce the issue
Did not reproduce the error as previous IOTA transferred are still unaccounted for.
...
Expected behaviour
That the IOTA transfer from the exchange to Firefly is successful and auditable, like the 5-6 previous transfers.
Actual behaviour
The message from the exchange is:
Error Message: Withdrawals are currently disabled for this address due to wallet maintenance.
Errors
IOTA Firefly Error file - 17Dec21.docx
The transfer failed on the following address:
iota1qzfy3cmlh5umjej2xstffu0mfl6ntvmgn22hhgavxkuwts9p65veuw4g8qk
If the exchange disabled withdrawals then this has nothing to do with Firefly. The funds were not yet transferred to your address (probably due to maintenance on their own wallet) as you can see here.
The empty ClientError will be fixed in the next release https://github.com/iotaledger/firefly/issues/1433, but as @r-c-k wrote your problem is only on exchange side
| gharchive/issue | 2021-12-17T05:39:29 | 2025-04-01T06:44:32.626894 | {
"authors": [
"Thoralf-M",
"clareiota",
"r-c-k"
],
"repo": "iotaledger/firefly",
"url": "https://github.com/iotaledger/firefly/issues/1772",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2162062026 | Add constructor for VerificationMethod in TS
Fixes https://github.com/iotaledger/identity.rs/issues/1320
@UMR1352 can you please take a look at this?
| gharchive/pull-request | 2024-02-29T20:37:25 | 2025-04-01T06:44:32.628651 | {
"authors": [
"abdulmth"
],
"repo": "iotaledger/identity.rs",
"url": "https://github.com/iotaledger/identity.rs/pull/1321",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1246274391 | UI: Reset svelte state when logging out
if a user logs out the svelte state shall be resetted otherwise for instance the selected detail view of an identity is still shown to the new logged in user
also search is still applied when logging out and in again
maybe also contact begona about a good concept to reset state
| gharchive/issue | 2022-05-24T09:37:42 | 2025-04-01T06:44:32.635216 | {
"authors": [
"dominic22"
],
"repo": "iotaledger/is-ui-components",
"url": "https://github.com/iotaledger/is-ui-components/issues/33",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2327127612 | 🛑 备用KMS激活服务器 01 is down
In 52a8233, 备用KMS激活服务器 01 (s11.ikms.eu.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 备用KMS激活服务器 01 is back up in 5f99680 after 10 minutes.
| gharchive/issue | 2024-05-31T07:16:38 | 2025-04-01T06:44:32.637776 | {
"authors": [
"iougemini"
],
"repo": "iougemini/ikms-uptime",
"url": "https://github.com/iougemini/ikms-uptime/issues/1955",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2738542939 | 🛑 备用KMS激活服务器 01 is down
In 1178e1a, 备用KMS激活服务器 01 (s11.ikms.eu.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 备用KMS激活服务器 01 is back up in aac1a7c after 11 minutes.
| gharchive/issue | 2024-12-13T14:36:12 | 2025-04-01T06:44:32.640298 | {
"authors": [
"iougemini"
],
"repo": "iougemini/ikms-uptime",
"url": "https://github.com/iougemini/ikms-uptime/issues/3817",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
733152776 | How to get face embedding / encoding?
Congrats for this nice project @ipazc!
I see the output of your algorithm is something like:
[
{
'box': [277, 90, 48, 63],
'keypoints':
{
'nose': (303, 131),
'mouth_right': (313, 141),
'right_eye': (314, 114),
'left_eye': (291, 117),
'mouth_left': (296, 143)
},
'confidence': 0.99851983785629272
}
]
i.e. it gives the bounding box, the keypoints (nose, mouth, eye, etc.).
But how to get an face embedding / face encoding, to be able to do face identification?
I may have misunderstood what you are asking, but this project isn't about getting face encodings for recognition purposes. For that, you will need something else, such as https://github.com/ageitgey/face_recognition.
@MattyB95 Maybe I misunderstood this project, but I thought MTCNN + FaceNet not only allows face detection, but also recognition / identification, by assigning a 128D vector embedding to each face.
https://github.com/davidsandberg/facenet
Is it correct @ipazc?
@josephernest I'm not completely familiar with FaceNet but wouldn't that be providing the face recognition/identification encodings. This project would be more about cropping the image to the facial region for that purpose, but I will let @ipazc give his verdict :)
MTCNN is only used to detect faces in an image,
The value in the the "box" are the corners for the face box in an image
To get the face encoding you will need to pass the the pixels resulted from cropping the image to a face corners to one of the models that are used for face encoding, as facenet, deepface,....
@imnimn Is there a Python implementation that packs: MTCNN for cropping the face box + facenet or deepface for the encoding?
@imnimn Is there a Python implementation that packs together: MTCNN to get the face box + facenet or deepface for the encoding?
Please check this article:
https://arsfutura.com/magazine/face-recognition-with-facenet-and-mtcnn/,
It is implementation is in this repository
https://github.com/arsfutura/face-recognition
| gharchive/issue | 2020-10-30T13:01:29 | 2025-04-01T06:44:32.661562 | {
"authors": [
"MattyB95",
"imnimn",
"josephernest"
],
"repo": "ipazc/mtcnn",
"url": "https://github.com/ipazc/mtcnn/issues/102",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1627714379 | Where is the changelog?
I'm doing some maintenance work to upgrade dependencies. Right now I'm at version 3.4.3. I see the latest version is 4.0.7. When doing updates, I like to go through a changelog to see if there is any breaking changes I should keep a lookout for. I seem to be unable to find the changelog for this library. I assume going from 3 to 4 is potentially a big upgrade. Can you point me to where the changelog is to help with this update?
Hi @lightswitch05 this is great idea and am sorry we did not already include one, we will do so going forward.
| gharchive/issue | 2023-03-16T15:10:39 | 2025-04-01T06:44:32.663203 | {
"authors": [
"jonathan-kosgei",
"lightswitch05"
],
"repo": "ipdata/python",
"url": "https://github.com/ipdata/python/issues/24",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1114180403 | p4c-xdp: Build and install p4c-xdp compiler
This loads the p4c-xdp compiler extension, builds it, and instals it as
a part of the host native install. Useful for experimenting with P4 to
XDP.
Signed-off-by: Kyle Mestery mestery@mestery.com
Closing for now.
| gharchive/pull-request | 2022-01-25T17:38:34 | 2025-04-01T06:44:32.664900 | {
"authors": [
"mestery"
],
"repo": "ipdk-io/ipdk",
"url": "https://github.com/ipdk-io/ipdk/pull/68",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2599432266 | Introduce ExternManager class architecture
This CL implements an architecture that can be used to isolate code that depends on the ES2K-specific IDPF extensions to P4Runtime from the rest of the code.
As things stand right now, we cannot upstream our P4InfoManager changes because they break non-IPDK Stratum platforms (barefoot, bcm, fmv2, etc.) unless they are built against our custom P4Runtime. The new architecture adds an interface that allows platform-specific extensions to P4InfoManager to be incorporated at build time.
The architecture also provides a mechanism that supports custom extensions to TdiTableManager to be incorporated at build time.
The IDPF-dependent code goes into the extensions, which are compiled and linked in as part of the ES2K build.
This CL adds the Es2kExternManager class and its relatives, together with a unit test and Bazel updates to support the new components.
It does NOT include the P4InfoManager and TdiTableManager changes that remove the existing IDPF support and replace it with the ExternManager code. The new files will be passive additions to the codebase until this is done (in a subsequent commit).
See issue https://github.com/ipdk-io/stratum-dev/issues/274 for more information. It includes a UML class diagram.
Development notes
These changes were developed in Draft PR https://github.com/ipdk-io/stratum-dev/pull/272. The CL consists of 37 files, which is a lot to review. It's also downright scary.
To mitigate risk, I reduced the number of changes to TdiTableManager and wrote an extensive unit test for Es2kExternManager. The latter turned out to be a worthwhile effort: it found several bugs, and the feedback led to a number of revisions to the code being tested. (Error handling has been completely overhauled.) The unit test covers 100% of the UUT.
I also decided to submit the new code first, including the unit test. Fewer files, very little overlap, and 100% safe because the only thing that compiles or links the new code is the unit test.
I copied the new files to a fresh branch, made a few additional edits, and created this PR.
Code review notes
tdi_table_handler.cc and tdi_table_handlers.h were extracted from tdi_table_manager.cc. I needed to share the functions with the resource handlers, which were also lifted from the TableManager. The originals will be deleted as part of integrating the old and new code.
When reviewing the ES2K resource handlers, you can compare the code with the originals in TdiTableManager. It should be essentially the same. The ResourceHandler method is named for the TableManager method from which it was lifted, with the addition of a Do prefix to distinguish between the TableManager and ResourceManager methods.
Closed in favor of PR https://github.com/ipdk-io/stratum-dev/pull/276, which implements Stage 1 of the design.
| gharchive/pull-request | 2024-10-19T17:32:24 | 2025-04-01T06:44:32.671040 | {
"authors": [
"ffoulkes"
],
"repo": "ipdk-io/stratum-dev",
"url": "https://github.com/ipdk-io/stratum-dev/pull/273",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1454285365 | Allow passing arbitrary fetchurl args per registry
Motivation
Same as #162 but with my own updates pushed to it
Closes #159
Checklist
[ ] added tests to verify new behavior
[ ] added an example template or updated an existing one
[ ] updated docs/API.md with changes
[ ] updated CHANGELOG.md
Note: still need to update the docs and changelog before merging
| gharchive/pull-request | 2022-11-18T01:30:20 | 2025-04-01T06:44:32.677630 | {
"authors": [
"ipetkov"
],
"repo": "ipetkov/crane",
"url": "https://github.com/ipetkov/crane/pull/168",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1297190148 | Add peergos talk
@b5
thanks @ianopolous! merged this content with #104
| gharchive/pull-request | 2022-07-07T10:49:19 | 2025-04-01T06:44:32.683243 | {
"authors": [
"b5",
"ianopolous"
],
"repo": "ipfs-shipyard/ipfs-thing-2022",
"url": "https://github.com/ipfs-shipyard/ipfs-thing-2022/pull/103",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
119272908 | D3 Tree Viz "help wanted" issues
@Dignifiedquire I did a bit of thinking and cleanup on this repo after pinging you in #8. I also wrote a readme for this repo, so the whole D3 tree thing should be a bit more scrutable now. Things I would love your help with if you have bandwidth and interest:
#15 d3 tree should use js-ipfs-api
#16 d3 tree should not load all nodes recursively; should initially load two levels
#14 d3 tree should expand downward to fit all nodes
#9 d3tree: support paths
@RichardLitt perhaps you'd be interested in helping out with #8: add d3 tree dataviz to "examples" on ipfs.io
I labeled all of these "help wanted" and have been organizing them in waffle: https://waffle.io/ipfs/dataviz
@harlantwood Thanks for doing this! This is pretty awesome. I can help with examples, sure!
Awesome, thanks @RichardLitt!
BTW I am open to this simple example evolving technically:
from CoffeeScript to es2015
from make to webpack/browserify
other directions that inspire potential contributors :smile:
| gharchive/issue | 2015-11-28T05:47:44 | 2025-04-01T06:44:32.687584 | {
"authors": [
"RichardLitt",
"harlantwood"
],
"repo": "ipfs/dataviz",
"url": "https://github.com/ipfs/dataviz/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
241986873 | All of POST request's results are 403 Forbidden
error:
ipfs gateway config:
config ~/.ipfs/config
"API": {
"HTTPHeaders": {
"Access-Control-Allow-Methods": [
"PUT",
"GET",
"POST"
],
"Access-Control-Allow-Origin": [
"*"
]
}
},
@shanelau I try to fix the ~/.ipfs/config as the same you, but it still go into the Web UI of IPFS when i run http://local:5001/ipfs/(PeerId)/#. It show me the 403 error.
Same problem ;(
I tried to copy @shanelau's snippet but it seems like this might be a bug with go-ipfs v0.5.0-rc1.
This Command Works for me
ipfs config --json API.HTTPHeaders.Access-Control-Allow-Origin '["'"$origin"'", "http://127.0.0.1:8080","http://localhost:3000"]' you can allow the request from multiple origins
| gharchive/issue | 2017-07-11T09:40:47 | 2025-04-01T06:44:33.316154 | {
"authors": [
"NatoBoram",
"aliawais0007",
"hleb-albau",
"shanelau",
"vutuyetmai130398"
],
"repo": "ipfs/ipfs-webui",
"url": "https://github.com/ipfs/ipfs-webui/issues/596",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
404962606 | Basic auth support
For people using a private IPFS cluster with a swarm key, it is also useful to have support for HTTP basic auth. Interaction with the gateway is easy, as there is no library and you can just use URLs like https://user:pass@gatewayurl.com/ipfs/hash. However, for using the IPFS API, this is not as simple.
Making py-ipfs-api support basic auth and is a useful and simple change that I'm willing to implement. Instead of vendoring this code I would love to submit a PR. Just want to ask before doing this - would you accept a PR for basic auth support?
+1 on this. This worked with ipfsapi but the new mutliaddr scheme doesn't support it and I'm trying to figure out how to connect.
@brianmcmichael could you try https://github.com/ldub/py-ipfs-api/ ? I'm having an issue testing it against my own IPFS node because of some SSL issue, but you should be able to do the following as mentioned in the README:
>>> import base64
>>> import ipfshttpclient
>>> creds = base64.b64encode(b"username:password").decode("utf-8")
>>> headers = {"Authorization" : "Basic " + creds}
>>> api = ipfshttpclient.connect('/dns/ipfs-api.example.com/tcp/443/https', headers=headers)
The TLS issue should be fixed (see #185), I'm still open to adding support for dedicated username and password parameters to the codebase btw. :wink:
| gharchive/issue | 2019-01-30T21:01:18 | 2025-04-01T06:44:33.321632 | {
"authors": [
"alexander255",
"brianmcmichael",
"ldub"
],
"repo": "ipfs/py-ipfs-api",
"url": "https://github.com/ipfs/py-ipfs-api/issues/154",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.