id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
2080134422 | 🛑 DNS2.electro is down
In 0dc6256, DNS2.electro (ns2.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS2.electro is back up in 0f438c6 after 33 minutes.
| gharchive/issue | 2024-01-13T06:44:10 | 2025-04-01T04:33:03.198953 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/11645",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2080164692 | 🛑 DNS1.electro is down
In 573c189, DNS1.electro (ns1.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS1.electro is back up in a02b72c after 1 hour, 29 minutes.
| gharchive/issue | 2024-01-13T07:49:20 | 2025-04-01T04:33:03.201792 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/11650",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2086248850 | 🛑 DNS2.electro is down
In 18831ab, DNS2.electro (ns2.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS2.electro is back up in c62cd9f after 14 minutes.
| gharchive/issue | 2024-01-17T13:56:07 | 2025-04-01T04:33:03.204070 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/12018",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2117327749 | 🛑 DNS2.electro is down
In d12b976, DNS2.electro (ns2.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS2.electro is back up in 1ce0dbe after 1 hour, 10 minutes.
| gharchive/issue | 2024-02-04T21:47:27 | 2025-04-01T04:33:03.206351 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/13550",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2128337018 | 🛑 DNS2.electro is down
In 87a87bf, DNS2.electro (ns2.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS2.electro is back up in 312cfaa after 16 minutes.
| gharchive/issue | 2024-02-10T11:32:18 | 2025-04-01T04:33:03.208567 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/13985",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2177744536 | 🛑 DNS2.electro is down
In dc362c2, DNS2.electro (ns2.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS2.electro is back up in 7a7339b after 31 minutes.
| gharchive/issue | 2024-03-10T14:26:00 | 2025-04-01T04:33:03.210826 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/16503",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1911543309 | 🛑 DE Server is down
In 5610bce, DE Server (de1.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DE Server is back up in f88b8c8 after 6 minutes.
| gharchive/issue | 2023-09-25T13:42:24 | 2025-04-01T04:33:03.213251 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/1706",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2219388761 | 🛑 DE Server is down
In 6b85d40, DE Server (de1.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DE Server is back up in 6b7a2bc after 1 hour, 26 minutes.
| gharchive/issue | 2024-04-02T01:23:14 | 2025-04-01T04:33:03.215517 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/18349",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1916445220 | 🛑 DE Server is down
In 3510614, DE Server (de1.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DE Server is back up in c906b2c after 8 minutes.
| gharchive/issue | 2023-09-27T22:50:20 | 2025-04-01T04:33:03.217739 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/1959",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1966934437 | 🛑 DNS1.electro is down
In 43b87b7, DNS1.electro (ns1.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS1.electro is back up in 1699d47 after 7 minutes.
| gharchive/issue | 2023-10-29T10:37:21 | 2025-04-01T04:33:03.219963 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/5058",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2005552563 | 🛑 DNS2.electro is down
In bb43cbd, DNS2.electro (ns2.elteam.ir) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS2.electro is back up in b6135cd after 9 minutes.
| gharchive/issue | 2023-11-22T04:28:21 | 2025-04-01T04:33:03.222255 | {
"authors": [
"SIRMaxis"
],
"repo": "SIRMaxis/uptime-robot",
"url": "https://github.com/SIRMaxis/uptime-robot/issues/7205",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2509339142 | Please consider fixing issues detected by repository checker
Notification from ioBroker Check and Service Bot
Dear adapter developer,
I'm the ioBroker Check and Service Bot. I'm an automated tool processing routine tasks for the ioBroker infrastructure. I have recently checked the repository for your adapter energiefluss for common errors and appropiate suggestions to keep this adapter up to date.
This check is based the current head revisions (master / main branch) of the adapter repository
Please see the result of the check below.
ioBroker.energiefluss
-
ERRORS:
[ ] :heavy_exclamation_mark: [E162] js-controller 3.3.22 listed as dependency but 4.0.24 is required as minimum, 5.0.19 is recommended. Please update dependency at io-package.json.
[ ] :heavy_exclamation_mark: [E204] Version "3.4.0" listed at common.news at io-package.json does not exist at NPM. Please remove from news section.
[ ] :heavy_exclamation_mark: [E999] GLOBAL ERROR: SyntaxError: Expected property name or '}' in JSON at position 7, {}
WARNINGS:
[ ] :eyes: [W028] Minimum node.js version 18 recommended. Please adapt "{'engines' : { 'node' >= '16' } }" at package.json.
[ ] :eyes: [W040] "keywords" within package.json should contain "ioBroker"
[ ] :eyes: [W115] "common.tier" is required in io-package.json. Please check https://github.com/ioBroker/ioBroker.docs/blob/master/docs/en/dev/objectsschema.md#adapter.
[ ] :eyes: [W181] "common.license" in io-package.json is deprecated. Please define object "common.licenseInformation"
[ ] :eyes: [W184] "common.main" is deprecated and ignored. Please remove from io-package.json. Use "main" at package.json instead.
[ ] :eyes: [W184] "common.materialize" is deprecated for admin >= 5 at io-package.json. Please use property "adminUI".
[ ] :eyes: [W184] "common.materializeTab" is deprecated for admin >= 5 at io-package.json. Please use property "adminUI".
SUGGESTIONS:
[ ] :pushpin: [S522] Please consider migrating to admin 5 UI (jsonConfig).
Please review issues reported and consider fixing them as soon as appropiate.
Errors reported by repository checker should be fixed as soon as possible. Some of them require a new release to be considered as fixed. Please note that errors reported by checker might be considered as blocking point for future updates at stable repository.
Warnings reported by repository checker should be reviewed. While some warnings can be ignored due to good reasons or a dedicated decision of the developer, most warnings should be fixed as soon as appropiate.
Suggestions reported by repository checker should be reviewed. Suggestions can be ignored due to a decision of the developer but they are reported as a hint to use a configuration which might get required in future or at least is used be most adapters. Suggestions are always optional to follow.
You may start a new check at any time by adding the following comment to this issue:
@iobroker-bot recheck
Please note that I (and the server at GitHub) have always plenty of work to do. So it may last up to 30 minutes until you see a reaction. I will drop a comment here as soon as I start processing.
Feel free to contact me (@iobroker-bot) if you have any questions or feel that an issue is incorrectly flagged.
And THANKS A LOT for maintaining this adapter from me and all users.
Let's work together for the best user experience.
your
ioBroker Check and Service Bot
@mcm1957 for evidence
Last update at Fri, 06 Sep 2024 02:40:57 GMT
created on error
| gharchive/issue | 2024-09-06T02:40:57 | 2025-04-01T04:33:03.236801 | {
"authors": [
"ioBroker-Bot"
],
"repo": "SKB-CGN/ioBroker.energiefluss",
"url": "https://github.com/SKB-CGN/ioBroker.energiefluss/issues/390",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
528528512 | Unused-variable warnings ignore GLOG
While building this project, warnings like following example occur many times.
/home/odroid/Desktop/test-cof/include/pointers/allocator.h: In instantiation of ‘void flash::unmap_file(flash::flash_ptr<X>) [with T = float]’:
/home/odroid/Desktop/test-cof/drivers/gemm.cpp:72:26: required from here
/home/odroid/Desktop/test-cof/include/pointers/allocator.h:49:9: warning: unused variable ‘ret’ [-Wunused-variable]
int ret = munmap(
When I look into the files that make warnings, 'unused variables' are actually used in GLOG sentences but not counted as used.
This bothers us when we modify and debug the codes. It should be solved.
The current direction of research is mainly NAS that finds a deep learning model suitable for a given memory size rather than using less memory.
Therefore, it is more valuable to train a model for the GPU and generate an optimized GPU kernel than to apply the technique to the GPU.
These works are in progress in the following repositories.
https://github.com/SKKU-ESLAB/ANT-Model-DB
https://github.com/SKKU-ESLAB/Auto-Compression
So, I close this issue.
| gharchive/issue | 2019-11-26T06:59:25 | 2025-04-01T04:33:03.239623 | {
"authors": [
"GH-Jo",
"LeeHayun"
],
"repo": "SKKU-ESLAB/CNN-on-flash",
"url": "https://github.com/SKKU-ESLAB/CNN-on-flash/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
364969609 | Help! dark theme
Does anyone want to try to implement this theme in readthedocs/sphinx?
https://userstyles.org/styles/159458/read-the-docs-dark?utm_campaign=stylish_stylepage
Seems relevant: https://stackoverflow.com/questions/14622698/customize-sphinxdoc-theme
| gharchive/issue | 2018-09-28T17:03:51 | 2025-04-01T04:33:03.244256 | {
"authors": [
"yinleon"
],
"repo": "SMAPPNYU/youtube-data-api",
"url": "https://github.com/SMAPPNYU/youtube-data-api/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2190362381 | 🛑 Ayi Academy is down
In c0f6c11, Ayi Academy (https://ayi.academy/) was down:
HTTP code: 521
Response time: 230 ms
Resolved: Ayi Academy is back up in 78fadc7 after 11 minutes.
| gharchive/issue | 2024-03-16T23:44:01 | 2025-04-01T04:33:03.274480 | {
"authors": [
"SOPORTEAYIGROUP"
],
"repo": "SOPORTEAYIGROUP/AYIGROUP",
"url": "https://github.com/SOPORTEAYIGROUP/AYIGROUP/issues/2044",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2197321037 | 🛑 CAOS is down
In 1b50e0b, CAOS (http://caos.ayi-asociados.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: CAOS is back up in 57a3212 after 17 minutes.
| gharchive/issue | 2024-03-20T11:39:16 | 2025-04-01T04:33:03.276758 | {
"authors": [
"SOPORTEAYIGROUP"
],
"repo": "SOPORTEAYIGROUP/AYIGROUP",
"url": "https://github.com/SOPORTEAYIGROUP/AYIGROUP/issues/2590",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1304410608 | Add RDF
Switch to standardized ontologies (using https://www.ebi.ac.uk/ols and/or https://bioportal.bioontology.org/) and use RDF format to make AncientMetagenomeDir available through semantic web technologies (querying with SPARQL for example)
Related to #97 and #257
Problem is often because we have 'ancient' categories, they don't fit into the ontologies nicely and the maintainers complain ;)
| gharchive/issue | 2022-07-14T08:00:02 | 2025-04-01T04:33:03.283501 | {
"authors": [
"jfy133",
"maxibor"
],
"repo": "SPAAM-community/AncientMetagenomeDir",
"url": "https://github.com/SPAAM-community/AncientMetagenomeDir/issues/937",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
609150411 | Completions and validations don't have the same result in NBT paths
See here reported by Ellivers.
:tada: This issue has been resolved in version 2.0.1 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/issue | 2020-04-29T15:30:26 | 2025-04-01T04:33:03.289830 | {
"authors": [
"SPGoding"
],
"repo": "SPGoding/datapack-language-server",
"url": "https://github.com/SPGoding/datapack-language-server/issues/442",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1099370340 | Remove obsolete files related to cpuid, etc.
This fixes compilation on non-X86 platforms
Performance test reportHPXLocal PerformanceComparisonBENCHMARKFORK_JOIN_EXECUTORPARALLEL_EXECUTORSCHEDULER_EXECUTORFor Each-(=)(=)InfoPropertyBeforeAfterHPXLocal Commit01e4980e2184ca113b853ae3a36869f713fcc4b9df927d028abf7bdaff19d373d858eb79e0bf5e88HPXLocal Datetime2021-12-21T15:01:46+00:002022-01-11T16:22:19+00:00Compiler/apps/daint/SSL/HPX/packages/llvm-11.0.0/bin/clang++ 11.0.0/apps/daint/SSL/HPX/packages/llvm-11.0.0/bin/clang++ 11.0.0Hostnamenid00932nid01528EnvfileDatetime2021-12-21T16:09:15.238666+01:002022-01-11T17:29:01.165959+01:00ClusternamedaintdaintComparisonBENCHMARKNO-EXECUTORFuture Overhead - Create Thread Hierarchical - Latch+InfoPropertyBeforeAfterHPXLocal Commitf499a2233385060b8a2612ab88163e62b0881888df927d028abf7bdaff19d373d858eb79e0bf5e88HPXLocal Datetime2021-12-10T13:50:04+00:002022-01-11T16:22:19+00:00Compiler/apps/daint/SSL/HPX/packages/llvm-11.0.0/bin/clang++ 11.0.0/apps/daint/SSL/HPX/packages/llvm-11.0.0/bin/clang++ 11.0.0Hostnamenid00243nid01528EnvfileDatetime2021-12-10T15:19:42.442217+01:002022-01-11T17:29:17.529626+01:00ClusternamedaintdaintComparisonBENCHMARKFORK_JOIN_EXECUTORPARALLEL_EXECUTORSCHEDULER_EXECUTORStream Benchmark - Add-(=)(=)Stream Benchmark - Scale(=)(=)(=)Stream Benchmark - Triad(=)-(=)Stream Benchmark - Copy(=)=(=)InfoPropertyBeforeAfterHPXLocal Commitf64fbd02165a132a6276cedd14c586910abb79e4df927d028abf7bdaff19d373d858eb79e0bf5e88HPXLocal Datetime2021-11-12T11:29:27+00:002022-01-11T16:22:19+00:00Compiler/apps/daint/SSL/HPX/packages/llvm-11.0.0/bin/clang++ 11.0.0/apps/daint/SSL/HPX/packages/llvm-11.0.0/bin/clang++ 11.0.0Hostnamenid00007nid01528EnvfileDatetime2021-11-12T12:57:50.824026+01:002022-01-11T17:29:32.228704+01:00ClusternamedaintdaintExplanation of SymbolsSymbolMEANING=No performance change (confidence interval within ±1%)(=)Probably no performance change (confidence interval within ±2%)(+)/(-)Very small performance improvement/degradation (≤1%)+/-Small performance improvement/degradation (≤5%)++/--Large performance improvement/degradation (≤10%)+++/---Very large performance improvement/degradation (>10%)?Probably no change, but quite large uncertainty (confidence interval with ±5%)??Unclear result, very large uncertainty (±10%)???Something unexpected…
This is superseded by https://github.com/STEllAR-GROUP/hpx/pull/5742
| gharchive/pull-request | 2022-01-11T16:22:17 | 2025-04-01T04:33:03.367219 | {
"authors": [
"StellarBot",
"hkaiser"
],
"repo": "STEllAR-GROUP/hpx-local",
"url": "https://github.com/STEllAR-GROUP/hpx-local/pull/23",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
265259292 | hpx.ini is not used correctly at initialization
Using an hpx.ini file to set the number of threads for hpx does not work, i.e. a file hpx.ini:
[hpx]
os_threads = n
which is loaded with --hpx:config=hpx.ini (same directory as executable) does not change os_threads (confirmed with --hpx:dump-config).
Changing the number of threads with --hpx:ini=hpx.os_threads=n or --hpx:threads=n works correctly.
This does not affect all config values. I was able to change for example the stack sizes.
Confirmed - will fix.
@msimberg please verify whether #2953 is fixing your problem.
@hkaiser Thanks, this mostly fixes the problem. It seems like it sets the default number of threads to 1 though rather than the number of cores.
It also seems like it doesn't pick up any of the other ways of setting the ini file (HPX_INI, PWD, HOME etc.). If I set any valid path with --hpx:config it does load an ini file from those locations as well. Not sure if this is related or not. Should I open a separate issue for that?
It seems like it sets the default number of threads to 1 though rather than the number of cores.
Confirmed, thanks for noticing this. Should be fixed now.
It also seems like it doesn't pick up any of the other ways of setting the ini file (HPX_INI, PWD, HOME etc.). If I set any valid path with --hpx:config it does load an ini file from those locations as well. Not sure if this is related or not.
Could it be a matter of which file takes precedence? OTO, it might be fixed by the change I made.
Should I open a separate issue for that?
No need, I think.
The original issue is solved now, thank you! The additional issue of loading from HPX_INI etc. still remains though.
@msimberg Could you explain one more time how to reproduce the issue you're trying to report, please? I'm not sure I understood what you're after.
All good now, thank you!
(For the record, the second issue was that an ini file gets loaded automatically from PWD, HOME etc. only when the --hpx:config option is set.)
| gharchive/issue | 2017-10-13T11:37:37 | 2025-04-01T04:33:03.373221 | {
"authors": [
"hkaiser",
"msimberg"
],
"repo": "STEllAR-GROUP/hpx",
"url": "https://github.com/STEllAR-GROUP/hpx/issues/2947",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
95750064 | Apex integration
Hey Hartmut! I think everything is merged and tested. Please let me know if there are problems, I will continue to debug.
Can we now close this (after #1726 was created)?
| gharchive/pull-request | 2015-07-17T21:36:34 | 2025-04-01T04:33:03.374351 | {
"authors": [
"hkaiser",
"khuck"
],
"repo": "STEllAR-GROUP/hpx",
"url": "https://github.com/STEllAR-GROUP/hpx/pull/1666",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
352408727 | Use source cache and workspace between job steps on CircleCI
Fixes #3424.
Note: this does a shallow local clone to avoid copying the whole 900 MB repository using the workspace between work steps. This is not strictly necessary and can be reverted if we ever want to query the git repo for more information than the current hash. I did this because copying the whole repo using the workspace was significantly slower than doing the same with the cache, but since there doesn't seem to be a way to save to the cache anymore on PRs from forked repos this is a small optimization to get it back to the same performance as before.
The problem with using workspaces was parallel updates. If that's fine now, this should be good
Yeah, I remember the problems. I don't think anything changed so it might still come back and haunt us, but for a few builds this has at least worked...
| gharchive/pull-request | 2018-08-21T06:57:15 | 2025-04-01T04:33:03.376009 | {
"authors": [
"msimberg",
"sithhell"
],
"repo": "STEllAR-GROUP/hpx",
"url": "https://github.com/STEllAR-GROUP/hpx/pull/3425",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2240705871 | 🛑 SPCast / Loadbalancer is down
In 92502e8, SPCast / Loadbalancer (https://loadbalancer.sp.radio.fm/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: SPCast / Loadbalancer is back up in b432aed after 7 minutes.
| gharchive/issue | 2024-04-12T18:37:21 | 2025-04-01T04:33:03.408165 | {
"authors": [
"scysys"
],
"repo": "STREAMPANEL/status.streampanel.net",
"url": "https://github.com/STREAMPANEL/status.streampanel.net/issues/1674",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
703515490 | error: command 'gcc' failed with exit status 1
When I run
python setup.py build_ext --inplace
get this error.
running build_ext building 'toolkit.utils.region' extension gcc -pthread -B /root/anaconda3/envs/pysot/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes /usr/local/cuda-9.0/include -fPIC -I./toolkit/utils -Itoolkit/utils/src -I/root/anaconda3/envs/pysot/include/python3.7m -c toolkit/utils/region.c -o build/temp.linux-x86_64-3.7/toolkit/utils/region.o toolkit/utils/region.c: In function ‘__pyx_pw_7toolkit_5utils_6region_7Polygon_5__str__’: toolkit/utils/region.c:4335:70: warning: ‘__pyx_v_i’ may be used uninitialized in this function [-Wmaybe-uninitialized] __pyx_t_10 = PyFloat_FromDouble((__pyx_v_self->_c_region_polygon->x[__pyx_v_i])); if (u ^ toolkit/utils/region.c:4201:8: note: ‘__pyx_v_i’ was declared here long __pyx_v_i; ^ gcc: warning: /usr/local/cuda-9.0/include: linker input file unused because linking not done gcc -pthread -B /root/anaconda3/envs/pysot/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes /usr/local/cuda-9.0/include -fPIC -I./toolkit/utils -Itoolkit/utils/src -I/root/anaconda3/envs/pysot/include/python3.7m -c toolkit/utils/src/region.c -o build/temp.linux-x86_64-3.7/toolkit/utils/src/region.o gcc: warning: /usr/local/cuda-9.0/include: linker input file unused because linking not done gcc -pthread -shared -B /root/anaconda3/envs/pysot/compiler_compat -L/root/anaconda3/envs/pysot/lib -Wl,-rpath=/root/anaconda3/envs/pysot/lib -Wl,--no-as-needed -Wl,--sysroot=/ /usr/local/cuda-9.0/lib64 /usr/local/cuda-9.0/include build/temp.linux-x86_64-3.7/toolkit/utils/region.o build/temp.linux-x86_64-3.7/toolkit/utils/src/region.o -o /wangyufan/code/Siamese/pysot/toolkit/utils/region.cpython-37m-x86_64-linux-gnu.so /root/anaconda3/envs/pysot/compiler_compat/ld: /usr/local/cuda-9.0/lib64: file not recognized: Is a directory collect2: error: ld returned 1 exit status error: command 'gcc' failed with exit status 1
My gcc version is 5.4.0, cuda version is 9.0.176, cudnn version is 7.0.5
I have installed all packages needed
Here is my package list:
certifi 2020.6.20 cffi 1.14.2 colorama 0.4.3 cycler 0.10.0 Cython 0.29.21 kiwisolver 1.2.0 llvmlite 0.34.0 matplotlib 3.3.2 mkl-fft 1.1.0 mkl-random 1.1.1 mkl-service 2.3.0 numba 0.51.2 numpy 1.19.2 olefile 0.46 opencv-python 4.4.0.42 Pillow 7.2.0 pip 20.2.2 protobuf 3.13.0 pycparser 2.20 pyparsing 2.4.7 python-dateutil 2.8.1 PyYAML 5.3.1 setuptools 49.6.0.post20200814 six 1.15.0 tensorboardX 2.1 torch 0.4.1.post2 torchvision 0.2.1 tqdm 4.49.0 wheel 0.35.1 yacs 0.1.8
I don't kown where is the mistake? Thx.
have you resolved this issue?
if yes kindly drop your solution
| gharchive/issue | 2020-09-17T11:48:21 | 2025-04-01T04:33:03.412773 | {
"authors": [
"2a2a",
"dcode-baranidharan"
],
"repo": "STVIR/pysot",
"url": "https://github.com/STVIR/pysot/issues/434",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1958564906 | PytorchStreamReader failed reading zip archive: failed finding central directory
when I run app.py
Collecting usage statistics. To deactivate, set browser.gatherUsageStats to False.
You can now view your Streamlit app in your browser.
Network URL: http://10.119.70.148:8501
External URL: http://144.48.107.18:8501
2023-10-24 13:47:51.466322: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: SSE4.1 SSE4.2 AVX AVX2 AVX512F AVX512_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
text_encoder/model.safetensors not found
Loading pipeline components...: 100%|███████████████████████████████| 8/8 [00:03<00:00, 2.25it/s]
2023-10-24 13:48:09.295 Uncaught app exception
Traceback (most recent call last):
onda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_utils.py", line 245, in _get_or_create_cached_value
cached_result = cache.read_result(value_key)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "aconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_resource_api.py", line 447, in read_result
raise CacheKeyNotFoundError()
streamlit.runtime.caching.cache_errors.CacheKeyNotFoundError
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/homaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_utils.py", line 293, in _handle_cache_miss
cached_result = cache.read_result(value_key)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/homeanaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_resource_api.py", line 447, in read_result
raise CacheKeyNotFoundError()
streamlit.runtime.caching.cache_errors.CacheKeyNotFoundError
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/scriptrunner/script_runner.py", line 565, in _run_script
exec(code, module.__dict__)
File "/home//download/zero123plus/app.py", line 206, in <module>
SAMAPI.get_instance()
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_utils.py", line 194, in wrapper
return cached_func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_utils.py", line 223, in __call__
return self._get_or_create_cached_value(args, kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_utils.py", line 248, in _get_or_create_cached_value
return self._handle_cache_miss(cache, value_key, func_args, func_kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/streamlit/runtime/caching/cache_utils.py", line 302, in _handle_cache_miss
computed_value = self._info.func(*func_args, **func_kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home//download/zero123plus/app.py", line 41, in get_instance
sam = sam_model_registry[model_type](checkpoint=sam_checkpoint)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/segment_anything/build_sam.py", line 15, in build_sam_vit_h
return _build_sam(
^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/segment_anything/build_sam.py", line 105, in _build_sam
state_dict = torch.load(f)
^^^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/torch/serialization.py", line 797, in load
with _open_zipfile_reader(opened_file) as opened_zipfile:
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home//anaconda3/envs/train_sd/lib/python3.11/site-packages/torch/serialization.py", line 283, in __init__
super().__init__(torch._C.PyTorchFileReader(name_or_buffer))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: PytorchStreamReader failed reading zip archive: failed finding central directory
Can you double check that the checkpoint has been completely downloaded?
The SAM checkpoint did not download completely. Note that it is 2.4GB, larger than 32 bit int, so some downloaders e.g. 32 bit wget would truncate the result. This also happens if you are using our script but have internet issues.
Closing for now due to lack of activity. If you are still meeting the same problem, add more comments or open another issue. Thank you.
| gharchive/issue | 2023-10-24T05:49:56 | 2025-04-01T04:33:03.425874 | {
"authors": [
"Colin97",
"Shiyao-Huang",
"eliphatfs"
],
"repo": "SUDO-AI-3D/zero123plus",
"url": "https://github.com/SUDO-AI-3D/zero123plus/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
237818207 | exposed want_ironic in openstack-mkcloud
Without it want_ironic is lost when e.g. rebuilding ironic jobs.
Would it make sense to have want_ironic_proposal instead? (Look at the other values...)
@jsuchome currently it has nothing to do with proposal. it is only enabling underlying setup of networking for ironic.
| gharchive/pull-request | 2017-06-22T11:43:57 | 2025-04-01T04:33:03.432266 | {
"authors": [
"jsuchome",
"skazi0"
],
"repo": "SUSE-Cloud/automation",
"url": "https://github.com/SUSE-Cloud/automation/pull/2022",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
107672358 | $HOSTNAME in vhosts.d/portus.conf
It looks like $HOSTNAME is not set in this script? Something leads to an apache configuration file with this content:
[....]
SetEnv PORTUS_SECRET_KEY_BASE XXXXXXXXXXXXXXXXX
SetEnv PORTUS_MACHINE_FQDN ${HOSTNAME}
SetEnv PORTUS_KEY_PATH /srv/Portus/config/server.key
[...]
I guess this env is read by the application here. So that the issued token contains:
{
"iss": "${HOSTNAME}",
"sub": "dirk",
[...]
and lets the docker login command fail:
Error response from daemon: no successful auth challenge for https://portus.xxx:5000/v2/ - errors: [token auth attempt to https://portus.xxx:5000/v2/ realm "https://portus.xxx/v2/token" failed with status: 401 Unauthorized]
Docker registry log:
2015-09-22T09:03:30.829112+00:00 portus registry[1108]: time="2015-09-22T09:03:30Z" level=error msg="token from untrusted issuer: \"${HOSTNAME}\""
/cc @jordimassaguerpla
Hi Dirk! Thanks for the report. We have already fixed that in
https://github.com/SUSE/Portus/commit/effdebdc6f9193a46b3b0a5644fbac6daf3b66ba
I guess your appliance is not the latest one.
Could you send us the output of
rpm -qi Portus
This will tell us which is the latest commit.
Thanks again!
actually the appliance does not call setup_apache and that is why this is not working
Hi,
here is the output:
rpm -qi Portus
Name : Portus
Version : 20150918120318
Release : 1.1
Architecture: x86_64
Install Date: Fri Sep 18 17:29:10 2015
Group : System/Management
Size : 53624951
License : Apache-2.0
Signature : RSA/SHA1, Fri Sep 18 12:37:38 2015, Key ID a193fbb572174fc2
Source RPM : Portus-20150918120318-1.1.src.rpm
Build Date : Fri Sep 18 12:36:56 2015
Build Host : build85
Relocations : (not relocatable)
Vendor : obs://build.opensuse.org/Virtualization
URL : https://github.com/SUSE/Portus
Summary : Authorization service and fronted for Docker registry (v2)
Description :
Portus targets version 2 of the Docker registry API. It aims to act both as an authoritzation server and as a user interface for the next generation of the Docker registry.
This package has been built with commit 75ce9316e3f126b49d2c9948030eda99c98acb49 from branch master on date Fri, 18 Sep 2015 12:03:18 +0000
Distribution: Virtualization:containers:Portus / openSUSE_13.2
Looks like this is newer than the commit, but I'll try the newest build (mine is Build62.6, I'll use 62.9)
I just created a PR that fixes this issue
https://github.com/SUSE/Portus/pull/360
also the appliance
https://build.opensuse.org/package/rdiff/Virtualization:containers:Portus/PortusAppliance?linkrev=base&rev=63
This has been fixed in commit 9c32ed9840a2f04d20c856295864b43289402d7f. Basically, this is now handled by the new portusctl tool. Feel free to re-open this issue.
| gharchive/issue | 2015-09-22T09:06:08 | 2025-04-01T04:33:03.442092 | {
"authors": [
"DirkHeinke",
"jordimassaguerpla",
"mssola"
],
"repo": "SUSE/Portus",
"url": "https://github.com/SUSE/Portus/issues/358",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
506570901 | WIP: Unify timeout handling in tests
Why is this PR needed?
Presently different tests using different approaches for handling timeouts when executing commands in the nodes, including those executed with ssh and kubectl`.
All this diversity of mechanism makes hard to reason about how long are tests actually waiting before failing, making difficult to debug the problems.
Fixes # https://github.com/SUSE/avant-garde/issues/939
What does this PR do?
Unify the timeout handling in tests using the wait function.
Use --timeout=0 option for kubectl wait calls to return immediately
and use the backoff/retry mechanism in the wait function.
Merge restrictions
(Please do not edit this)
We are in v4-maintenance phase, so we will restrict what can be merged to prevent unexpected surprises:
What can be merged (merge criteria):
2 approvals:
1 developer: code is fine
1 QA: QA is fine
there is a PR for updating documentation (or a statement that this is not needed)
by customized query i mean kubectl get pods -o jsonpath e.g. - this might be more reliable
@MaximilianMeister I agree but I think this change should be a different PR just to keep this scope of this PR around handling the timeouts.
@MaximilianMeister at the end I followed your suggestion of checking for the pod status because using the wait command caused some issues (basically, it generated an exception in the run_shellcommand function which was not properly handled) .
| gharchive/pull-request | 2019-10-14T10:34:12 | 2025-04-01T04:33:03.470529 | {
"authors": [
"MaximilianMeister",
"pablochacin"
],
"repo": "SUSE/skuba",
"url": "https://github.com/SUSE/skuba/pull/773",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
571447591 | versions: updated the kured version
There is a new version for kured available (1.3.0) which fixes some problems
we've been having in regards to v1.16 support. We've already done the packaging
heavy lifting, so the only thing left was to update the version here.
See SUSE/avant-garde#1346
Signed-off-by: Miquel Sabaté Solà msabate@suse.com
Merge restrictions
(Please do not edit this)
We are in v4-maintenance phase, so we will restrict what can be merged to prevent unexpected surprises:
What can be merged (merge criteria):
2 approvals:
1 developer: code is fine
1 QA: QA is fine
there is a PR for updating documentation (or a statement that this is not needed)
The package has to be submitted before we merge this
The package has already been accepted and the container for kured 1.3.0 has been generated.
Awesome, let's roll! :)
After a rebase, tests are passing. Review again :smile:
| gharchive/pull-request | 2020-02-26T15:00:53 | 2025-04-01T04:33:03.474033 | {
"authors": [
"evrardjp",
"mssola"
],
"repo": "SUSE/skuba",
"url": "https://github.com/SUSE/skuba/pull/972",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2598632915 | Stays at "Processing" and never converts.
Hi everyone. I've deleted the folder and started all over again just to see if that was the case, but it is not. I have everything in stalled, even ffmpeg and python version as well. The link opens, i am able to upload an audio sample and even synthesize, but i never ever get the conversion. Right now, it is literally at the 380-second mark. I understand it shouldn't take this long, so what am I doing incorrectly, or what am I missing?
maybe you could help with checking your torch version, see if it's -cpu or -cuda.
and make sure your connection to huggingface is fine, otherwise using a local ckpt is better
| gharchive/issue | 2024-10-19T01:19:44 | 2025-04-01T04:33:03.514504 | {
"authors": [
"BMKBAI",
"SWivid"
],
"repo": "SWivid/F5-TTS",
"url": "https://github.com/SWivid/F5-TTS/issues/171",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2407295064 | 🛑 Rahanet Movie AF is down
In 0124cb8, Rahanet Movie AF (http://movie.af/) was down:
HTTP code: 500
Response time: 560 ms
Resolved: Rahanet Movie AF is back up in 2ce4334 after 10 minutes.
| gharchive/issue | 2024-07-14T06:49:31 | 2025-04-01T04:33:03.522317 | {
"authors": [
"Saboor-Hakimi"
],
"repo": "Saboor-Hakimi/upptime",
"url": "https://github.com/Saboor-Hakimi/upptime/issues/655",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2160149917 | SO-85
Get client functionality & server (index.ts) [Supabase]
Functionality for fetching members based off of project or organization ties.
Loaders for dashboard routing (though may not be best implementation)
SideNavbar.tsx - added effect to check role, and remove the dashboard link if not admin
😪😪😿
Merging. :(
TBD: member per project tests...
| gharchive/pull-request | 2024-02-29T01:07:13 | 2025-04-01T04:33:03.524767 | {
"authors": [
"jvniorrr"
],
"repo": "SacOverflow/CAMEL-Services",
"url": "https://github.com/SacOverflow/CAMEL-Services/pull/22",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2278911031 | Type cast + less than
Describe the bug
The parser confuses type argument lists and the less than operator. That's the same issues we've had with type argument lists of calls.
To Reproduce
Place this code into the editor:
val filtered = houseRentData.filterRows((data) -> data.getValue("Bathroom") as Int < 5);
Expected behavior
No error. The < should be treated as the less than operator, not the start of a type argument list.
Screenshots (optional)
No response
Additional Context (optional)
No response
We cannot generally fix this. Adjusting the lexer only fixes the issue in some cases. Other languages, like TypeScript, have the same behavior.
We could give as and later is the lowest precedence, so parentheses must always be used when they occur inside another expression.
Or we could leave it as is, and document that parentheses must be used in this specific case (type cast followed by less than).
:tada: This issue has been resolved in version 0.15.0 :tada:
The release is available on:
npm package (@latest dist-tag)
npm package (@latest dist-tag)
v0.15.0
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/issue | 2024-05-04T10:11:49 | 2025-04-01T04:33:03.533897 | {
"authors": [
"lars-reimann"
],
"repo": "Safe-DS/DSL",
"url": "https://github.com/Safe-DS/DSL/issues/1150",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2643665651 | Syntax for file wrong after snippet suggestion
Make sure you have done the following
[X] I have updated to the latest version of blink.cmp
[X] I have read the README
Bug Description
Using LazyVim I've enabled the new blink extra. I'm experiencing issues when interacting with Friendly snippets. Normal suggestions / copilot suggestions work fine. But when selecting a friendly snippet the syntax for the file gets changed. Setting it manually afterwards, restores / fixes the bug.
https://github.com/user-attachments/assets/4b988652-1a35-4e5d-a0e3-3d28031c14f8
I'm running smarty a template language written in PHP. I'm not using treesitter highlighting. Just vim / regex based. Since this works the best for this language in my experience.
My snippets are in vscode format.
Relevant configuration
No response
neovim version
NVIM v0.11.0-dev-4072+g2377443cd-Homebrew
blink.cmp version: branch, tag, or commit
75cadbcd2657ed01326ca2b0e5e4d78a77127ca3
What in the world :joy: This seems to have been fixed at some point between v0.5.1 and main. I'll be cutting a new release soon
Scratch that, I was just able to reproduce on main. There's some ghosts in my code...
| gharchive/issue | 2024-11-08T10:52:42 | 2025-04-01T04:33:03.605103 | {
"authors": [
"Saghen",
"danjessen"
],
"repo": "Saghen/blink.cmp",
"url": "https://github.com/Saghen/blink.cmp/issues/295",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1591238852 | Create solution2.py
Description of Change
References
Checklist
[x] Added detailed description of changes.
[ ] Relevant documentation/comments is changed or added.
[x] Added file name matches File name guidelines.
[ ] PR title follows semantic commit guidelines.
Notes:
Added python solution
| gharchive/pull-request | 2023-02-20T06:39:32 | 2025-04-01T04:33:03.608287 | {
"authors": [
"ShubhamAsbe"
],
"repo": "SahilK-027/LeetCode",
"url": "https://github.com/SahilK-027/LeetCode/pull/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
539992616 | Enabling XMRig as the active mining plugin.
Taking a shortcut from the full fallback PR to try and fix the present NiceHash issues.
Superseded by #258.
| gharchive/pull-request | 2019-12-19T00:01:58 | 2025-04-01T04:33:03.619621 | {
"authors": [
"TimTinkers",
"seniorquico"
],
"repo": "SaladTechnologies/salad-applications",
"url": "https://github.com/SaladTechnologies/salad-applications/pull/253",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
387733537 | Refactoring
Changes:
implemented normalizr, state in the App component is a bit more confusing now, but unless we add in redux or something I think it's as good as we're gonna get. In state, there is:
an entities object. This holds all of the normalized data.
a movies object. This holds all of the data related to the movies page, same as before.
a movie object. This holds all of the data related to the singular movie page.
Added in a ripple effect on the cards.
Various other less exciting bits!
Looks great, thanks!
| gharchive/pull-request | 2018-12-05T12:25:39 | 2025-04-01T04:33:03.622165 | {
"authors": [
"SalemPCF",
"jakewtaylor"
],
"repo": "SalemPCF/creagle-movies",
"url": "https://github.com/SalemPCF/creagle-movies/pull/4",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1765920455 | [Snyk] Fix for 1 vulnerabilities
This PR was automatically created by Snyk using the credentials of a real user.Snyk has created this PR to fix one or more vulnerable packages in the `npm` dependencies of this project.
Changes included in this PR
Changes to the following files to upgrade the vulnerable dependencies to a fixed version:
packages/pwa-kit-runtime/package.json
Vulnerabilities that will be fixed
With an upgrade:
Severity
Issue
Breaking Change
Exploit Maturity
Regular Expression Denial of Service (ReDoS) SNYK-JS-SEMVER-3247795
Yes
Proof of Concept
Commit messages
Package name: cross-env
The new version differs by 1 commits.
61ebf59 fix: bumped cross-spawn to version 7.0.0 (#211)
See the full diff
Package name: semver
The new version differs by 8 commits.
e7b78de chore: release 7.5.2
58c791f fix: diff when detecting major change from prerelease (#566)
5c8efbc fix: preserve build in raw after inc (#565)
717534e fix: better handling of whitespace (#564)
2f738e9 chore: bump @ npmcli/template-oss from 4.14.1 to 4.15.1 (#558)
aa016a6 chore: release 7.5.1
d30d25a fix: show type on invalid semver error (#559)
09c69e2 chore: bump @ npmcli/template-oss from 4.13.0 to 4.14.1 (#555)
See the full diff
Check the changes in this PR to ensure they won't cause issues with your project.
Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs.
For more information:
🧐 View latest project report
🛠 Adjust project settings
📚 Read more about Snyk's upgrade and patch logic
Learn how to fix vulnerabilities with free interactive lessons:
🦉 Regular Expression Denial of Service (ReDoS)
W-13673737 created
| gharchive/pull-request | 2023-06-20T18:24:32 | 2025-04-01T04:33:03.631848 | {
"authors": [
"cc-prodsec",
"shethj"
],
"repo": "SalesforceCommerceCloud/pwa-kit",
"url": "https://github.com/SalesforceCommerceCloud/pwa-kit/pull/1288",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1830138657 | #1393: Add CheckboxGroup to shared ui components
Description
During the transition to the shared/ui library, an important ChakraUI component is omitted - CheckboxGroup. Since the Radio + RadioGroup are there, and the Checkbox is also there, I can only assume CheckboxGroup was omitted by mistake.
Types of Changes
[x] Bug fix (non-breaking change that fixes an issue)
[x] New feature (non-breaking change that adds functionality)
[ ] Documentation update
[ ] Breaking change (could cause existing functionality to not work as expected)
[ ] Other changes (non-breaking changes that does not fit any of the above)
Changes
include CheckboxGroup component in shared/ui folder
include useCheckboxGroup hook in shared/ui folder
include useCheckbox hook in shared/ui folder
How to Test-Drive This PR
Add a sample code somewhere in the app similar to this one
<CheckboxGroup defaultValues={['1']}> <Checkbox value="1">Option A</Checkbox> <Checkbox value="2">Option B</Checkbox> </CheckboxGroup>
Create custom checkbox component(s) using useCheckboxGroup and useCheckbox
Checklists
General
[ ] Changes are covered by test cases
[ ] CHANGELOG.md updated with a short description of changes (not required for documentation updates)
Accessibility Compliance
You must check off all items in one of the follow two lists:
[x] There are no changes to UI
or...
[ ] Changes were tested with a Screen Reader (iOS VoiceOver or Android Talkback) and had no issues
[ ] Changes comply with WCAG 2.0 guidelines levels A and AA
[ ] Changes to common UI patterns and interactions comply with WAI-ARIA best practices
Localization
[ ] Changes include a UI text update in the Retail React App (which requires translation)
See https://github.com/SalesforceCommerceCloud/pwa-kit/issues/1393#issuecomment-1660507942
| gharchive/pull-request | 2023-07-31T22:24:15 | 2025-04-01T04:33:03.640905 | {
"authors": [
"mdenchev-aiopsgroup",
"wjhsf"
],
"repo": "SalesforceCommerceCloud/pwa-kit",
"url": "https://github.com/SalesforceCommerceCloud/pwa-kit/pull/1394",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
221342058 | Push enhancements
Critical Changes
Changes
Push org lists (text files with one org ID per line) can now have comments and blank lines. The first word on the line is assumed to be the org ID and anything after that is ignored.
Issues Closed
Fixes #294
Fixes #306
Hooray for the batch retry functionality! Thanks for this.
| gharchive/pull-request | 2017-04-12T17:56:15 | 2025-04-01T04:33:03.642571 | {
"authors": [
"Joble",
"cpolcyn"
],
"repo": "SalesforceFoundation/CumulusCI",
"url": "https://github.com/SalesforceFoundation/CumulusCI/pull/307",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
833294617 | [EDA] People & Groups: Contact Information - Page and Contact Language
Critical Changes
Changes
Issues Closed
New Metadata
Apex Classes
ContactLanguageSettingsModel
ContactLanguageSettingsModel_TEST
ContactLanguageSettingsMapper
ContactLanguageSettingsMapper_TEST
ContactLanguageSettingsService
ContactLanguageSettingsService_TEST
ContactInformationSettingsVMapper
ContactInformationSettingsVMapper_TEST
ContactInformationSettingsController
ContactInformationSettingsController_TEST
ContactLanguageSettingsVModel
ContactLanguageSettingsVModel_TEST
Unpackaged Metadata
Deleted Metadata
Testing Notes
QA Passes - Merging - Closing WI
| gharchive/pull-request | 2021-03-17T00:28:03 | 2025-04-01T04:33:03.646633 | {
"authors": [
"m-kinkade",
"spinkelman"
],
"repo": "SalesforceFoundation/EDA",
"url": "https://github.com/SalesforceFoundation/EDA/pull/1621",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
479809780 | New component: Password
Do you plan to create an input type="password" specific component? Similar like the TextField just with masked characters.
@Satukan Here is how you can accomplish the password:
<MatTextField @bind-value="@loginParameters.Password" Label="Password" Icon="lock_outline" IconTrailing="true" FullWidth="true" Required="true" Type="password">
You can see a real example here: https://blazorboilerplate.com/account/login
My repo is here: https://github.com/enkodellc/blazorboilerplate
@Satukan if that solved your issue please close this! Thanks for using MatBlazor!
| gharchive/issue | 2019-08-12T19:23:46 | 2025-04-01T04:33:03.660124 | {
"authors": [
"Satukan",
"enkodellc"
],
"repo": "SamProf/MatBlazor",
"url": "https://github.com/SamProf/MatBlazor/issues/183",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
689265601 | Add required attribute to MatSelect
Describe the feature request
I would like to have a matselect component with a required attribute, which, as it says, it requires the user to select a value from one of the possible values.
Expected behavior
The attribute forces the user to select a value from one of the possible values.
A solution is to bind the MatSelect with a property with Required attribute and use the DataAnnotationsValidator
| gharchive/issue | 2020-08-31T14:47:18 | 2025-04-01T04:33:03.661639 | {
"authors": [
"fededim"
],
"repo": "SamProf/MatBlazor",
"url": "https://github.com/SamProf/MatBlazor/issues/698",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1026878867 | add new WFIT task
example inputs are on Midway under $SNANA_DEBUG/submit_batch_wfit
Pippin should auto-create a submit_batch input file of the form
WFIT_DEBUG.INPUT
where each INPDIR is an output directory of create_covariance.
Launch task with
submit_batch_jobs.sh WFIT_DEBUG.INPUT
Example pippin-input syntax is in pippin+wfit.yml
#73
| gharchive/issue | 2021-10-14T22:34:08 | 2025-04-01T04:33:03.670315 | {
"authors": [
"OmegaLambda1998",
"RickKessler"
],
"repo": "Samreay/Pippin",
"url": "https://github.com/Samreay/Pippin/issues/70",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
169714174 | Cannot add a scene object in X3D
scene.addSceneObject() does not work after PL https://github.com/Samsung/GearVRf/pull/774
@NolaDonato @mwitchwilliams any idea about what might be going wrong?
I’m up and running so I’ll be happy to look at it Monday.
Mitch
From: rahul [mailto:notifications@github.com]
Sent: Friday, August 5, 2016 4:30 PM
To: Samsung/GearVRf GearVRf@noreply.github.com
Cc: Mitchell Williams m1.williams@partner.samsung.com; Mention mention@noreply.github.com
Subject: Re: [Samsung/GearVRf] Cannot add a scene object in X3D (#781)
@NolaDonatohttps://github.com/NolaDonato @mwitchwilliamshttps://github.com/mwitchwilliams any idea about what might be going wrong?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHubhttps://github.com/Samsung/GearVRf/issues/781#issuecomment-237988807, or mute the threadhttps://github.com/notifications/unsubscribe-auth/ARSfvuUl6oWshItwZdVDV2UzxSlIVDteks5qc8dhgaJpZM4JeItg.
I will look at it. I am going to change jassimp to add a scene object after I read in an X3D file. Will that be a sufficient test?
From: rahul [mailto:notifications@github.com]
Sent: Friday, August 5, 2016 4:30 PM
To: Samsung/GearVRf GearVRf@noreply.github.com
Cc: Nola Donato nola.donato@samsung.com; Mention mention@noreply.github.com
Subject: Re: [Samsung/GearVRf] Cannot add a scene object in X3D (#781)
@NolaDonatohttps://github.com/NolaDonato @mwitchwilliamshttps://github.com/mwitchwilliams any idea about what might be going wrong?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHubhttps://github.com/Samsung/GearVRf/issues/781#issuecomment-237988807, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AQw1S8dVvyG2fN8e9cqVJKgvb4f3vUcxks5qc8dhgaJpZM4JeItg.
That should recreate it. I noticed the issue with the X3D demo.
I replaced astro_boy.dae with astro.x3d in the gvr-jassimp sample. It correctly displays the astronaut, the bench and the trees. I know that GVRAssetLoader calls GVRScene.addSceneObject to add each model to the scene. So adding a scene object after importing X3D seems to work.
Can you tell me a bit more about exactly when after importing the X3D model that you added the scene object? Are you absolutely sure that this new scene object is not already a child of another parent?
Thanks,
Nola
Here is the modified sample that does not work:
https://github.com/rahul27/GearVRf-Demos/commit/62129540745aee6d83510eee82655aa66236ccee
Make sure you sync to master so that the changes form PL 774 are included when you build the framework.
X3D (and I believe this is true of all Jassimp model loaders) never adds objects directly to GVRSceneObject 'scene' and thus never uses scene.addSceneObject() . The model loader passes a GVRModelSceneObject 'root' in the X3Dobject constructor. When the X3D parser adds GVRSceneObjects, it calls "root.addChildObject(myNewObject); or to currentSceneObject.addChildObject( myNewObject ) when we are not at the root. I'll look at the demo and suggest a fix soon but I believe that is the source of the issue
My sense is the renderer parses the Scene Heirarchy at this root object.
Found the bug. It is an NPE on the following line:
https://github.com/gearvrf/GearVRf-Demos/blob/master/gvr-x3d-demo/app/src/main/java/org/gearvrf/x3ddemo/X3DparserScript.java#L119
@NolaDonato seems to have fixed it in PL https://github.com/gearvrf/GearVRf-Demos/pull/288
| gharchive/issue | 2016-08-05T23:29:00 | 2025-04-01T04:33:03.682309 | {
"authors": [
"NolaDonato",
"mwitchwilliams",
"rahul27"
],
"repo": "Samsung/GearVRf",
"url": "https://github.com/Samsung/GearVRf/issues/781",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1263060281 | [Device] Add ExecutionEnv
This commit will add ExecutionEnv.
ONE-vscode-DCO-1.0-Signed-off-by: struss rrstrous@nate.com
From #744
after adding this, I will add each envType.
Thanks for @jyoungyun , UML for this completed.
https://github.com/Samsung/ONE-vscode/pull/744#issuecomment-1152136506
| gharchive/pull-request | 2022-06-07T10:15:52 | 2025-04-01T04:33:03.684594 | {
"authors": [
"struss"
],
"repo": "Samsung/ONE-vscode",
"url": "https://github.com/Samsung/ONE-vscode/pull/754",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
737777448 | Only require address_state on API
Description
Jira Ticket: N/A
This PR removes the requirement that address_line_1, address_city, and address_zip are set when creating or updating a Patient via the API.
Important Changes
patient.rb
No longer require the fields mentioned above.
Testing
To test this, you can try adding a patient via the API that is missing the above fields, and note that it is now added with no errors. Also, you will see that address_state is still required. If a patient is missing that field, it will not be added, and an error will be returned.
This is small enough that I think 1 reviewer is sufficient
| gharchive/pull-request | 2020-11-06T13:58:34 | 2025-04-01T04:33:03.786678 | {
"authors": [
"ngfreiter",
"tstrass"
],
"repo": "SaraAlert/SaraAlert",
"url": "https://github.com/SaraAlert/SaraAlert/pull/501",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2035496581 | 🛑 Secure is down
In 3286b85, Secure (https://secure.selgaraje.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Secure is back up in e936383 after 9 minutes.
| gharchive/issue | 2023-12-11T11:50:20 | 2025-04-01T04:33:03.798419 | {
"authors": [
"Sasillo"
],
"repo": "Sasillo/upptime",
"url": "https://github.com/Sasillo/upptime/issues/213",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1404454895 | 🛑 Cloud is down
In b7b0a1f, Cloud (https://cloud.selgaraje.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Cloud is back up in 860cfd0.
| gharchive/issue | 2022-10-11T11:13:33 | 2025-04-01T04:33:03.800669 | {
"authors": [
"Sasillo"
],
"repo": "Sasillo/upptime",
"url": "https://github.com/Sasillo/upptime/issues/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
487094501 | [Site request] Deviantart
Adding the md5 hasher to deviantart would be good.
I'm bad at github.
Suggestion has been added in ea16b0fb7f43eeeed8d68f59266c049f885ca8d2.
| gharchive/issue | 2019-08-29T17:49:18 | 2025-04-01T04:33:03.801562 | {
"authors": [
"Sasquire"
],
"repo": "Sasquire/Idems-Sourcing-Suite",
"url": "https://github.com/Sasquire/Idems-Sourcing-Suite/issues/5",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
1360608108 | Enhance the "Explanation" section of the documentation
Description:
With this PR I'm enhancing the pages of the "Explanation" section of the documentation.
The scope of changes is relatively small. Most of the time I'm simply adding "," or ".", or breaking paragraphs to make it easier to read. On other pages, I merely added references to some documentation that will help new people learn this tool and F#.
Let me know if there's something you'd like me to change.
I saw that the project documentation is not update yet when I visit the URL https://saturnframework.org/. What is it lacking to be updated?
Oh, blegh, it looks like docs changes don't get updated upon merge to the main branch. I'll see if I can adjust that
All right, thanks @cartermp 🚀
Ugh, okay, so there's probably an issue with using a very old version of Fornax here, so the docs aren't actually getting built and run anymore. Will investigate
Quick update - there's something that's a little fundamentally fucky going on here, likely related to the docs gen engine just being slow. I've looked at it a bit but haven't spent hours on it yet.
Ack, I noticed that there is something weird with the .NET setup in the Github Actions environment. It failed to find the SDK version that you tried 5.0.17.
https://github.com/SaturnFramework/Saturn/actions/runs/3032621248/jobs/4880300795
5.0.17 is a runtime version, not an SDK version - if you use a version like 5.0.x instead that should do what you want.
Oh yeah, this is me fucking around with github actions
I'm taking a look at this command and noticed some interesting points:
Most of the errors I got running locally were related to the file docs/generators/apiref.fsx;
I was able to fix some of them by updating the path for dll files in the beginning of the script. FSharp.Formatting is creating a folder named netstandard2.1 instead of netstandard2.0, and the name of the files inside it changed as well.
Now I'm only getting an error for the FSharp.MetadataFormat.dll, when running dotnet fornax build inside the docs/ folder.
Logs:
❯ dotnet fornax build
Load Errors: [|~/company/codes/foss/Saturn/docs/generators/apiref.fsx (6,1)-(6,88) parameter error Unable to find the file '../../packages/docs/FSharp.Formatting/lib/netstandard2.0/FSharp.MetadataFormat.dll' in any of
~/dotnet/shared/Microsoft.NETCore.App/6.0.2
~/dotnet/packs/Microsoft.NETCore.App.Ref/6.0.2/ref/net6.0
~/company/codes/foss/Saturn/docs
~/.nuget/packages/fornax/0.14.0/tools/net5.0/any/|]
Open Errors: [|input.fsx (1,6)-(1,12) typecheck error The namespace or module 'Apiref' is not defined.|]
Get generator Errors: [|input.fsx (1,16)-(1,24) typecheck error The value or constructor 'generate' is not defined. Maybe you want one of the following:
Generating
GenerationPhase
GeneratorConfig
GeneratorOutput
GeneratorTrigger|]
[15:53:13] multiple files generation failed
I'll keep investigating this, but apparently this MetadataFormat.dll does not exist anymore.
Great, the last import remaining was renamed. Found the information here:
https://github.com/fsprojects/FSharp.Formatting/blob/d0ac84e3a39221cc536d34a57e3f0f9dfec0f2df/RELEASE_NOTES.md
FSharp.MetadataFormat --> FSharp.Formatting.ApiDocs
So the imports must be:
// docs/generators/apiref.fsx
#r "../../packages/docs/FSharp.Formatting/lib/netstandard2.1/FSharp.Formatting.CodeFormat.dll"
#r "../../packages/docs/FSharp.Formatting/lib/netstandard2.1/FSharp.Formatting.Markdown.dll"
#r "../../packages/docs/FSharp.Formatting/lib/netstandard2.1/FSharp.Formatting.Literate.dll"
#r "../../packages/docs/FSharp.Formatting/lib/netstandard2.1/FSharp.Formatting.ApiDocs.dll"
PR: https://github.com/SaturnFramework/Saturn/pull/357
| gharchive/pull-request | 2022-09-02T20:59:45 | 2025-04-01T04:33:03.833735 | {
"authors": [
"64J0",
"baronfel",
"cartermp"
],
"repo": "SaturnFramework/Saturn",
"url": "https://github.com/SaturnFramework/Saturn/pull/354",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
521956939 | Crashes after logging in
connection = MixerConnection.ConnectViaLocalhostOAuthBrowser(clientID, scopes).Result;
Opens up the browser, logs in.
After login though, it just crashes. It goes to my Main function
static void Main() => MainAsync().GetAwaiter().GetResult(); with the error:
System.AggregateException: 'One or more errors occurred.'
Inner Exception
RestServiceRequestException: Bad Request
The setup is as follows:
Website: google.com
hosts: localhost
No secret key.
What is the actual contents of the RestServiceRequestException? I would need to see what the values are for the Request, StatusCode, Reason, & Content fields on it.
I'm not sure. I can't somehow check it as when I leave the step to see results, it just jumps to my Main. Is there a way to provide this somehow? If you want, it might be easier to communicate via discord? たここ#1196
Just throw a try-catch around that for type RestServiceRequestException and then you can explore the contents of the exception while debugging your application.
Content: "{"error":"invalid_client","error_description":"OAuth client secret invalid."}"
StatusCode: System.Net.HttpStatusCode.BadRequest
Request: "https://mixer.com/api/v1/oauth/token"
Reason: "Bad Request"
I'm not suer how my client secret is invalid when your ConnectViaLocalhostOAuthBrowser function only wants "clientID" and scopes.
The ConnectViaLocalhostOAuthBrowser method has multiple overloads, you're just using one that doesn't specify a client secret. However, because you're getting that error, you must have clicked the option on mixer.com/lab to generate a client secret for your application. Because of this, you are now required to always specify a client secret. So update your code to specify your client secret or build a new application without selecting the option to generate a client secret.
I don't have a client secret selected on /lab/.
| gharchive/issue | 2019-11-13T04:46:22 | 2025-04-01T04:33:03.848545 | {
"authors": [
"SaviorXTanren",
"takoz53"
],
"repo": "SaviorXTanren/mixer-client-csharp",
"url": "https://github.com/SaviorXTanren/mixer-client-csharp/issues/64",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2326142082 | Fixed The gap between the Faqs content and navbar is very less .
I fixed the overlapping of navbar and faq content by giving some margin to the top .
closes : #181
highlighting some edite
as we are about to declare winner want to make sure some edits need to be done, anyhow this wont affect your score drastically
| gharchive/pull-request | 2024-05-30T17:01:59 | 2025-04-01T04:33:03.854445 | {
"authors": [
"Ayushmaanagarwal1211",
"sanjay-kv"
],
"repo": "Sayak-Bhunia/mystory",
"url": "https://github.com/Sayak-Bhunia/mystory/pull/186",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
449599957 | Reports to show the difference in influence from the nearest faction and the change from last tick
Difference from the nearest faction
Change from last tick
Closed by e05e787
Unfortunately, the point 1 (Difference from the nearest faction) could not be implemented due to technical constraints.
| gharchive/issue | 2019-05-29T05:05:41 | 2025-04-01T04:33:03.855912 | {
"authors": [
"SayakMukhopadhyay"
],
"repo": "SayakMukhopadhyay/bgsbot",
"url": "https://github.com/SayakMukhopadhyay/bgsbot/issues/44",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
624101846 | 当文档内容有层级关系时无法正常工作
如图, 如果想在插入的文本框中使用参数填充功能无效,只能得到avatar这一个参数
文档地址(如果需要的话 https://self-file-server.oss-cn-shanghai.aliyuncs.com/disk/template.docx)
Hi,表格布局更专业哦 http://deepoove.com/poi-tl/#_常见问题
大神是指这个意思吗? 谢了😂
大神是指这个意思吗? 谢了😂
祝你成功~
| gharchive/issue | 2020-05-25T07:24:34 | 2025-04-01T04:33:03.863237 | {
"authors": [
"Sayi",
"zt1820057934"
],
"repo": "Sayi/poi-tl",
"url": "https://github.com/Sayi/poi-tl/issues/386",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2272302580 | 🛑 Prueba_Error_500 is down
In 34cf8ab, Prueba_Error_500 ($URL_ERROR_500) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Prueba_Error_500 is back up in 51931b9 after 10 minutes.
| gharchive/issue | 2024-04-30T19:25:59 | 2025-04-01T04:33:03.869722 | {
"authors": [
"Scaldy"
],
"repo": "Scaldy/ALADI_Sites_Monitor",
"url": "https://github.com/Scaldy/ALADI_Sites_Monitor/issues/448",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1382973253 | 🛑 PLGN is down
In 498f416, PLGN ($URL_PLGN) was down:
HTTP code: 0
Response time: 0 ms
Resolved: PLGN is back up in f80c318.
| gharchive/issue | 2022-09-22T20:24:05 | 2025-04-01T04:33:03.871776 | {
"authors": [
"Scaldy"
],
"repo": "Scaldy/ALADI_Sites_Monitor",
"url": "https://github.com/Scaldy/ALADI_Sites_Monitor/issues/52",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1642963135 | Drawable trait and pipeline construction assistance
handy drawable trait
infer the vertex buffer layouts in a const expression
each generated rust module will have a "create_render_pipeline" function that accepts a "PartialRenderPipelineDescriptor" and internally completes a true wgpu::RenderPipelineDescriptor to create the render pipeline
On point 3 I've defined "PartialRenderPipelineDescriptor" in lib.rs. This means wgsl_to_wgpu would have to be added to dependencies too, and not just build-dependencies, which I understand is not usual. This feels awkward to me so I welcome ideas on how to best organize/place this struct.
Thanks for taking a look. I'm still pretty new to gpu and shaders, but very committed to using this lovely library.
Here are some extra things to consider for the Drawable trait.
You could actually implement it with a tuple. Since the trait carries a lifetime 'd you can also do so with a tuple of references. So a single object/type doesn't have to be responsible for all bind groups, as long as you can bring together all these objects at the draw call.
However I didn't think about stuff being updated at different frequencies. It might be possible to update the trait to handle this though. For example each "get_bind_group" could return an Option<BindGroup> instead of a BindGroup, and it will only send it to the render_pass when it gets Some(bind_group). Then if some application has a rarely updated BindGroup, maybe their tuple that implements Drawable has some Optional elements (e.g. camera) that only get brought in when an update is actually necessary.
Perhaps this approach is not obvious, and so a contribution to the example project would be due. (Eyes on https://github.com/ScanMountGoat/wgsl_to_wgpu/issues/25)
On static enforcement of the order and number of buffer layouts, I think it's a bit of yes and no.
While const INFERRED_BUFFER_LAYOUTS does assume the number and order of buffer layouts, at the end of the day it is a const expression, not a type (no static enforcement). So anyone can provide their own expression of the layouts instead if they want.
However, as you know, the my_shader_rs::create_render_pipeline(device, partial_descriptor) does make use of the inferred buffer layouts. So this is a bit of layout presumption/enforcement.
Then with the Drawable trait, it provides default implementations for the inferred layout, but can be overriden. I'd consider this to not be layout enforcement. Sample:fn get_vertex_buffer0(&self) -> wgpu::BufferSlice<'d> {
return self.get_vertex_input_buffer();
}
fn get_vertex_buffer1(&self) -> wgpu::BufferSlice<'d> {
return self.get_instance_input_buffer();
}
fn get_vertex_input_buffer(&self) -> wgpu::BufferSlice<'d>;
fn get_instance_input_buffer(&self) -> wgpu::BufferSlice<'d>;
This becomes difficult to generate when modules have multiple entry points for each stage.
I did not even realize my own novice assumptions here. Yeah it looks like I use find_map to inadvertently assume there is only one vertex entry point and up-to-one fragment entry point. And so if I have N vertex entry points and M fragment entry points, and grant that the outputs/inputs are all compatible, there could be up to N*M valid pipeline constructions, yes?
So a single object/type doesn't have to be responsible for all bind groups, as long as you can bring together all these objects at the draw call.
Correct. wgpu already tracks the bound resources against the pipeline layout at runtime, so this would just be for checking for errors at compile time. While it can sometimes be helpful to think about renderable "objects", the graphics APIs themselves don't have this abstraction. The data can be set at different times or even more than once as long as everything is initialized in time for the draw or dispatch call.
Then if some application has a rarely updated BindGroup, maybe their tuple that implements Drawable has some Optional elements (e.g. camera) that only get brought in when an update is actually necessary.
This sounds like the builder pattern. The main issue is that bind groups can be set more than once. The draw is only invalid if any of the bind groups haven't been set yet. This should be possible to track in theory using something like a builder, but it would need to be its own issue and subsequent PR to work out the details.
And so if I have N vertex entry points and M fragment entry points, and grant that the outputs/inputs are all compatible, there could be up to N*M valid pipeline constructions, yes?
It's unlikely that all of these will be used in practice. The actual vertex buffer layouts will depend on which entry point the user selects. Some entry point combinations will have mismatches between each shader stage and could be considered "invalid". I'm not sure how difficult this is to detect or if it's worth preventing at build time.
After giving this some more thought, the code for 2. should be possible to implement in this PR if you're interested. You could create code that generates the entire wgpu::VertexState for each vertex entry point in the shader module. This would probably need to be a function since the step modes need to be provided by the user.
I would only want to consider 3. if it significantly reduced the amount of boilerplate code. The partial pipeline type ends up being almost as complicated as the full wgpu::RenderPipelineDescriptor. All we can really generate for the fragment state is the entry point, which could could just be addressed later by adding string constants for the entry point names.
The design of 1. is problematic in it's current state, but it does raise some interesting questions. Feel free to open an issue for further discussion. Enforcing correct initialization of bind groups and buffers are both worth looking into some more at some point even if it just ends up being part of the docs or example project.
This is all helpful feedback. Most of all I you've changed my perspective on the Draw trait. Even with the "Options" approach, it adds enough nuance to implementing the trait that it is not much simpler than just using the render pass directly.
I'll open a separate PR (less git history clutter) for 2 and close this PR.
I'll continue to experiment on these other ideas with your feedback in mind. I do enjoy type level builders.
| gharchive/pull-request | 2023-03-27T23:17:03 | 2025-04-01T04:33:03.885099 | {
"authors": [
"AndrewBrownK",
"ScanMountGoat"
],
"repo": "ScanMountGoat/wgsl_to_wgpu",
"url": "https://github.com/ScanMountGoat/wgsl_to_wgpu/pull/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2650090450 | [MODEL EVALUATION REQUEST] infly/OpenCoder-8B-Instruct
Model ID
infly/OpenCoder-8B-Instruct
Model type
Decoder model (e.g., GPT)
Evaluation languages
[X] Danish
[X] Swedish
[X] Norwegian (Bokmål or Nynorsk)
[X] Icelandic
[X] Faroese
[X] German
[X] Dutch
[X] English
Merged model
Not a merged model
Bug! #609
| gharchive/issue | 2024-11-11T18:40:48 | 2025-04-01T04:33:03.888570 | {
"authors": [
"saattrupdan"
],
"repo": "ScandEval/ScandEval",
"url": "https://github.com/ScandEval/ScandEval/issues/585",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
587783070 | Resize scanner when keyboard opens
Hi,
We are migrating the scandit plugin from 5.0 to 5.13.3.
The scanner is shown on 50% of the screen and other 50% shows an input box to enter value manually. When keyboard opens for the input box, the app webview shrinks but the scanner view is not shrinking or resizing accordingly. This makes the input box hide behind scanner.
This behavior is perfect in 5.0 version but somehow not working for 5.13.3
Can anyone help me here.
Hi @akshayajeevan,
Our support team will be able to help you best with this issue, please reach out to them at support@scandit.com.
Cheers,
Petra
I resized the scanner using higher margins in setMargins() when keyboard opens and back to lower margins when keyboard closes. This works fine when scanner is running.
That's great to hear! Thanks for letting us know this worked out for you.
| gharchive/issue | 2020-03-25T15:30:26 | 2025-04-01T04:33:03.892002 | {
"authors": [
"akshayajeevan",
"petradonka"
],
"repo": "Scandit/barcodescanner-sdk-cordova",
"url": "https://github.com/Scandit/barcodescanner-sdk-cordova/issues/262",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2144158064 | ⚠️ Jenkins Plugins Artifactory has degraded performance
In 600b61b, Jenkins Plugins Artifactory (https://get.jenkins.io/plugins/artifactory/) experienced degraded performance:
HTTP code: 200
Response time: 1170 ms
Resolved: Jenkins Plugins Artifactory performance has improved in c5565ec after 12 minutes.
| gharchive/issue | 2024-02-20T11:27:06 | 2025-04-01T04:33:03.920068 | {
"authors": [
"eifelmicha"
],
"repo": "Schmitzis/Uptime",
"url": "https://github.com/Schmitzis/Uptime/issues/401",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1812128838 | Move the 'free text' search up to a more prominent part of the main scicat page
Issue Name
Move the 'free text' search up to a more prominent part of the main scicat page
Summary
This is so that the box can be wider to allow more information to be displayed and entered, and so that the search is highlighted as a larger part of the process.
Transferred to frontend issue #1147
| gharchive/issue | 2023-07-19T14:51:38 | 2025-04-01T04:33:03.939544 | {
"authors": [
"markbasham",
"nitrosx"
],
"repo": "SciCatProject/scicat-backend-next",
"url": "https://github.com/SciCatProject/scicat-backend-next/issues/609",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2546034749 | refactor: migrate to LineSearch.jl
sooo many bad design decisions I had made are now returning to haunt me....
TODOs
[ ] use reinit!
Tests do pass locally. Now we need to just wait for SciMLJacobianOperators to merge, followed by LineSearch to get a tag.
bad push. fortunately have a copy on my lab machine...
| gharchive/pull-request | 2024-09-24T18:06:17 | 2025-04-01T04:33:03.954057 | {
"authors": [
"avik-pal"
],
"repo": "SciML/NonlinearSolve.jl",
"url": "https://github.com/SciML/NonlinearSolve.jl/pull/461",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2514006543 | Can't access Hatchet web admin GUI
Describe the bug
Fresh deploy of r2r in docker r2r serve --docker --config-name=local_llm_neo4j_kg. All containers come up healthy but when I attempt to access the Hatchet GUI to monitor running workflows I get a blank page that reads:
"Log in to Hatchet
No login methods are enabled."
To Reproduce
Steps to reproduce the behavior:
Launch r2r with r2r serve --docker --config-name=local_llm_neo4j_kg
Navigate to http://r2r-host.local:7274
Expected behavior
Expect to see that Hatchet GUI and to be able to login with default credentials per https://r2r-docs.sciphi.ai/cookbooks/orchestration
Screenshots
Desktop (please complete the following information):
OS: Ubuntu
Browser Firefox 130.0 (64-bit)
Version 24.04.1 LTS
Additional context
r2r version
3.1.14
I have seen this issue before, it seems to be something to do with caching inside the Hatchet web application.
Can you try clearing your cookies or logging in via incognito to see if you then correctly route to the login page?
Nailed it again, Sir. Logging in an Incognito window worked as expected. Thanks, much!
| gharchive/issue | 2024-09-09T13:39:40 | 2025-04-01T04:33:03.962161 | {
"authors": [
"br00t4c",
"emrgnt-cmplxty"
],
"repo": "SciPhi-AI/R2R",
"url": "https://github.com/SciPhi-AI/R2R/issues/1093",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1763186899 | Remove native libraries from LLama.csproj and replace it with a targets file.
The LLama.csproj deploys its native runtime dlls and dylibs into the output directory of whatever project references is. This differs from the NuGet, where the runtime files are not packaged. If you wanted to use this repo as a submodule into another project, you can't reference LLama.csproj directly because it's packaging those native libraries there.
Ideally, the same targets files you're shipping on nuget would be in this repo. Until then, I have a PR that can help fix this for those wanting to use the source directly. I removed the native libraries from LLama.csproj and put them into a targets file, and changed the example projects to reference that instead. If you don't want to use the runtime files, you can set IncludeBuiltInRuntimes to False and they won't be bundled, letting you bundle your own.
This does not affect the NuGet, as I can't find how you're building them in this repo. I'm assuming you're doing that locally with your own scripts.
Great work! Thanks a lot for your contribution!
This does not affect the NuGet, as I can't find how you're building them in this repo. I'm assuming you're doing that locally with your own scripts.
Yes, I build them locally by manually changing the files with Nuget Package Explorer. BTW, could you please give some guides about how to config the backend package settings so that it would be generated when building the project? (I'm not familiar with the MSBuild but I'll learn it if you'd like to give some guides, many thanks!)
| gharchive/pull-request | 2023-06-19T10:00:40 | 2025-04-01T04:33:03.965011 | {
"authors": [
"AsakusaRinne",
"drasticactions"
],
"repo": "SciSharp/LLamaSharp",
"url": "https://github.com/SciSharp/LLamaSharp/pull/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2406309132 | Include debug symbols with NuGet packages
Just adding symbols per the docs so users can step into the code more reliably.
Thanks!
| gharchive/pull-request | 2024-07-12T20:31:30 | 2025-04-01T04:33:03.966156 | {
"authors": [
"dpmm99",
"martindevans"
],
"repo": "SciSharp/LLamaSharp",
"url": "https://github.com/SciSharp/LLamaSharp/pull/851",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
235672652 | int to float promotion
See this comment.
Investigate casting and fix to ensure usage of numpy.promote_types.
No longer an issue since we're no longer using NaNs to represent the mask.
| gharchive/issue | 2017-06-13T19:51:11 | 2025-04-01T04:33:03.967337 | {
"authors": [
"bjlittle",
"djkirkham"
],
"repo": "SciTools/iris",
"url": "https://github.com/SciTools/iris/issues/2602",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
98918892 | 'fixed broken link in netcdf.py'
Fixed broken link to netcdf conventions on ESRL website.
The link now goes over the line limit, but I wasn't sure whether to keep this in or make a new line halfway through?
The link now goes over the line limit, but I wasn't sure whether to keep this in or make a new line halfway through?
The pep8 utility doesn't mind so I'd leave it as is.
NB. There's another instance of this URL at line 1693 which could also do with being updated.
Thanks - I've added that as well.
Thanks @scmc72! It might have been a simple one, but it is always good to get the first commit under your belt :wink:
| gharchive/pull-request | 2015-08-04T08:39:29 | 2025-04-01T04:33:03.970250 | {
"authors": [
"pelson",
"rhattersley",
"scmc72"
],
"repo": "SciTools/iris",
"url": "https://github.com/SciTools/iris/pull/1749",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2410957776 | clash-verge-rev@1.7.5: download via aria2 failed
Updating 'clash-verge-rev' (1.7.3 -> 1.7.5)
Downloading new version
Starting download with aria2 ...
Download: 07/16 19:58:02 [ERROR] CUID#8 - Download aborted. URI=https://github.com/clash-verge-rev/clash-verge-rev/releases/download/v1.7.5/Clash.Verge_1.7.5_x64_portabDownload: Exception: [AbstractCommand.cc:351] errorCode=3 URI=https://github.com/clash-verge-rev/clash-verge-rev/releases/download/v1.7.5/Clash.Verge_1.7.5_x64_portableDownload: Download Results: ound
Download: gid |stat|avg speed |path/URI
Download: ======+====+===========+=======================================================
Download: d3e823|ERR | 0B/s|E:/Application/Scoop/cache/clash-verge-rev#1.7.5#56ecd57.zip
Download: Status Legend:
Download: (ERR):error occurred.
Download: aria2 will resume download if the transfer is restarted.
Download: If there are any errors, then see the log file. See '-l' option in help/man page for details.
ERROR Download failed! (Error 3) Resource was not found
ERROR https://github.com/clash-verge-rev/clash-verge-rev/releases/download/v1.7.5/Clash.Verge_1.7.5_x64_portable.zip
referer=https://github.com/clash-verge-rev/clash-verge-rev/releases/download/v1.7.5/
dir=E:\Application\Scoop\cache
out=clash-verge-rev#1.7.5#56ecd57.zip
ERROR & 'E:\Application\Scoop\apps\aria2\current\aria2c.exe' --input-file='E:\Application\Scoop\cache\clash-verge-rev.txt' --user-agent='Scoop/1.0 (+http://scoop.sh/) PowerShell/7.4 (Windows NT 10.0; Win64; x64; Core)' --allow-overwrite=true --auto-file-renaming=false --retry-wait=2 --split=5 --max-connection-per-server=5 --min-split-size=5M --console-log-level=warn --enable-color=false --no-conf=true --follow-metalink=true --metalink-preferred-protocol=https --min-tls-version=TLSv1.2 --stop-with-process=19708 --continue --summary-interval=0 --auto-save-interval=1
Please update and re-check.
Seems assets for v1.7.5 were re-uploaded after release, hashes updated by #13631.
| gharchive/issue | 2024-07-16T11:58:45 | 2025-04-01T04:33:04.025313 | {
"authors": [
"aliesbelik",
"sparkslyz"
],
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/issues/13628",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
2512366696 | digital@0.31: hash check failed
ERROR Hash check failed!
App: extras/digital
URL: https://github.com/hneemann/Digital/releases/download/v0.31/Digital.zip
First bytes: 50 4B 03 04 0A 00 00 08
Expected: f165c40e55b702fe4e42935ac3aaa9f74cd1d18b9e296204064fff2e2c13f222
Actual: 12f014c8b99140554f8f7464ebc771bbe3de6af39c83c20463492bcb892afc69
Fixed by #13981.
| gharchive/issue | 2024-09-08T12:34:21 | 2025-04-01T04:33:04.027969 | {
"authors": [
"aliesbelik",
"yinyue200"
],
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/issues/13980",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
2546214654 | Prerequisites
[X] I have searched all issues/PRs to ensure it has not already been reported or fixed.
Criteria
[X] Reasonably well-known and widely used (e.g. if it's a GitHub project, it should have at least 100 stars and/or 50 forks)
[X] English interface (or at least English documentation)
[X] Latest stable version
[X] Full version (i.e. not a trial version)
[X] Fairly standard install (e.g. uses a version-specific download URL, no elaborate pre/post install scripts)
Name
aws-nuke (ekristen-fork)
Description
Allows deleting all resources of an AWS account. Actively maintained fork of aws-nuke.
Homepage
https://github.com/ekristen/aws-nuke
Download Link(s)
https://github.com/ekristen/aws-nuke/releases
Some Indication of Popularity/Repute
110 stars.
It's a fork of aws-nuke (which is no longer actively maintained), and is recommended as the go-to by the original aws-nuke package.
Unsure how best to name this -
ekristen-aws-nuke, aws-nuke-ekristen, or just aws-nuke 👀
| gharchive/issue | 2024-09-24T19:30:48 | 2025-04-01T04:33:04.032564 | {
"authors": [
"callumevans"
],
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/issues/14082",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
2567808513 | gimp@2.10.38-1: decompress error
gimp fails to unpacked and installed
gives following error in log file:
Signature detected: Inno Setup Setup Data (6.3.0)
This is not directly supported, but i'll try to unpack it as version 5602
; Version detected: 6300
Duplicates #14126.
| gharchive/issue | 2024-10-05T10:06:27 | 2025-04-01T04:33:04.034160 | {
"authors": [
"BLACKNOTES",
"aliesbelik"
],
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/issues/14152",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
1470572486 | mpv: Remove pre_install
mpv does not have an updater.bat, my bad.
Related to #9662
[x] I have read the Contributing Guide.
@niheaven
🤣
| gharchive/pull-request | 2022-12-01T02:43:31 | 2025-04-01T04:33:04.036162 | {
"authors": [
"niheaven",
"sitiom"
],
"repo": "ScoopInstaller/Extras",
"url": "https://github.com/ScoopInstaller/Extras/pull/9884",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
111840251 | Update to d3fc v1.5.0 causes bollinger bands to not display the series underneath
In v1.3.0, we can see the series below the bollinger bands.
Should probably ask Chris or Colin whether they think series should be drawn in front of indicators.
Already fixed in d3fc (not released yet though): https://github.com/ScottLogic/d3fc/commit/30fb8c4dd294c6dcf2853b57949a1df5a8683ef2
Fixed in d3fc 2.0.0 and above.
| gharchive/issue | 2015-10-16T14:10:46 | 2025-04-01T04:33:04.063850 | {
"authors": [
"djmiley",
"jleft"
],
"repo": "ScottLogic/d3fc-showcase",
"url": "https://github.com/ScottLogic/d3fc-showcase/issues/292",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1162581545 | Add ScottMyszka's file
Created a branch, created a file and made a commit, and opened a pull request
Hello
| gharchive/pull-request | 2022-03-08T12:11:19 | 2025-04-01T04:33:04.064665 | {
"authors": [
"ScottMyszka"
],
"repo": "ScottMyszka/github-slideshow",
"url": "https://github.com/ScottMyszka/github-slideshow/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1844451680 | Adding multiple Signal plots can decrease performance
ScottPlot Version: 4.1.66
Operating System: Windows
Application Type: WPF
Question: Adding multiple Signal plots can decrease performance.
// a code sample may improve communication
var wpfPlot = new WpfPlot();
for (int k = 0; k < 50; k++)
{
double[] data = Generate.RandomWalk(10000);
wpfPlot .AddSignalConst(data);
wpfPlot .Plot.AddSignalConst(data);
}
wpfPlot.Refresh(true);
Is there any solution?
Hi @Chfei1,
I modified the code a little bit but performance is fantastic. Here's a console app:
ScottPlot.Plot plt = new();
ScottPlot.RandomDataGenerator gen = new();
for (int k = 0; k < 50; k++)
{
double[] data = gen.RandomWalk(10_000);
plt.AddSignalConst(data);
}
// display it in a Windows Form
Form form = new() { Width = 600, Height = 400 };
ScottPlot.FormsPlot formsPlot1 = new() { Dock = DockStyle.Fill };
formsPlot1.Reset(plt);
formsPlot1.Refresh();
form.Controls.Add(formsPlot1);
form.ShowDialog();
Let me know if generating random data this way does not improve performance in the way you expect 👍
thanks,However, the larger the window, the lower the performance.
thanks,However, the larger the window, the lower the performance.
This is consistent with expectations. The laeger the bitmap, the longer it takes to render in memory.
ScottPlot 5 is still in preview but it offers hardware accelerated rendering for vastly improved performance over what System.Drawing (what ScottPlot 4 uses) can achieve. I look forward to it getting out of preview soon!
| gharchive/issue | 2023-08-10T05:40:44 | 2025-04-01T04:33:04.069411 | {
"authors": [
"Chfei1",
"swharden"
],
"repo": "ScottPlot/ScottPlot",
"url": "https://github.com/ScottPlot/ScottPlot/issues/2861",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1444862373 | CI: Bump actions
Purpose:
Node.js 12 actions are deprecated. The versions of the actions involved have been updated in this PR.
Details:
https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/
Thanks @MareMare! This wasn't on my radar but I'm glad you mentioned it 👍
| gharchive/pull-request | 2022-11-11T03:17:14 | 2025-04-01T04:33:04.071255 | {
"authors": [
"MareMare",
"swharden"
],
"repo": "ScottPlot/ScottPlot",
"url": "https://github.com/ScottPlot/ScottPlot/pull/2255",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
424021024 | [BUG] - Symbol.asyncIterator is not defined.
Describe the bug
I'm getting the following error when trying to run instamancer
TypeError: Symbol.asyncIterator is not defined.
at __asyncGenerator (/instamancer/src/api/instagram.js:4:38)
at Hashtag.generator (/instamancer/src/api/instagram.js:145:16)
at spawn (/instamancer/src/cli.js:238:41)
at <anonymous>
at process._tickCallback (internal/process/next_tick.js:188:7)
To Reproduce
I ran instamancer hashtag instagood -d after running through the installation steps on the README.
Setup (please complete the following information):
Node version - 6.9.0
Typescript version - 3.3.4000
Instamancer version - 1.3.1
You'll need to use an up to date version of Node. v11 or v10 LTS should work.
Duplicate of #1
| gharchive/issue | 2019-03-22T02:34:20 | 2025-04-01T04:33:04.086369 | {
"authors": [
"ScriptSmith",
"jessqnnguyen"
],
"repo": "ScriptSmith/instamancer",
"url": "https://github.com/ScriptSmith/instamancer/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1212969285 | 🛑 Tzar API is down
In c369c54, Tzar API (https://api.tzar.fr/health) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Tzar API is back up in d4a4bf6.
| gharchive/issue | 2022-04-22T23:38:23 | 2025-04-01T04:33:04.088712 | {
"authors": [
"xonlly"
],
"repo": "Scriptor-Group/status",
"url": "https://github.com/Scriptor-Group/status/issues/74",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
566499971 | Add Push-Button "Taster" as setting
Add the Push-Button funcionality to the common settings of the "Button Toogle". Like it is done in hqwidgets "On/Off".
Bei den widgets "hqwidgets" gibt es beim Button on/off die Funktion "Taster" in den allgemeinen Einstellungen. Für mich ist die sehr hilfreich, da ich öfter mit Tastern arbeite.
Versteh ich nicht. Kann doch ebenso als Taster verwendet werden. Man darf dann halt keine Icon / Farbe für on angeben, dann ist das ein Taster.
Bitte genaues Beispiel angeben und wie dann damit der Datenpunkt geschaltet werden soll, also angeben welche. Zustand die Werte annehmene sollen
IST: Beim "materialdesign - Button Toolge" wechselt der Datenpunkt bei jedem Drücken von false auf true und wieder auf false
Button gedrückt = true
Button nicht mehr gedrückt = true
Button gedrückt = false
SOLL: Es soll ein Datenpunkt (state, boolean (true/false)) nur dann von false auf true wechseln, wenn der entsprechende Button im VIS gedrückt ist.
Button gedrückt = true
Button nicht (mehr) gedrückt = false
Button gedrückt = true
D.h. der Datenpunkt soll nur solange true sein, solange der Button im VIS gedrückt ist.
Beim "hqwidget - On/Off" gibt es bei den Eigenschaften die Option "Taster". Dann reagiert dieser wie oben gewünscht.
OK jetzt hab ich es verstanden. Kann ich einbauen.
enthalten in c89f1423d96b63d52010fef55729ac40c2f02935
| gharchive/issue | 2020-02-17T20:19:46 | 2025-04-01T04:33:04.093367 | {
"authors": [
"IO-Link",
"Scrounger"
],
"repo": "Scrounger/ioBroker.vis-materialdesign",
"url": "https://github.com/Scrounger/ioBroker.vis-materialdesign/issues/58",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1631025175 | Create baseline using Finstreder
[ ] #14
[ ] #15
[ ] #16
[ ] #17
[ ] #18
[ ] #19
[ ] #20
[ ] #21
[ ] #22
The paper - https://arxiv.org/pdf/2206.14589v1.pdf
And the repo - https://gitlab.com/Jaco-Assistant/finstreder
| gharchive/issue | 2023-03-19T16:53:57 | 2025-04-01T04:33:04.113331 | {
"authors": [
"kordc"
],
"repo": "SebChw/audio-research-toolkit",
"url": "https://github.com/SebChw/audio-research-toolkit/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1725084550 | dataset
sequence.py
from pathlib import Path
import weakref
import cv2
import h5py
import numpy as np
import torch
from torch.utils.data import Dataset
from dataset.representations import VoxelGrid
from utils.eventslicer import EventSlicer
class Sequence(Dataset):
**### def __init__(self, seq_path: Path = 'C:\Users\hase0\OneDrive\바탕 화면\Dataset\thun_00_a_events_left', mode: str='train', delta_t_ms: int=50, num_bins: int=15):**
visualization -> eventreader.py
class EventReaderAbstract:
def __init__(self, filepath: Path = "C:\Users\hase0\OneDrive\바탕 화면\Dataset\thun_00_a_events_left\events.h5"):
assert filepath.is_file()
assert filepath.name.endswith('.h5')
self.h5f = h5py.File(str(filepath), 'r')
self._finalizer = weakref.finalize(self, self.close_callback, self.h5f)
eventvideo.py
if name == 'main':
parser = argparse.ArgumentParser('Visualize Events')
parser.add_argument('event_file', type=str, help='Path to events.h5 file')
parser.add_argument('output_file', help='Path to write video file')
parser.add_argument('--delta_time_ms', '-dt_ms', type=float, default=50.0, help='Time window (in milliseconds) to summarize events for visualization')
args = parser.parse_args()
event_filepath = Path(args.event_file)
video_filepath = Path(args.output_file)
dt = args.delta_time_ms
height = 480
width = 640
assert video_filepath.parent.is_dir(), "Directory {} does not exist".format(str(video_filepath.parent))
writer = skvideo.io.FFmpegWriter(video_filepath)
for events in tqdm(EventReader(event_filepath, dt)):
p = events['p']
x = events['x']
y = events['y']
t = events['t']
img = render(x, y, p, height, width)
writer.writeFrame(img)
writer.close()
eventvideo.py
if name == 'main':
parser = argparse.ArgumentParser('Visualize Events')
parser.add_argument('event_file', type=str, help='Path to events.h5 file')
parser.add_argument('output_file', help='Path to write video file')
parser.add_argument('--delta_time_ms', '-dt_ms', type=float, default=50.0, help='Time window (in milliseconds) to summarize events for visualization')
args = parser.parse_args()
event_filepath = Path(args.event_file)
video_filepath = Path(args.output_file)
dt = args.delta_time_ms
height = 480
width = 640
assert video_filepath.parent.is_dir(), "Directory {} does not exist".format(str(video_filepath.parent))
writer = skvideo.io.FFmpegWriter(video_filepath)
for events in tqdm(EventReader(event_filepath, dt)):
p = events['p']
x = events['x']
y = events['y']
t = events['t']
img = render(x, y, p, height, width)
writer.writeFrame(img)
writer.close()
[ 1.] error : usage: Visualize Events [-h] [--delta_time_ms DELTA_TIME_MS]
C:/Users/hase0/OneDrive/바탕 화면/events.h5 C:/Users/hase0/OneDrive/바탕 화면/output.mp4
Visualize Events: error: the following arguments are required: C:/Users/hase0/OneDrive/바탕 화면/events.h5, C:/Users/hase0/OneDrive/바탕 화면/output.mp4
d
| gharchive/issue | 2023-05-25T04:56:32 | 2025-04-01T04:33:04.128413 | {
"authors": [
"Sebongbak"
],
"repo": "Sebongbak/DSEC",
"url": "https://github.com/Sebongbak/DSEC/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2012488628 | Add new languages
feat: support toml (uses prettier-plugin-toml). Thanks @JounQin @so1ve
feat: support ini (uses prettier-plugin-ini). Thanks @kddnewton
feat: support latex (uses prettier-plugin-latex). Thanks @siefkenj
feat: support properties (uses prettier-plugin-properties). Thanks @eemeli
feat: support yaml (no plugins required). Thanks @ikatyang
feat: support json (no plugins required).
👇 Use the package built by codesandbox-ci below to test the new languages support.
https://github.com/un-ts/prettier/tree/master/packages/sh
Is there a hope of getting a new release soon (like within a week) with all these new languages?
I'd like to try it out, but wondering whether I should just wait a bit or should I pull this repo and fiddle around with it.
Is there a hope of getting a new release soon (like within a week) with all these new languages?
@nene Sorry for the delay, my mind is currently occupied by another project. I'll make some cleanups and hopefully release a new version in 48 hrs.
@Sec-ant You'd better find some collaborators.
@nene You should be able to test the prettier-plugin-sql-cst support in v0.4.0.
Note that the default sql plugin used for embedded SQL language formatting is still prettier-plugin-sql. So an additional option embeddedSqlPlugin is needed to set it to prettier-plugin-sql-cst. An example of the configuration would be:
{
"plugins": ["prettier-plugin-sql-cst", "prettier-plugin-embed"],
"embeddedSqlIdentifiers": ["sql"],
"embeddedSqlPlugin": "prettier-plugin-sql-cst",
"embeddedSqlParser": "sqlite",
"sqlKeywordCase": "upper",
}
And then you should be able to format code blocks that looks like this:
/* sql */`
ATTACH DATABASE 'my_file.sqlite' AS my_schema
`
or this:
sql`
ATTACH DATABASE 'my_file.sqlite' AS my_schema
`
If you need other strings to be recognized as identifiers, like sqlite or bigquery, add them to the embeddedSqlIdentifiers option.
{
"embeddedSqlIdentifiers": ["sql", "sqlite", "bigquery"],
}
I know the documentation of this plugin is still very immature and in early stage. If you have any questions regarding how to use this plugin, feel free to ask me. Comments, new issues, everywhere is ok.
Or is that possible to generate those information via linguist-languages like https://github.com/un-ts/prettier/blob/master/scripts/languages.ts#L33?
Thanks. I'll be playing around with it and will report if I run into issues.
I know the documentation of this plugin is still very immature and in early stage.
I find the documentation to be great. Considering this plugin is just a few months old, I'd say the docs are spectacular.
| gharchive/pull-request | 2023-11-27T15:09:34 | 2025-04-01T04:33:04.137110 | {
"authors": [
"JounQin",
"Sec-ant",
"nene"
],
"repo": "Sec-ant/prettier-plugin-embed",
"url": "https://github.com/Sec-ant/prettier-plugin-embed/pull/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2057824421 | Add Native Flow Response reply type definition
This PR adds the type definition for an interactive nfp_reply message. This message is received when sharing an address, or completing an flow.
Thanks for the PR! I didn't know this webhook notification existed. There are two nit details which will trigger a CI error, and one from the official documentation:
Here it says the "interactive" object will also contain an "action" property, and the "nfm_reply" might contain a "name" and "body" strings.
If you don't mind, could you fix and add those extra types? If not, don't worry, I can keep working from this.
Thanks!
Prefer using unknown over any, since it's type safer and doesn't make ESLint cry :)
Also, run npm run prettier to fix the nit issues.
Got ya, will fix the issues and run Lu ting locally before proposing the final changes. Thanks!
I added 2 more commits to the PR, super minor nit picks and replaced "address_response" with "address_message" based on the documentation.
Let me know if there's something else you would like to add, and if not I can merge it and release it tomorrow.
Thanks for contributing!
Awesome, this all looks good now. I do have some other changes, but they are not related to this one. They have to deal with post() being non-blocking and serverless environments don't like those scenario's.
Anyhow... thanks for the great work!
| gharchive/pull-request | 2023-12-27T22:43:36 | 2025-04-01T04:33:04.147280 | {
"authors": [
"Secreto31126",
"pjvds"
],
"repo": "Secreto31126/whatsapp-api-js",
"url": "https://github.com/Secreto31126/whatsapp-api-js/pull/282",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
787807379 | Support scalar-query to get a Json value out of a database
When relying on Json features of SQL Server, it could be handy to retrieve the value of a stored procedure or query that is already returnd as a Json string.
<json-source>
<scalar-query>exec MyStoredProcedure ("[{"id"=0}, {"id"=1}]")<scalar-query>
<json-path>...</json-path>
</json-source>
Implemented and documented in 1.23.0-beta.164 or on nuget via Update-Package NBi.Framework -version 1.23.0-beta0164.
Implemented and documented in 1.23.0-beta.164 or on nuget via Update-Package NBi.Framework -version 1.23.0-beta0164.
| gharchive/issue | 2021-01-17T22:15:34 | 2025-04-01T04:33:04.159970 | {
"authors": [
"Seddryck"
],
"repo": "Seddryck/NBi",
"url": "https://github.com/Seddryck/NBi/issues/632",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2109089973 | [Test] Add missing flow-based test for bip39 passphrase entry
Description
Adds a missing flow-based test for the process of entering an optional bip39 passphrase, reviewing the resulting fingerprint, editing the passphrase, once again reviewing the results, and then finalizing the seed.
This pull request is categorized as a:
[x] Test case
Checklist
[x] I’ve run pytest and made sure all unit tests pass before sumbitting the PR
If you modified or added functionality/workflow, did you add new unit tests?
[x] N/A
I have tested this PR on the following platforms/os:
[x] Other
LGTM
| gharchive/pull-request | 2024-01-31T01:06:25 | 2025-04-01T04:33:04.167090 | {
"authors": [
"kdmukai",
"newtonick"
],
"repo": "SeedSigner/seedsigner",
"url": "https://github.com/SeedSigner/seedsigner/pull/524",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
564416854 | MLX9060 I2C lib issue: AttributeError: 'SMBus' object has no attribute 'msg'
Hey,
I am having trouble with your MLX9060 I2C lib
I installed grove.py, followed tutorial :
https://pypi.org/project/seeed-python-mlx90640/#description
And everything goes fine up to the moment when I try to initialize the sensor:
import seeed_mlx90640
mlx = seeed_mlx90640.grove_mxl90640()
What I get is :
pi@raspberrypi:~ $ python3 infrared.py ## just a script name with copy&paste of the above two lines
Traceback (most recent call last):
File "infrared.py", line 2, in
mlx = seeed_mlx90640.grove_mxl90640()
File "/home/pi/.local/lib/python3.7/site-packages/seeed_mlx90640.py", line 58, in init
self.refresh_rate = RefreshRate.REFRESH_0_5_HZ
File "/home/pi/.local/lib/python3.7/site-packages/seeed_mlx90640.py", line 83, in refresh_rate
self._I2CReadWords(0x800D, controlRegister)
File "/home/pi/.local/lib/python3.7/site-packages/seeed_mlx90640.py", line 699, in _I2CReadWords
write = self.bus.msg.write(self.addr,[addr>>8,addr&0xFF])
File "/home/pi/.local/lib/python3.7/site-packages/grove/i2c.py", line 51, in getattr
return getattr(self.instance, name)
AttributeError: 'SMBus' object has no attribute 'msg'
I'm at loss and google gives me nothing. I can see the software is young. Any hints?
Thank you
you can use https://github.com/Seeed-Studio/grove.py#install-grovepy to update the newest Grove.py to resolve this problem.
Aright, so I did it again. I did:
`pi@raspberrypi:~/2 $ git clone https://github.com/Seeed-Studio/grove.py
Cloning into 'grove.py'...
remote: Enumerating objects: 5, done.
remote: Counting objects: 100% (5/5), done.
remote: Compressing objects: 100% (5/5), done.
remote: Total 1265 (delta 0), reused 2 (delta 0), pack-reused 1260
Receiving objects: 100% (1265/1265), 636.12 KiB | 1.20 MiB/s, done.
Resolving deltas: 100% (828/828), done.
pi@raspberrypi:~/2 $ cd grove.py/
pi@raspberrypi:~/2/grove.py $ sudo pip install .
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Processing /home/pi/2/grove.py
Requirement already satisfied: RPi.GPIO in /usr/lib/python2.7/dist-packages (from grove.py==0.6) (0.7.0)
Requirement already satisfied: rpi_ws281x in /usr/local/lib/python2.7/dist-packages (from grove.py==0.6) (4.2.3)
Requirement already satisfied: smbus2 in /usr/local/lib/python2.7/dist-packages (from grove.py==0.6) (0.3.0)
Building wheels for collected packages: grove.py
Running setup.py bdist_wheel for grove.py ... done
Stored in directory: /tmp/pip-ephem-wheel-cache-7jy7Mp/wheels/75/e9/7c/4e8984224bddd70b038a8d79d539c117533a9b707de39bc770
Successfully built grove.py
Installing collected packages: grove.py
Found existing installation: grove.py 0.6
Uninstalling grove.py-0.6:
Successfully uninstalled grove.py-0.6
Successfully installed grove.py-0.6
pi@raspberrypi:~/2/grove.py $ sudo pip3 install .
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Processing /home/pi/2/grove.py
Requirement already satisfied: RPi.GPIO in /usr/lib/python3/dist-packages (from grove.py==0.6) (0.7.0)
Requirement already satisfied: rpi_ws281x in /usr/local/lib/python3.7/dist-packages (from grove.py==0.6) (4.2.3)
Requirement already satisfied: smbus2 in /usr/local/lib/python3.7/dist-packages (from grove.py==0.6) (0.3.0)
Building wheels for collected packages: grove.py
Running setup.py bdist_wheel for grove.py ... done
Stored in directory: /tmp/pip-ephem-wheel-cache-eyjkh07x/wheels/75/e9/7c/4e8984224bddd70b038a8d79d539c117533a9b707de39bc770
Successfully built grove.py
Installing collected packages: grove.py
Found existing installation: grove.py 0.6
Uninstalling grove.py-0.6:
Successfully uninstalled grove.py-0.6
Successfully installed grove.py-0.6
`
Python version
pi@raspberrypi:~/grove.py $ python -V Python 3.7.3
I tried
curl -sL https://github.com/Seeed-Studio/grove.py/raw/master/install.sh | sudo bash -s -
But it does something like
`curl -sL https://github.com/Seeed-Studio/grove.py/raw/master/install.sh | sudo bash -s -
Warning: apt-key output should not be parsed (stdout is not a terminal)
Hit:1 http://archive.raspberrypi.org/debian buster InRelease
Hit:2 http://raspbian.raspberrypi.org/raspbian buster InRelease
Hit:3 https://seeed-studio.github.io/pi_repo stretch InRelease
Reading package lists... Done
Building dependency tree
Reading state information... Done
268 packages can be upgraded. Run 'apt list --upgradable' to see them.
I2C interface enabled...
Reading package lists... Done
Building dependency tree
Reading state information... Done
python-rpi.gpio is already the newest version (0.7.0~buster-1).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
python3-rpi.gpio is already the newest version (0.7.0~buster-1).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
libmraa1 is already the newest version (1.9.0-git20191021-pi20191021).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
python-mraa is already the newest version (1.9.0-git20191021-pi20191021).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
python-upm is already the newest version (1.6.0-git20191021-pi20191021).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
python3-mraa is already the newest version (1.9.0-git20191021-pi20191021).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
python3-upm is already the newest version (1.6.0-git20191021-pi20191021).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
libbma456 is already the newest version (1.0.0-1).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Reading package lists... Done
Building dependency tree
Reading state information... Done
libbmi088 is already the newest version (0.0.2-1).
0 upgraded, 0 newly installed, 0 to remove and 268 not upgraded.
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: rpi-ws281x in /usr/local/lib/python2.7/dist-packages (4.2.3)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: rpi-ws281x in /usr/local/lib/python3.7/dist-packages (4.2.3)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: smbus in /usr/local/lib/python2.7/dist-packages (1.1.post2)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: smbus in /usr/local/lib/python3.7/dist-packages (1.1.post2)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: smbus2 in /usr/local/lib/python2.7/dist-packages (0.3.0)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: smbus2 in /usr/local/lib/python3.7/dist-packages (0.3.0)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: bme680 in /usr/local/lib/python2.7/dist-packages (1.0.5)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: bme680 in /usr/local/lib/python3.7/dist-packages (1.0.5)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: rpi-vl53l0x in /usr/local/lib/python2.7/dist-packages (0.0.3)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: rpi-vl53l0x in /usr/local/lib/python3.7/dist-packages (0.0.3)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: sgp30 in /usr/local/lib/python2.7/dist-packages (0.1.6)
Requirement already satisfied: smbus2 in /usr/local/lib/python2.7/dist-packages (from sgp30) (0.3.0)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Requirement already satisfied: sgp30 in /usr/local/lib/python3.7/dist-packages (0.1.6)
Requirement already satisfied: smbus2 in /usr/local/lib/python3.7/dist-packages (from sgp30) (0.3.0)
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Collecting https://github.com/Seeed-Studio/grove.py/archive/master.zip
Cache entry deserialization failed, entry ignored
Downloading https://github.com/Seeed-Studio/grove.py/archive/master.zip
| 1.3MB 1.9MB/s^Z
[1]+ Stopped curl -sL https://github.com/Seeed-Studio/grove.py/raw/master/install.sh | sudo bash -s -
`
And freezes forever so I had to kill it, like, three times I tried. I understood that installing through pip3 should be equivalent. Also, just pip3 install should be enough and i did pip & pip3 but only because standalone pip3 install threw same errors so I just tried something else.
Some sample from grove.py lib
`import time
from grove.gpio import GPIO
led = GPIO(12, GPIO.OUT)
button = GPIO(22, GPIO.IN)
`
Throws no errors. Just initializing the mlx class.
It is very helpful for you to update the library by using sudo pip --force-reinstall . since the pip' mechanism.
Hello,
I'm receiving the same error running from a virtual environment:
File "/home/pi/volvis_sam_integration/venv/lib/python3.7/site-packages/seeed_mlx90640.py", line 58, in __init__
self.refresh_rate = RefreshRate.REFRESH_0_5_HZ
File "/home/pi/volvis_sam_integration/venv/lib/python3.7/site-packages/seeed_mlx90640.py", line 83, in refresh_rate
self._I2CReadWords(0x800D, controlRegister)
File "/home/pi/volvis_sam_integration/venv/lib/python3.7/site-packages/seeed_mlx90640.py", line 699, in _I2CReadWords
write = self.bus.msg.write(self.addr,[addr>>8,addr&0xFF])
File "/home/pi/volvis_sam_integration/venv/lib/python3.7/site-packages/grove/i2c.py", line 51, in __getattr__
return getattr(self.instance, name)
AttributeError: 'SMBus' object has no attribute 'msg'
python --version
Python 3.7.3
pip list
Package Version
--------------------- ---------
grove.py 0.6
numpy 1.18.2
opencv-contrib-python 3.4.4.19
opencv-python 3.4.4.19
picamera 1.13
pip 20.0.2
pyFirmata 1.1.0
pyserial 3.4
rpi-ws281x 4.2.3
RPi.GPIO 0.7.0
seeed-python-mlx90640 1.1.0
setuptools 46.1.3
smbus 1.1.post2
smbus2 0.3.0
wheel 0.34.2
I followed the Readme.md instructions to install grove.py. What am I missing?
Thanks in advance
do you use sudo pip --force-reinstall . to install.
I cloned grovy repo and reinstalled it. Now it seems to work. Anyway I will try that.
Thanks!
| gharchive/issue | 2020-02-13T03:32:04 | 2025-04-01T04:33:04.218390 | {
"authors": [
"Hansen0314",
"arituerto",
"impeer"
],
"repo": "Seeed-Studio/Seeed_Python_MLX90640",
"url": "https://github.com/Seeed-Studio/Seeed_Python_MLX90640/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
314874385 | Server does not appear to run void MapInstance::start() at all?
Issue Summary and Steps to Reproduce
Server does not appear to run void MapInstance::start() at all? Consequently, none of these functions fire:
m_world_update_timer.reset(new SEGSTimer(this,(void *)World_Update_Timer,world_update_interval,false)); // world simulation ticks
m_resend_timer.reset(new SEGSTimer(this,(void *)State_Transmit_Timer,resend_interval,false)); // state broadcast ticks
m_session_store.create_reaping_timer(this,Session_Reaper_Timer,reaping_interval); // session cleaning
Steps to Test
Add qWarning() << "MapInstance Start"; to the first line to test.
void MapInstance::start()
{
qWarning() << "MapInstance Start";
//...
SEGS build
3bfadac
The map instances are started for the first client that tries to connect to the given instance.
the logic goes like this:
in GameHandler::on_map_req ExpectMapClientRequest is sent to map server
MapServer::on_expect_client does map_manager().get_template(request_data.m_map_name) and tpl->get_instance()
MapTemplate::get_instance spins the instance up
hrm.. I can't get my qWarning() << "MapInstance Start"; code to ever fire.. maybe I'm overlooking something?
Whatever this issue was, it's not longer happening. Maybe it was a mirage from late night coding =P
| gharchive/issue | 2018-04-17T01:40:50 | 2025-04-01T04:33:04.226433 | {
"authors": [
"broxen",
"nemerle"
],
"repo": "Segs/Segs",
"url": "https://github.com/Segs/Segs/issues/335",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1669392830 | 🛑 🌯 Burritalks is down
In 7015dbc, 🌯 Burritalks (https://www.burritalks.io) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 🌯 Burritalks is back up in b04cfd3.
| gharchive/issue | 2023-04-15T13:21:04 | 2025-04-01T04:33:04.415397 | {
"authors": [
"SelrahcD"
],
"repo": "SelrahcD/uptime",
"url": "https://github.com/SelrahcD/uptime/issues/321",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1680244136 | 🛑 🌯 Burritalks is down
In 522439d, 🌯 Burritalks (https://www.burritalks.io) was down:
HTTP code: 0
Response time: 0 ms
Resolved: 🌯 Burritalks is back up in 67477d3.
| gharchive/issue | 2023-04-24T00:36:30 | 2025-04-01T04:33:04.418023 | {
"authors": [
"SelrahcD"
],
"repo": "SelrahcD/uptime",
"url": "https://github.com/SelrahcD/uptime/issues/327",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
183378889 | Dropdown 'openOnSpace' triggered through TextArea
Hey guys,
I've multiple Dropdown components at one page, and also other form components like TextArea included. When I use the dropdown components everything works fine, but when I then use my TextArea and press space inside, all my previous selected dropdowns are opening.
Looks like it has something to do with the search-attribute for the dropdown, when I remove this, it works.
Steps
Select item in dropdown
select textarea
start typing
press space
Expected Result
The listener for space-key should be removed for dropdowns if losing focus.
Actual Result
Dropdown/s is/are opened.
Version
0.56.0
Testcase
http://codepen.io/treckerflx/pen/ALPdPg?editors=0010
Thank you for the superb report. I've labeled this issue for fixing.
I've got a fix in progress, PR coming soon.
Released in semantic-ui-react@0.56.3
Thank you guys, incredible fast fix. :) 👍
No problem!
| gharchive/issue | 2016-10-17T10:23:57 | 2025-04-01T04:33:04.422425 | {
"authors": [
"fel1xw",
"levithomason"
],
"repo": "Semantic-Org/Semantic-UI-React",
"url": "https://github.com/Semantic-Org/Semantic-UI-React/issues/692",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
357245258 | chore(package): update deps
This PR updates projects dependencies and adds some changes to tests.
autoGenerateKey
Is now true by default in shorthand tests because it's true in factories, this allows make our tests more strict.
AccordionPanel
The update of enzyme allows us enable some tests for AccordionPanel :+1:
Codecov Report
Merging #3131 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #3131 +/- ##
=======================================
Coverage 99.92% 99.92%
=======================================
Files 163 163
Lines 2742 2742
=======================================
Hits 2740 2740
Misses 2 2
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 24d4334...9ffe067. Read the comment docs.
Released in semantic-ui-react@0.82.4.
| gharchive/pull-request | 2018-09-05T14:03:51 | 2025-04-01T04:33:04.428995 | {
"authors": [
"codecov-io",
"layershifter",
"levithomason"
],
"repo": "Semantic-Org/Semantic-UI-React",
"url": "https://github.com/Semantic-Org/Semantic-UI-React/pull/3131",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
669629487 | docs(Modal): improve examples, use React hooks
This PR:
improves examples for Modal component
refactors them use React hooks instead of class components
adds bundle size fixture
Codecov Report
Merging #4012 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #4012 +/- ##
=======================================
Coverage 99.84% 99.84%
=======================================
Files 185 185
Lines 3251 3251
=======================================
Hits 3246 3246
Misses 5 5
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 5ee5915...a2b3071. Read the comment docs.
| gharchive/pull-request | 2020-07-31T09:45:39 | 2025-04-01T04:33:04.434716 | {
"authors": [
"codecov-commenter",
"layershifter"
],
"repo": "Semantic-Org/Semantic-UI-React",
"url": "https://github.com/Semantic-Org/Semantic-UI-React/pull/4012",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
119819042 | Pointing Dropdown
Why does the pointing dropdown use box-shadow instead of border?
This is what the arrow looks like in Chrome 46 on OS X.
I was expecting something more like this:
@khornberg same feeling here; and what's more using box-shadow produces a tiny border-like effect on those border which shouldn't have it. (I'm not sure if it will be visible on every monitor, probably not, so i've zoomed it in to make it clearer.)
As you hopefully see pointer have a shadow leak, thus producing a border on bottom-left and bottom-right edges.
This may seem not a big thing, but it's a problem with clients having some mac monitors or other IPS'es as it's more visible, even without zooming in.
I got the second image with by doing this..
I don't remember exactly what I did and I've never gotten around to making the changes in code.
From
To
Fix
.ui.pointing.dropdown > .menu:after {
box-shadow: none !important;
background: #ffffff !important;
border: #e0e0e1 solid !important;
border-left-width: 1px !important;
border-right-width: 0px !important;
border-top-width: 1px !important;
border-bottom-width: 0px !important;
}
.ui.top.right.pointing.dropdown > .menu:after {
top: -.36em !important;
}
Note that those are manual overrides after SUI is built.
not now probot
please
| gharchive/issue | 2015-12-01T22:07:57 | 2025-04-01T04:33:04.440247 | {
"authors": [
"khornberg",
"plunntic"
],
"repo": "Semantic-Org/Semantic-UI",
"url": "https://github.com/Semantic-Org/Semantic-UI/issues/3421",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
350876936 | Haw to add semantic-ui-vue in Nuxtjs Project
who help me please !!!
https://github.com/Semantic-UI-Vue/Semantic-UI-Vue#nuxtjs-module
| gharchive/issue | 2018-08-15T16:30:31 | 2025-04-01T04:33:04.441467 | {
"authors": [
"aysnet1",
"mariolamacchia"
],
"repo": "Semantic-UI-Vue/Semantic-UI-Vue",
"url": "https://github.com/Semantic-UI-Vue/Semantic-UI-Vue/issues/222",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
265512801 | Add classMixin
Related issue: #23
Is it possible to implement it in the Statistic component to remove the inGroup prop? https://github.com/Semantic-UI-Vue/Semantic-UI-Vue/blob/master/src/views/Statistic/Statistic.jsx#L18
Added the mixin to Statistic. I noticed that even the official Semantic-UI group behavior is inconsistent between Statistic and other components (e.g. Item, Button, etc.).
Thanks a lot for your hard work, this looks amazing.
Added the mixin to Statistic. I noticed that even the official Semantic-UI group behavior is inconsistent between Statistic and other components (e.g. Item, Button, etc.)
Yes, true. This is actually a good question for the Semantic UI team. Another good reason to keep this logic out of the components
| gharchive/pull-request | 2017-10-14T18:28:29 | 2025-04-01T04:33:04.443853 | {
"authors": [
"mariolamacchia",
"viettrung9012"
],
"repo": "Semantic-UI-Vue/Semantic-UI-Vue",
"url": "https://github.com/Semantic-UI-Vue/Semantic-UI-Vue/pull/54",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
831245083 | logging crashes
after a while the cache clogs up so badly it takes a while to process anything triggering a timeout and exiting itself
Removed the cache file and using a transport to report the bug to the website were it can be viewed
| gharchive/issue | 2021-03-14T19:59:29 | 2025-04-01T04:33:04.466960 | {
"authors": [
"Seniatical"
],
"repo": "Seniatical/Mecha-Karen-Source-Code",
"url": "https://github.com/Seniatical/Mecha-Karen-Source-Code/issues/83",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1942203564 | 🛑 MedicoVet is down
In 26c7ff2, MedicoVet (https://medicovet.eu) was down:
HTTP code: 429
Response time: 856 ms
Resolved: MedicoVet is back up in 5a4e986 after 6 minutes.
| gharchive/issue | 2023-10-13T15:52:00 | 2025-04-01T04:33:04.483348 | {
"authors": [
"siimliimand"
],
"repo": "SeoWeb/upptime",
"url": "https://github.com/SeoWeb/upptime/issues/118",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2131863447 | 🛑 MedicoVet is down
In e56663c, MedicoVet (https://medicovet.eu) was down:
HTTP code: 429
Response time: 1918 ms
Resolved: MedicoVet is back up in ccf79d8 after 8 minutes.
| gharchive/issue | 2024-02-13T09:26:31 | 2025-04-01T04:33:04.485559 | {
"authors": [
"siimliimand"
],
"repo": "SeoWeb/upptime",
"url": "https://github.com/SeoWeb/upptime/issues/674",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1393489704 | Второй этап обработки данных, создание итогового вектора фрагментов
Идея в том, чтобы реализовать класс, который возвращает по сэмплу зашумлённый сэмпл с выбранным шумом.
Объект назовём MusicEffectGen.
Следующая структура:
MusicEffectGen
2.1. Различные функции, накладывающие звуковые эффекты
2.2. Функция суммирующая все остальные.
2.2.1. Получает в аргумент тип шума.
2.2.2. Возвращает функцию из 2.1. с нужным шумом.
2.2.3. Название : функция хранятся в специальном словаре.
[ ] 1. Собрать набор различных эффектов для обработки фрагментов музыки.
[ ] 1.1. Зацикливаемые фрагменты звуков по типу: шум толпы; звонкие, глухие удары; фейерверк; шум механизмов, машин; сигналы поездов и тп; скрип и тд, дисторшн.
[ ] 1.2. Обработка музыки непосредственно: затихание, непостоянная громкость(громкость по синусоиде), просто тихая и т.п.
[ ] 2. В DataGen cформировать вектор из обработанных музыкальных фрагментов.
simple di
mple
| gharchive/issue | 2022-10-01T17:43:48 | 2025-04-01T04:33:04.507757 | {
"authors": [
"Sergey-Shulnyaev",
"arti-shok"
],
"repo": "Sergey-Shulnyaev/ZamZam",
"url": "https://github.com/Sergey-Shulnyaev/ZamZam/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.