added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:36:41.124291
| 2022-10-21T07:38:59
|
1417898950
|
{
"authors": [
"Atanasov-AA",
"aschiakros",
"balsih"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:119",
"repo": "AkrosAG/Akros-Marketplace",
"url": "https://github.com/AkrosAG/Akros-Marketplace/issues/186"
}
|
gharchive/issue
|
Update my ads
Story
As an advertiser I want to be able to update my ads because the price or other values might change during the existence of an ad.
Preconditions
user is logged in
Post-Conditions
changes are saved in the corresponding database tables
Dependencies
This story can be started after Story Nr. 184 has been finished.
DoR
[x] External dependencies are identified or eliminated
[x] Design (at least LoFi?)
[x] Has acceptance criteria
Acceptance Criteria
[ ] AC 1: the user can see an update icon below each of his ads in "settings/my-ads".
[ ] AC 2: the user is able to click on the update icon of one of his ads and is then showed the same view as in the component "create", except that all values are prefilled as they were.
AC 1 could look like this ->
DoD
[ ] Story is tested against acceptance criteria
[ ] Unit test should be passed
[ ] Integration test is done ( if applicable)
[ ] Non-functional requirements are met
[ ] Story ok-ed by Product Owner
[ ] Peer Code Review performed
[ ] Any configuration or build changes documented
[ ] any database changes are documented Database
Changes are pushed to branch 186-update-my-ads
Might be combined with 292
@LanaKast the update functionality is not implemented, so I assume that it will be a big task because a form for collecting the user data should be implemented in the frontend and the backend should provide an endpoint for updating the topic.
|
2025-04-01T06:36:41.125629
| 2024-10-08T10:01:20
|
2572717249
|
{
"authors": [
"AkshitLakhera",
"Gauravtb2253"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:120",
"repo": "AkshitLakhera/PenCraft-Full-Stack-Blogging-Application",
"url": "https://github.com/AkshitLakhera/PenCraft-Full-Stack-Blogging-Application/issues/84"
}
|
gharchive/issue
|
Adding input validations
Hello I can add proper input validation for the login and register pages
Pls assign this to me
@Gauravtb2253 currently we are facing some problem in our auth apis ,we are working it.After it get fixed we take issues related to it.
|
2025-04-01T06:36:41.142983
| 2018-12-07T14:20:50
|
388673750
|
{
"authors": [
"AladW",
"alfunx"
],
"license": "isc",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:121",
"repo": "AladW/aurutils",
"url": "https://github.com/AladW/aurutils/pull/475"
}
|
gharchive/pull-request
|
aur-sync: update options in man page
Fixes #474
I think I got ahead of you: https://github.com/AladW/aurutils/commit/ae6cfb56ead53e16991492dcbc4be777844d2a1b
I didn't see a reason for leaving out the short options, so I readded them, rather than change the man page.
Sure that makes sense!
|
2025-04-01T06:36:41.158262
| 2024-07-19T11:05:23
|
2418636032
|
{
"authors": [
"blackfalcon",
"jason-capsule42"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:122",
"repo": "AlaskaAirlines/auro-backtotop",
"url": "https://github.com/AlaskaAirlines/auro-backtotop/pull/36"
}
|
gharchive/pull-request
|
perf: update to latest dependencies and linting
Alaska Airlines Pull Request
Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.
Resolves: #35
Summary:
Please summarize the scope of the changes you have submitted, what the intent of the work is and anything that describes the before/after state of the project.
Type of change:
Please delete options that are not relevant.
[ ] New capability
[x] Revision of an existing capability
[ ] Infrastructure change (automation, etc.)
[ ] Other (please elaborate)
Checklist:
[x] My update follows the CONTRIBUTING guidelines of this project
[x] I have performed a self-review of my own update
By submitting this Pull Request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
Pull Requests will be evaluated by their quality of update and whether it is consistent with the goals and values of this project. Any submission is to be considered a conversation between the submitter and the maintainers of this project and may require changes to your submission.
Thank you for your submission!
-- Auro Design System Team
:tada: This PR is included in version 2.1.9-beta.2 :tada:
The release is available on:
npm package (@beta dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:36:41.197645
| 2023-07-19T18:13:49
|
1812469782
|
{
"authors": [
"howardwu",
"ljedrz"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:123",
"repo": "AleoHQ/snarkVM",
"url": "https://github.com/AleoHQ/snarkVM/pull/1811"
}
|
gharchive/pull-request
|
Introduce ToBits::to_bits_le_into and use it in Vec/[T]
This PR targets the feat/narwhal branch, but it's quite likely it would result in a performance improvement in other use cases as well; it introduces a ToBits::to_bits_le_into method that allows us to avoid a lot of allocations when calling Vec::to_bits on large collections.
In a 1-minute run of test_state_coherence in snarkOS/narwhal, this PR reduces the number of allocations from ~4.2M to ~2M, and temporary allocations from ~2.2M to ~0.7M, as measured with heaptrack.
I've investigated the option of calling to_bits_le directly in BatchHeader::compute_batch_id, but that would require plenty of additional implementations of ToBits, and it wouldn't have the potential to reduce other allocations in snarkVM, so I decided against that.
I'm filing this as a draft until all the tests have run - it's possible that I might need to introduce another impl of ToBits::to_bits_le_into, which should be trivial.
before:
after:
The CI failures appear to be unrelated, so it's ready for review.
I am retargeting this PR to testnet3 since the changes are applicable to all snarkVM constructs that use this.
Rebased against testnet3 and added one additional commit which provides a small further improvement; that one was also found via profiling.
Superseded by https://github.com/AleoHQ/snarkVM/pull/1836; the drive-by commit will be filed as its own PR.
|
2025-04-01T06:36:41.324015
| 2023-05-26T08:02:28
|
1727114623
|
{
"authors": [
"AlexPshul",
"joelybahh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:124",
"repo": "AlexPshul/nxazure",
"url": "https://github.com/AlexPshul/nxazure/issues/16"
}
|
gharchive/issue
|
Question: Does this work generically with mono-repos? Or strictly NX at this stage?
I'm working in turbo-repo. My solution has been to write a script that overrides the path on the .json files so I can use the deployment extension, but if this can be configured to work in turborepo that would be amazing.
Hi @joelybahh ,
Unfortunately, I'm less familiar with turbo-repo and how it works.
Internally, in the executors, I'm using specific NX tools, so I don't know if this will work on other tools and monorepo frameworks.
@AlexPshul Thanks for the prompt reply!
I might do some digging into the source code in any case, although it's NX specific, I might be able to extract out the Azure parts and try to apply the same concepts to Turborepo, thanks again!
|
2025-04-01T06:36:41.354060
| 2021-10-07T18:47:17
|
1020351923
|
{
"authors": [
"AlexanderPrendota"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:125",
"repo": "AlexanderPrendota/kotlin-compiler-server",
"url": "https://github.com/AlexanderPrendota/kotlin-compiler-server/pull/378"
}
|
gharchive/pull-request
|
New samples
New samples in a repository https://github.com/JetBrains/kotlin-web-site
Files:
docs/topics/collection-transformations.md
docs/topics/constructing-collections.md
New samples in a repository https://github.com/JetBrains/kotlin-web-site
Files:
docs/topics/collection-transformations.md
docs/topics/constructing-collections.md
New samples in a repository https://github.com/JetBrains/kotlin-web-site
Commit: https://github.com/JetBrains/kotlin-web-site/commit/0cf51882694faa08341dc065cabdc631aca1fed8
Files:
docs/topics/jvm/java-to-kotlin-collections-guide.md
|
2025-04-01T06:36:41.359522
| 2024-05-15T23:18:08
|
2299009235
|
{
"authors": [
"LukashonakV",
"ZaheenJ"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:126",
"repo": "Alexays/Waybar",
"url": "https://github.com/Alexays/Waybar/issues/3275"
}
|
gharchive/issue
|
Unplugging and Replugging Laptop Crashes Waybar
Unplugging my laptop, waiting a few seconds, and then replugging my laptop consistently crashes waybar. Even after disabling the battery module, waybar still crashes. This could be related to #2519 and #2662. #2704 may solve this until inotify can possibly be replaced with netlink or upower events. The upower module would actually be perfect if it had format-charging like the battery module has.
Waybar Version: 0.10.3
Output:
[2024-05-15 15:31:26.609] [info] Using configuration file /home/zaheen/.config/waybar/config.jsonc
[2024-05-15 15:31:26.610] [info] Unable to receive desktop appearance: GDBus.Error:org.freedesktop.DBus.Error.UnknownMethod: No such interface “org.freedesktop.portal.Settings” on object at path /org/freedesktop/portal/desktop
[2024-05-15 15:31:26.610] [info] Using CSS file /home/zaheen/.config/waybar/style.css
[2024-05-15 15:31:26.637] [info] Hyprland IPC starting
[2024-05-15 15:31:26.637] [warning] $XDG_RUNTIME_DIR/hypr does not exist, falling back to /tmp/hypr
[2024-05-15 15:31:26.638] [info] Loading persistent workspaces from Waybar config
[2024-05-15 15:31:26.638] [info] Loading persistent workspaces from Hyprland workspace rules
[2024-05-15 15:31:26.672] [info] Loading persistent workspaces from Waybar config
[2024-05-15 15:31:26.672] [info] Loading persistent workspaces from Hyprland workspace rules
[2024-05-15 15:31:27.014] [info] Bar configured (width: 1900, height: 22) for output: eDP-1
[2024-05-15 15:31:27.014] [info] Bar configured (width: 1900, height: 22) for output: HDMI-A-1
[2024-05-15 15:31:27.051] [warning] Requested height: 22 is less than the minimum height: 23 required by the modules
[2024-05-15 15:31:27.051] [info] Bar configured (width: 1900, height: 23) for output: eDP-1
[2024-05-15 15:31:27.085] [warning] Requested height: 22 is less than the minimum height: 23 required by the modules
[2024-05-15 15:31:27.085] [info] Bar configured (width: 1900, height: 23) for output: HDMI-A-1
[2024-05-15 15:31:27.156] [warning] Requested height: 23 is less than the minimum height: 24 required by the modules
[2024-05-15 15:31:27.156] [info] Bar configured (width: 1900, height: 24) for output: eDP-1
[2024-05-15 15:31:27.215] [warning] Requested height: 23 is less than the minimum height: 24 required by the modules
[2024-05-15 15:31:27.215] [info] Bar configured (width: 1900, height: 24) for output: HDMI-A-1
/usr/include/c++/13.2.1/bits/stl_vector.h:1128: constexpr std::vector<_Tp, _Alloc>::reference std::vector<_Tp, _Alloc>::operator[](size_type) [with _Tp = std::tuple<long unsigned int, long unsigned int>; _Alloc = std::allocator<std::tuple<long unsigned int, long unsigned int> >; reference = std::tuple<long unsigned int, long unsigned int>&; size_type = long unsigned int]: Assertion '__n < this->size()' failed.
fish: Job 1, 'waybar' terminated by signal SIGABRT (Abort)
Hi @ZaheenJ , for UPower features it's better to open feature issue.
Even with the battery module and upower modules disabled, replugging the laptop crashed Waybar which should not happen. I don't have access to the original computer on which the crash occurred, but I can't seem to reproduce the issue on my M1 Macbook on Asahi Linux. I can try debugging more when I have access to the other laptop in about two weeks.
Can't seem to reproduce anymore.
|
2025-04-01T06:36:41.361421
| 2022-05-13T20:08:29
|
1235643266
|
{
"authors": [
"LukashonakV",
"tmpm697"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:127",
"repo": "Alexays/Waybar",
"url": "https://github.com/Alexays/Waybar/pull/1555"
}
|
gharchive/pull-request
|
Calendar module: localization issue #1552
Fixed week numbers right alignment when are used different locales(week is started from the SU or MO)
Hardcode prefix 'W' is removed in order to provide user an opportunity to use his own prefix via pango markup.
|
2025-04-01T06:36:41.448524
| 2018-05-24T14:34:52
|
326150803
|
{
"authors": [
"Alexpux",
"OTLabs"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:128",
"repo": "Alexpux/MSYS2-packages",
"url": "https://github.com/Alexpux/MSYS2-packages/issues/1266"
}
|
gharchive/issue
|
MSYS2 ruby is stalled at 2.4.0-2
I have installed ruby in MSYS2 some time ago, it was at version 2.4.0-2. I see that at https://github.com/Alexpux/MSYS2-packages/tree/master/ruby version 2.5.0 is avalable. I am unable to update to this version with pacman:
`$ pacman -Q ruby
ruby 2.4.0-2
$ pacman -S ruby
warning: ruby-2.4.0-2 is up to date -- reinstalling
resolving dependencies...
looking for conflicting packages...
Packages (1) ruby-2.4.0-2
Total Installed Size: 16.16 MiB
Net Upgrade Size: 0.00 MiB
:: Proceed with installation? [Y/n]`
It would be nice to get ruby updated to version 2.5.1 too. Thank you!
@OTLabs 32-bit ruby on MSYS/Cygwin crashing:
https://bugs.ruby-lang.org/issues/13999
Ruby 2.6.0 in repo now
|
2025-04-01T06:36:41.495167
| 2023-05-05T19:14:44
|
1698077572
|
{
"authors": [
"allyhawkins",
"jaclyn-taroni"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:129",
"repo": "AlexsLemonade/scpca-docs",
"url": "https://github.com/AlexsLemonade/scpca-docs/issues/107"
}
|
gharchive/issue
|
Tweak FAQ and downloaded files section for participant IDs to allow for future additions
There will be projects released shortly that do not currently have multiple samples that map to the same participant, but we expect future releases to include samples to meet that condition. The investigator has asked us to include participant IDs so that we don't have to make any revisions later.
As such, we're going to need to change the language in this FAQ to cover these cases: https://github.com/AlexsLemonade/scpca-docs/blob/5c66caa36f99b764f9f770e282b5787e318f3029/docs/faq.md#why-do-some-samples-have-missing-participant-ids
Edit: Also need to address this section: https://github.com/AlexsLemonade/scpca-docs/blob/5c66caa36f99b764f9f770e282b5787e318f3029/docs/download_files.md#metadata, so I've updated the title accordingly.
Anyone who implements this should find all "participant" to be sure everything has been addressed.
Closed by #108
|
2025-04-01T06:36:41.497099
| 2023-12-06T16:34:16
|
2028954260
|
{
"authors": [
"jashapiro",
"sjspielman"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:130",
"repo": "AlexsLemonade/scpca-nf",
"url": "https://github.com/AlexsLemonade/scpca-nf/pull/606"
}
|
gharchive/pull-request
|
Add version updates before writing metadata
I realized that there was a possibility that when skipping steps in workflows we might inadvertently propagate old version numbers, so I added a function and steps to make sure that the version numbers are correct before we write out any metadata files.
The good news is that we are getting the version info for the output reports and final metadata directly from Nextflow, so those were always correct, but some intermediate scpca-meta.json files might not always have been.
I don't think I care much about this variation, but I can make it uniform...
Yeah, it's small 🥔 . totally up to you.
|
2025-04-01T06:36:41.505571
| 2021-09-30T20:01:35
|
1012570616
|
{
"authors": [
"allyhawkins",
"jashapiro"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:131",
"repo": "AlexsLemonade/scpcaTools",
"url": "https://github.com/AlexsLemonade/scpcaTools/pull/57"
}
|
gharchive/pull-request
|
Add plots to QC report
Here I am mostly adding the Total vs. gene count plot to the QC report. I also fixed the knee plot and made the session info hidden by default, so closing all three of those issues here.
I was also playing around with #54, which I ended up abandoning, but in that process I added a bit of formatting to the tables, including making them not full-width, and striping the rows.
A few things to focus on that I was thinking about as I did them:
Please also look at all axis labels and legend titles! (Should I make the knee plot use non-scientific notation?)
For consistency among reports, I made the mitochondrial percentage go from 0-100% for all reports. This may make lower numbers harder to see in mostly-good plots (where no cells are above 50%, say). I thought the consistency was more important, but we could tweak the color scale if we think it matters.
The other potentially controversial aesthetic decision I made was to move the figure legends into the figure panels, at least for the knee & UMI/gene plots. In both of these plots, the location of the data is predictable, so in the vast majority of cases the positions I chose will not overlap data. Still, there is a chance they will, so if there is worry, I can move them back out. (when we add the miQC plot, the top right ought to work as a clear space)
I am attaching an example report (zipped, b/c github), and the two main figures I worked on below, for reference:
closes #39
closes #40
closes #19
closes #55
Adding in the miQC plot as well, which looks like this (not included in the example report above):
Same story here where I moved the legend inside the plot. In most cases the upper right should have no genes, so it ought to be safe, but I can move back outside if that is preferred.
I updated legends and titles as suggested, using sentence case everywhere, and unifying labels.
I moved what I had put as figure captions to alt text because it was laregely redundant, as noted, but I thought we should have some kind of figure label beyond the title.
For the knee plot, I chose a more contrasting color scheme, but I didn't want to go to red, because that would imply (to me) that those points failed some QC. So I went with a dark green, and lightened up the grey. I also went with smaller points for the filtered cells. Let me know what you think.
Here is a full report file...
SCPCL000001_qc.html.zip
I don't necessarily think we need the change in size and have previously been taught not to double encode variables where we have both size and color showing if a cell is considered passed or excluded but I'll leave that final call up to you.
Yeah, I don't usually do this, but I thought it was worth it in this case, as I think it solves a problem where the two kinds of points overlap.
Yeah, I don't usually do this, but I thought it was worth it in this case, as I think it solves a problem where the two kinds of points overlap.
I would agree, it does help show the separation a lot more. I think for the purposes of this plot its okay to keep it in.
|
2025-04-01T06:36:41.514984
| 2020-05-02T17:18:15
|
611234815
|
{
"authors": [
"barto64"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:132",
"repo": "AlexxIT/SonoffLAN",
"url": "https://github.com/AlexxIT/SonoffLAN/issues/88"
}
|
gharchive/issue
|
Could you please add the EWelink Smart Wifi Switch Relay Module Timer DC 5 B/12 B/24 B/32 to your component ? ?
Hi Alex,
For your reference:
https://aliexpress.ru/i/33027260480.html
It seems to be a TISHRIC TSR620/DC
Sonoff cloud info about is:
"1000926297":{"settings":{"opsNotify":0,"opsHistory":1,"alarmNotify":1,"wxAlarmNotify":0,"wxOpsNotify":0,"wxDoorbellNotify":0},"group":"","online":true,"shareUsersInfo":[],"groups":[],"devGroups":[],"_id":"5d84c2daa4285d754675c139","name":"SW_Puerta_Calle","type":"10","deviceid":"1000926297","apikey":"23effe28-d521-4d8f-8a79-e23c97e5c9d9","extra":{"extra":{"uiid":6,"description":"20190516002","brandId":"5735f5f906d9751d4f109629","apmac":"d0:27:01:24:c2:7f","mac":"d0:27:01:24:c2:7e","ui":"单通道开关","modelInfo":"589833ac2f979b623e2f503f","model":"PSF-B01-GL","manufacturer":"郑州市中原区汇诚电子材料经营部","staMac":"60:01:94:D5:BC:41","chipid":"00D5BC41"},"_id":"5cdd2fbb211f3b10753dee66"},"params":{"bindInfos":{"alexa":["621f7a3a-6015-449f-a4ae-9d7403cdc5bc_26ca1996a20e8bd63617ab272d4eeede1d2d8e32"]},"sledOnline":"on","switch":"off","fwVersion":"3.4.0","rssi":-86,"staMac":"60:01:94:D5:BC:41","startup":"off","init":1,"pulse":"off","pulseWidth":500,"version":8},"createdAt":"2019-09-20T12:15:22.938Z","__v":0,"onlineTime":"2020-05-01T07:15:33.008Z","ip":"<IP_ADDRESS>","location":"","offlineTime":"2020-05-01T07:14:51.754Z","tags":{"m_c9d9_jime":"on"},"devicekey":"621f7a3a-6015-449f-a4ae-9d7403cdc5bc","deviceUrl":"","brandName":"New Smart ","showBrand":true,"brandLogoUrl":"","productModel":"G1","devConfig":{},"uiid":6},
Alternatively if you give some hint I can look to the code and try a contribution on my own. I look to your code and looks pretty well organized and done. Only issue is that my russian is really poor, not more of 10-15 words after 10y travelling frequently to Moscow for work.
Apologies, it work fine. My fault.
|
2025-04-01T06:36:41.575491
| 2023-07-08T23:48:22
|
1795168103
|
{
"authors": [
"neonWhiteout"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:133",
"repo": "AlgebraicJulia/StockFlow.jl",
"url": "https://github.com/AlgebraicJulia/StockFlow.jl/pull/35"
}
|
gharchive/pull-request
|
Added add_suffix! and add_prefix! functions to SystemStructure to easily change names, and corresponding set_sname!/set_snames!, etc to StockFlow
Edits to SystemStructure may require further documentation. Currently writing tests for it, though seemed to work in informal tests so far.
Specifically, these change stock, flow, sum variable, dynamic variable and parameters names to feet, stock flows, and open stock flows.
I think the changes are done on the wrong branch, this should be done on a branch based off 7e65d31
|
2025-04-01T06:36:41.605612
| 2018-07-12T08:50:51
|
340545474
|
{
"authors": [
"CyberiaResurrection",
"c-harris",
"donmbelembe",
"dsv4890"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:134",
"repo": "Algo-Web/POData-Laravel",
"url": "https://github.com/Algo-Web/POData-Laravel/issues/170"
}
|
gharchive/issue
|
how to configure this in laravel ?
what will be next step after adding trait in model ?
go to
http://localhost:8000/odata.svc
http://localhost:8000/odata.svc/$metadata
http://localhost:8000/odata.svc/ModelName
please let me know if it works for you because i'm getting this error
A non well formed numeric value encountered
Do i have to declare this path in my web.php file right? how can i use all
that function like $count, $top ect..
Best Regards,
Dhruvisha Vasavada
On Thu, Jul 12, 2018 at 2:33 PM donmbelembe<EMAIL_ADDRESS>wrote:
go to
http://localhost:8000/odata.svc
http://localhost:8000/odata.svc/$metadata
http://localhost:8000/odata.svc/ModelName
please let me know if it works for you because i'm getting this error
A non well formed numeric value encountered
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/Algo-Web/POData-Laravel/issues/170#issuecomment-404443346,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AkqxyYLE4jHmy87ge0bWc7MMOa9R-a2zks5uFxDfgaJpZM4VMc4V
.
Do i have to declare this path in my web.php file right? how can i use all
that function like $count, $top ect..
Have you exposed the model with the trait?
The routes are defined by the route service provider.
@donmbelembe where are you encountering that error
http://localhost:8000/odata.svc/$metadata this URL is working for me but when I write ModelName it redirecting back to a login page
I'm also new to Odata
so i don't know, just trying out
this link: http://localhost:8000/odata.svc/$metadata
produce this error _http://localhost:8000/odata.svc/$metadata_
when I write ModelName it redirecting to Home page
@c-harris yeah and when trying to access http://localhost:8000/odata.svc i'm getting A non well formed numeric value encountered on \vendor\symfony\http-foundation\Response.php
Do we need to add Class 'Symfony\Component\Yaml\Yaml' this manually in the composer
i have checked in composer.json this class is already there still I'm getting the same error on this link: http://localhost:8000/odata.svc/$metadata
I've installed composer require symfony/yaml
i'm not getting this error anymore Class 'Symfony\Component\Yaml\Yaml' not found
but still my model redirecting to /home and
http://localhost:8000/odata.svc/$metadata & http://localhost:8000/odata.svc gives this error A non well formed numeric value encountered on \vendor\symfony\http-foundation\Response.php
I'm also getting the same error with laravel 5.6
@donmbelembe , @dsv4890 , can you try again, please? I've rolled up @NoelDeMartin 's pull request that fixed at least one underlying issue.
thank you ,
but unfortunately after updating to the last version and set the trait to the User Model, i'm getting the following error
`php artisan package:discover
In MetadataProvider.php line 140:
Undefined index: User`
@donmbelembe , thanks for your feedback.
Blast. I'm working on a pull request at the moment that might get you out of a jam (#171 ) - if I give you the details, are you able to update with them and see if pain persists?
of course @CyberiaResurrection
Here we go - you'll need to tweak your project's composer.json file somewhat:
1 - Add the following section:
"repositories": [ { "type": "vcs", "url": "https://github.com/CyberiaResurrection/podata-laravel.git", "no-api": true } ],
2 - Change your podata-laravel dependency line as follows:
"algo-web/podata-laravel": "dev-CleanUpAsserts as dev-master",
After that, you'll need to run composer update. Could you let me know how that goes?
I'm still arguing with getting the test suite working under Laravel 5.6.
Thanks dude, I followed your steps, now http://localhost:8000/ and http://localhost:8000/odata.svc/$metadata are returning good response, however When trying to access my model like this http://localhost:8000/odata.svc/ModelName It's redirecting to home page
@donmbelembe , thanks heaps for your feedback and confirming a certain wombat (ie, @CyberiaResurrection) had not mucked that particular bit up.
@CyberiaResurrection should we perhaps consider disabling auth by default as it is not detailed in odata itself?
I direct my learned colleague's attention to the README.
@donmbelembe , can you try adding the following line to your project's .env file?
APP_DISABLE_AUTH=true
@CyberiaResurrection About your question, When I'm not logged in it is redirecting to the login page and when it's already logged in it is redirecting to /home. actually my home page has auth middleware
The trait is already set in my model but http://localhost:8000/odata.svc/ModelName is redirecting to /home
@CyberiaResurrection even with APP_DISABLE_AUTH=true i'm getting the same result
@donmbelembe , that's a new problem. APP_DISABLE_AUTH should do what it says on the tin - thanks for rumbling a spot where I've outsmarted myself.
@donmbelembe , I've pushed out what I hope is a fix for APP_DISABLE_AUTH - does that get you out of trouble?
I think It works now.
So http://localhost:8000/odata.svc/User return an xml with 404 status code but http://localhost:8000/odata.svc/Users return data, so this mean we must call call the model in plural ?
Maybe I should learn more about Odata, I'm testing first
@donmbelembe , thanks for your continued feedback and patience.
We followed Laravel convention and thus pluralised the model names to use as endpoints. That's an implementation choice in POData-Laravel, not anything intrinsic to OData itself.
Okey. Thanks to you also for your help
Thanks for your bug reports. As you've confirmed that your issue is fixed in that branch, I'll close this issue.
@donmbelembe , I've rolled those changes up, so you'll need to undo those changes I asked you to make in your project's composer.json file.
|
2025-04-01T06:36:41.628780
| 2020-08-11T01:03:02
|
676505131
|
{
"authors": [
"Grandmother",
"ishc3ice",
"koles289",
"lobomfz",
"mgleed"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:135",
"repo": "AliaksandrSiarohin/first-order-model",
"url": "https://github.com/AliaksandrSiarohin/first-order-model/issues/226"
}
|
gharchive/issue
|
Problem with docker build
I'm struggling to get the proper linux environment to work for this, so I've resorted to using docker. I eventually got the program to work with the provided docker build instructions about a week or so ago on WSL2, but it doesn't seem to work for me any more.
docker build -t first-order-model . fails for me on this step: full log
Collecting opencv-python (from face-alignment==1.1.0)
Downloading https://files.pythonhosted.org/packages/a1/d6/8422797e35f8814b1d9842530566a949d9b5850a466321a6c1d5a99055ee/opencv-python-<IP_ADDRESS>.tar.gz (88.0MB)
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-build-wuc0x2vh/opencv-python/setup.py", line 9, in <module>
import skbuild
ModuleNotFoundError: No module named 'skbuild'
Any ideas to get this working again?
Running on a fresh WSL2 Ubuntu build following the instructions here. Building with a fresh clone of this repo as well.
Thank you!
Hi, I have exactly the same problem after downloading OpenVino toolkit for linux with FPGA support version 2020.4.287.
the same issue is being solved in this thread, but I do not know, how to download changed files.
https://github.com/OpenVisualCloud/Dockerfiles/issues/549
Hi, this problem is described in issue in official opencv-python repo and in their FAQ.
Submited a pull request to fix it
Submited a pull request to fix it
This worked, thank you
Edit the Dockerfile, and add RUN pip3 install scikit-build==0.11.1 before the last Command
about like this:
FROM nvcr.io/nvidia/cuda:10.0-cudnn7-runtime-ubuntu18.04
RUN DEBIAN_FRONTEND=noninteractive apt-get -qq update \
&& DEBIAN_FRONTEND=noninteractive apt-get -qqy install python3-pip ffmpeg git less nano libsm6 libxext6 libxrender-dev \
&& rm -rf /var/lib/apt/lists/*
COPY . /app/
WORKDIR /app
RUN pip3 install scikit-build==0.11.1
RUN pip3 install setuptools
RUN pip3 install \
https://download.pytorch.org/whl/cu100/torch-1.0.0-cp36-cp36m-linux_x86_64.whl \
git+https://github.com/1adrianb/face-alignment \
-r requirements.txt
|
2025-04-01T06:36:41.732801
| 2017-12-12T14:19:43
|
281400490
|
{
"authors": [
"AllenFang",
"rikkirabz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:136",
"repo": "AllenFang/react-bootstrap-table",
"url": "https://github.com/AllenFang/react-bootstrap-table/issues/1805"
}
|
gharchive/issue
|
Ability to sort column either by number or letter
Hi! Is there any way to add 2 options to the header of a column (which has alpha numeric data) which allows the user to either sort by number or by letter? Thank you for a job well done!
@rikkirabz sorry for lately reply. Currently, it's hard to achieve your requirement in the react-bootstrap-table. But I'll consider to enhance it on react-bootstrap-table2 when I have time,
Thank you :)
|
2025-04-01T06:36:41.778694
| 2021-11-17T03:22:04
|
1055660086
|
{
"authors": [
"LuQQiu",
"lilyzhoupeijie"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:137",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/issues/14531"
}
|
gharchive/issue
|
worker heap memory use high
Alluxio Version:
alluxio 2.6.1
Describe the bug
alluxio worker use memory as follows:
this one only shows heap memory and only alluxio worker use without job worker
worker jvm setting:
ALLUXIO_WORKER_JAVA_OPTS: ' -Dalluxio.worker.rpc.port=30062 -Dalluxio.worker.web.port=30063
-Dalluxio.worker.container.hostname=${ALLUXIO_WORKER_CONTAINER_HOSTNAME} -Dalluxio.worker.memory.size=70Gi
-Dalluxio.worker.hostname=${ALLUXIO_WORKER_HOSTNAME} -Xmx20g -Xms20g -XX:MaxDirectMemorySize=50g'
And the metaspaceSize show as follows:
jinfo -flag MaxMetaspaceSize 1
-XX:MaxMetaspaceSize=18446744073709547520
jinfo -flag MetaspaceSize 1
-XX:MetaspaceSize=21807104
To Reproduce
spark write to allxuio
Expected behavior
can this be smaller?
6 spark jobs write to alluxio ,each spark job has 20 executor and 4 cpu * 8 memory per executor
@lilyzhoupeijie are you able to get one or more heap dump when the worker memory usage is high?
@yuzhu any suggestions?
K8S will calculate memory by adding process memory + process page cache and will issue OOM when using more than 100GB
process memory is small
K8S don't have ways to clear page cache
worker process with new process cleaning page cache
another manual way cleaning process page cache
Close because K8S constraint
|
2025-04-01T06:36:41.797331
| 2021-11-01T15:59:11
|
1041329901
|
{
"authors": [
"Xenorith",
"codecov-commenter",
"ggezer",
"jiacheliu3"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:138",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/pull/14363"
}
|
gharchive/pull-request
|
Fix BlockMasterIntegrityIntegrationTest for worker stream register
This test fails when worker registers with a stream, due to how the operation are rearranged for the streaming registration.
Although this is on the same codepath for unary RPC register, this does not affect the correctness.
@HelloHorizon fyi
Codecov Report
Merging #14363 (eed1910) into master (7f30116) will decrease coverage by 25.06%.
The diff coverage is 0.00%.
@@ Coverage Diff @@
## master #14363 +/- ##
=============================================
- Coverage 41.96% 16.90% -25.07%
+ Complexity 9341 2698 -6643
=============================================
Files 1492 1492
Lines 87257 87258 +1
Branches 10417 10417
=============================================
- Hits 36621 14753 -21868
- Misses 47667 71266 +23599
+ Partials 2969 1239 -1730
Impacted Files
Coverage Δ
.../java/alluxio/master/block/DefaultBlockMaster.java
40.62% <0.00%> (-35.40%)
:arrow_down:
...mon/src/main/java/alluxio/shell/CommandReturn.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...mon/src/main/java/alluxio/util/ExceptionUtils.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...n/src/main/java/alluxio/wire/AlluxioProxyInfo.java
0.00% <0.00%> (-100.00%)
:arrow_down:
.../src/main/java/alluxio/wire/AlluxioMasterInfo.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...src/main/java/alluxio/job/meta/JobIdGenerator.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...n/src/main/java/alluxio/stress/BaseParameters.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...src/main/java/alluxio/client/UnderStorageType.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...main/java/alluxio/underfs/options/ListOptions.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...main/java/alluxio/worker/block/io/BlockReader.java
0.00% <0.00%> (-100.00%)
:arrow_down:
... and 684 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 7f30116...eed1910. Read the comment docs.
alluxio-bot, merge this please
alluxio-bot, cherry-pick this to branch-2.7 please
alluxio-bot, cherry-pick this to branch-2.7 please
|
2025-04-01T06:36:41.800899
| 2018-04-20T20:15:57
|
316398406
|
{
"authors": [
"AmplabJenkins",
"calvinjia"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:139",
"repo": "Alluxio/alluxio",
"url": "https://github.com/Alluxio/alluxio/pull/7183"
}
|
gharchive/pull-request
|
[SMALLFIX] Concurrency test speed up
100ms sleep is okay on my local machine.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Alluxio-Pull-Request-Builder/19332/
Test PASSed.
Merged build finished. Test PASSed.
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Alluxio-Pull-Request-Builder/19367/
Test PASSed.
|
2025-04-01T06:36:41.804896
| 2021-01-31T21:05:01
|
797811731
|
{
"authors": [
"AlmasB"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:140",
"repo": "AlmasB/FXGL",
"url": "https://github.com/AlmasB/FXGL/issues/957"
}
|
gharchive/issue
|
Design and implement default InventoryView
In the fxgl-trade module we have Inventory API, we now need to provide a default InventoryView with some reasonable settings such as width and height and a default look.
Most of the time, developers will produce their own views using Inventory.
We should consider if the "sort" functionality responsibility of the view or the model. Example: sort by name...
Something like the inventory in Zephyria
|
2025-04-01T06:36:41.823375
| 2020-07-09T14:40:23
|
654115024
|
{
"authors": [
"JamesSmartCell",
"colourfreak",
"hboon",
"tomekalphawallet"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:141",
"repo": "AlphaWallet/alpha-wallet-android",
"url": "https://github.com/AlphaWallet/alpha-wallet-android/issues/1504"
}
|
gharchive/issue
|
Improve Smart QR Code Scanner [Fiji]
It is a bit complex task.
Zeplin:
When scanning ETH Address
a) Send to this Address https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f641b060131115f1bc1c6
b) Add to Address Book https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f648d82a344183b52c43c
c) Watch Wallet https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f64c5fae46e11eefb46ae
When scanning Payment Request
https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f6531bc13e3160fe28da9
When scanning to add custom token
https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f63483eea5014c74bec75
When scanned QR is not ETH
https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f65e9599980122923f129
When QR is a link (to open in browser)
https://app.zeplin.io/project/5d088205bff2d15de6a4397b/dashboard?seid=5e6f6503de850113ba370a97
Flow (from iOS)
https://drive.google.com/a/alphawallet.com/file/d/14ge8G3pIM513myKZbDrUyjA2bvRGzFXR/view?usp=sharing
When scanning ETH Address - Add to Contacts
When scanning ETH Address - Watch Wallet
When scanning ETH Address - Send to the Address
When scanning Payment Request
When scanning to Add Custom Token
When scanned QR is not ETH
When QR is a link (to open in browser)
Which ones of these aren't working?
@tomekalphawallet could you clarify the 'add custom token' flow? Currently we access that through the + button on the wallet screen. Does this issue replace that behaviour, and if so do we need to remove that plus button? Or do you still get to the scan via clicking the plus, then clicking the scan button on the 'add custom token' activity?
It does not replace that behaviour, but goes in parallel. So if you want to add custom token you can:
a) launch a qr code scanner, scan a smart contract qr, hit "Add Token" to confirm adding prefilled form (symbol, balance, contract address, token's name)
b) you can add a custom token by: hit add/hide token in the wallet tab, hit "+", and enter details.
Makes sense?
--
Also, can I request you to update + button? I got a feedback that is not so visible. We can make it more prominent. The new one is the same size (24x24). So should be easy to replace.
https://zpl.io/2yAgxNo
This icon is used:
Change Wallet
Address Book
Add/Hide Tokens
@ChintanRathod I applied Tomek's suggestion to display an error within the QR scanner and resume scanning, rather than fold back to home and show scan error (as I mistakenly directed you to do), so no need to apply that fix now.
Addressbook integration is still being designed, so you can just add this as a hidden option that doesn't go anywhere for now.
@mpaschenko This is a very old issue, but still waiting for implementation. Please check out how this works in iOS.
Run QR Code camera (top right in the wallet tab) and scan the above QR in iOS.
iOS
@hboon The only one task left from Fiji.
Bump: @JamesSmartCell @mpaschenko
Hi @colourfreak can you clarify - in the graphic above underneath 'When scanning Payment Request', the ActionSheet looks like it pops up directly after the scan rather than going to the send screen. I like this (I have seen cases where people needed to implement something this simple at conferences, and failed terribly) this is your intention right?
Also @colourfreak when scanning a pure address, then clicking on 'send' it defaults to Ethereum Mainnet. It would be good to be able to pick which network we want to send to, although this isn't critical.
Hi @colourfreak can you clarify - in the graphic above underneath 'When scanning Payment Request', the ActionSheet looks like it pops up directly after the scan rather than going to the send screen. I like this (I have seen cases where people needed to implement something this simple at conferences, and failed terribly) this is your intention right?
@JamesSmartCell I remember that Weiwu wanted me to add this. I was not sure about the usage. However, I have seen something similar at a conference.
|
2025-04-01T06:36:41.827108
| 2020-07-26T12:39:15
|
665772810
|
{
"authors": [
"AW-STJ",
"colourful-land"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:142",
"repo": "AlphaWallet/alpha-wallet-android",
"url": "https://github.com/AlphaWallet/alpha-wallet-android/issues/1533"
}
|
gharchive/issue
|
allow nonce of 0?
Feedback from a user:
Question - I’m trying to specify a nonce of 0 to replace a pending tx of the same nonce (0 - first transfer from wallet). But Alpha Wallet is telling me the nonce must be a positive integer - it won’t let me specify 0.
I verified that this is a bug - 0 should be allowed. And 0 is a positive integer.
Can you try to reproduce this problem by:
Install a fresh AlphaWallet and generate a new address
Send some test-ether to the new address
Send it out
Speed up
See if you can capture the error in full and paste here.
Your Ethereum address starts with a nonce of 0 and increases by 1 with each transaction that's confirmed.
I think this can be a hotfix release candidate. I don't know how to fix it but shouldn't be difficult to figure out (by any one including chintan)
Looks like this issue is from a previous build. Nonce control is not available in the latest Playstore version of AlphaWalelt as the current version includes SpeedUp function.
|
2025-04-01T06:36:41.829058
| 2021-11-13T19:27:28
|
1052753299
|
{
"authors": [
"hboon",
"vladyslav-iosdev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:143",
"repo": "AlphaWallet/alpha-wallet-ios",
"url": "https://github.com/AlphaWallet/alpha-wallet-ios/pull/3428"
}
|
gharchive/pull-request
|
Add thin gray lines behind the chart with edge values #3418
Closes #3418
Force pushed after rebase master
Also see if you can add the labels on the right?
but this PR, is already displaying them, on right side of the chart
Oh, what was I thinking. Hah!. Thanks
|
2025-04-01T06:36:41.836228
| 2020-08-04T13:10:57
|
672781096
|
{
"authors": [
"Centril",
"dyc3"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:144",
"repo": "AltSysrq/proptest",
"url": "https://github.com/AltSysrq/proptest/issues/204"
}
|
gharchive/issue
|
How do I use HashMapStrategy? Documentation unclear, has no examples.
Usage of HashMapStrategy is unclear from the documentation. I need to randomly create hashmaps that meet certain conditions. From what I can see, I'm supposed to supply it with 2 other strategies to generate the keys and the values, but I'm not exactly sure how I'm supposed to make the SizeRange. Should I even be using the hash_map function?
Related documentation:
https://docs.rs/proptest/0.10.0/proptest/collection/fn.hash_map.html
https://docs.rs/proptest/0.10.0/proptest/collection/struct.HashMapStrategy.html
HashMapStrategy is literally a thin wrapper around a strategy for creating Vec<(Key, Value)> up to the specified size and filtered to not go below the minumum.
So if you e.g., use hash_map(key_strat, val_strat, 5...10), then what happens is that a vec((key_strat, val_strat), 5..10) is created, and then this is mapped using vec.into_iter().collect(), giving you a HashMap of 5..10 elements with keys drawn from key_strat and values from val_strat.
Note that most RangeXYZ types implement Into<SizeRange>, which is what hash_map accepts, so you can use the normal Rust range syntax.
I hope that answers your question.
Ah, I see. Can I use prop_filter on it? or do I just have to ignore values I can't use?
Sure, you can use .prop_filter on the key_strat, value_strat, or the final hash_map(...) strategy. Doing so on the latter will give you a &HashMap<K, V> which you can apply a predicate to. You can also ignore irrelevant maps in each test itself as well, but that will scale less well if the strategy is used a lot. If you can create a correct key/value/map by construction, then that is preferable to filtering (see the book / docs on prop_filter for elaboration).
Thanks for your help! Closing...
|
2025-04-01T06:36:41.868641
| 2023-08-21T05:39:18
|
1858632643
|
{
"authors": [
"Slach",
"alex-zaitsev",
"prashant-shahi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:145",
"repo": "Altinity/clickhouse-operator",
"url": "https://github.com/Altinity/clickhouse-operator/issues/1221"
}
|
gharchive/issue
|
sudden disappearance of CHI pods
After running the clickhouse for couple of months, the CHI pod has disappeared on its own.
@hodgesrm had advised on adding stop: "no" for fix. However, I had already resolved by changing shard count and reverting, to trigger re-creation of CHI pod (by changing state of CHI out of completed to force pod re-creation).
Nevertheless, we had been unable to find the root cause of the issue. So, this issue could re-appear in the future.
CHI definition:
apiVersion: clickhouse.altinity.com/v1
kind: ClickHouseInstallation
metadata:
name: my-release-clickhouse
namespace: default
labels:
helm.sh/chart: clickhouse-23.12.2
app.kubernetes.io/name: clickhouse
app.kubernetes.io/instance: my-release
app.kubernetes.io/component: clickhouse
app.kubernetes.io/version: "22.8.8"
app.kubernetes.io/managed-by: Helm
spec:
defaults:
templates:
dataVolumeClaimTemplate: data-volumeclaim-template
serviceTemplate: service-template
configuration:
users:
admin/password: redacted
admin/networks/ip:
- "<IP_ADDRESS>/8"
- "<IP_ADDRESS>/10"
- "<IP_ADDRESS>/12"
- "<IP_ADDRESS>/24"
- "<IP_ADDRESS>/15"
- "<IP_ADDRESS>/16"
admin/profile: default
admin/quota: default
profiles:
default/allow_experimental_window_functions: "1"
default/allow_nondeterministic_mutations: "1"
clusters:
- name: "cluster"
templates:
podTemplate: pod-template
layout:
replicasCount: 1
shardsCount: 1
settings:
format_schema_path: /etc/clickhouse-server/config.d/
prometheus/endpoint: /metrics
prometheus/port: 9363
user_defined_executable_functions_config: /etc/clickhouse-server/functions/custom-functions.xml
user_scripts_path: /var/lib/clickhouse/user_scripts/
files:
events.proto: |
syntax = "proto3";
message Event {
string uuid = 1;
string event = 2;
string properties = 3;
string timestamp = 4;
uint64 team_id = 5;
string distinct_id = 6;
string created_at = 7;
string elements_chain = 8;
}
zookeeper:
nodes:
- host: my-release-zookeeper-0.my-release-zookeeper-headless
port: 2181
templates:
podTemplates:
- name: pod-template
metadata:
labels:
helm.sh/chart: clickhouse-23.12.2
app.kubernetes.io/name: clickhouse
app.kubernetes.io/instance: my-release
app.kubernetes.io/component: clickhouse
app.kubernetes.io/version: "22.8.8"
app.kubernetes.io/managed-by: Helm
annotations: {}
podDistribution:
- topologyKey: kubernetes.io/hostname
type: ReplicaAntiAffinity
spec:
serviceAccountName: my-release-clickhouse
priorityClassName: ""
securityContext:
fsGroup: 101
runAsGroup: 101
runAsUser: 101
volumes:
- name: shared-binary-volume
emptyDir: {}
- name: custom-functions-volume
configMap:
name: my-release-clickhouse-custom-functions
initContainers: []
containers:
- name: clickhouse
image: docker.io/clickhouse/clickhouse-server:22.8.8-alpine
imagePullPolicy: IfNotPresent
command:
- /bin/bash
- -c
- /usr/bin/clickhouse-server --config-file=/etc/clickhouse-server/config.xml
ports:
- name: http
containerPort: 8123
- name: client
containerPort: 9000
- name: interserver
containerPort: 9009
volumeMounts:
- name: data-volumeclaim-template
mountPath: /var/lib/clickhouse
- name: shared-binary-volume
mountPath: /var/lib/clickhouse/user_scripts
- name: custom-functions-volume
mountPath: /etc/clickhouse-server/functions
resources:
requests:
cpu: 100m
memory: 200Mi
serviceTemplates:
- name: service-template
generateName: my-release-clickhouse
metadata:
labels:
helm.sh/chart: clickhouse-23.12.2
app.kubernetes.io/name: clickhouse
app.kubernetes.io/instance: my-release
app.kubernetes.io/component: clickhouse
app.kubernetes.io/version: "22.8.8"
app.kubernetes.io/managed-by: Helm
spec:
type: ClusterIP
ports:
- name: http
port: 8123
nodePort: null
- name: tcp
port: 9000
nodePort: null
volumeClaimTemplates:
- name: data-volumeclaim-template
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 100Gi
storageClassName: standard
Context: Slack Message
Is your kind: ClickHouseInstallation managed by the same Helm chart with clickhouse-operator?
Could you share the following commands?
kubectl get deploy --all-namespaces -l app=clickhouse-operator
kubectl get pod --all-namespaces -l clickhouse.altinity.com/chi
@Slach Yes, it is managed by the same chart with clickhouse-operator.
kubectl get deploy --all-namespaces app.kubernetes.io/component=operator
NAMESPACE NAME READY UP-TO-DATE AVAILABLE AGE
default my-release-clickhouse-operator 1/1 1 1 60d
test my-release-clickhouse-operator 1/1 1 1 38d
kubectl get pod --all-namespaces -l clickhouse.altinity.com/chi
NAMESPACE NAME READY STATUS RESTARTS AGE
default chi-my-release-clickhouse-cluster-0-0-0 1/1 Running 0 3d21h
test chi-my-release-clickhouse-cluster-0-0-0 1/1 Running 0 4d19h
could you share operator logs from test namespace?
kubectl logs -n test -l app.kubernetes.io/component=operator --since=7d
I was getting the error below:
error: invalid argument "7d" for "--since" flag: time: unknown unit "d" in duration "7d"
Hence, replaced 7d with 168h:
$ kubectl logs -n test -l app.kubernetes.io/component=operator --since=168h
Defaulted container "my-release-clickhouse-operator" out of: my-release-clickhouse-operator, my-release-clickhouse-metrics-exporter
E0822 10:51:33.116468 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/test/services?resourceVersion=101836934": dial tcp <IP_ADDRESS>:443: i/o timeout
I0822 10:51:38.125884 1 trace.go:205] Trace[301511361]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (22-Aug-2023 10:51:08.124) (total time: 30001ms):
Trace[301511361]: [30.00114905s] [30.00114905s] END
E0822 10:51:38.125902 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: Get "https://<IP_ADDRESS>:443/apis/apps/v1/namespaces/test/statefulsets?resourceVersion=101836529": dial tcp <IP_ADDRESS>:443: i/o timeout
I0822 10:51:38.355293 1 trace.go:205] Trace[2060365486]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (22-Aug-2023 10:51:08.354) (total time: 30001ms):
Trace[2060365486]: [30.001212232s] [30.001212232s] END
E0822 10:51:38.355312 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/test/configmaps?resourceVersion=101837009": dial tcp <IP_ADDRESS>:443: i/o timeout
I0822 10:51:42.173930 1 trace.go:205] Trace[250496038]: "Reflector ListAndWatch" name:pkg/client/informers/externalversions/factory.go:117 (22-Aug-2023 10:51:12.172) (total time: 30001ms):
Trace[250496038]: [30.001189387s] [30.001189387s] END
E0822 10:51:42.173949 1 reflector.go:138] pkg/client/informers/externalversions/factory.go:117: Failed to watch *v1.ClickHouseOperatorConfiguration: failed to list *v1.ClickHouseOperatorConfiguration: Get "https://<IP_ADDRESS>:443/apis/clickhouse.altinity.com/v1/namespaces/test/clickhouseoperatorconfigurations?resourceVersion=101836611": dial tcp <IP_ADDRESS>:443: i/o timeout
$ kubectl logs -n default -l app.kubernetes.io/component=operator --since=168h
Defaulted container "my-release-clickhouse-operator" out of: my-release-clickhouse-operator, my-release-clickhouse-metrics-exporter
E0822 10:51:38.591455 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/default/services?resourceVersion=101836934": dial tcp <IP_ADDRESS>:443: i/o timeout
I0822 10:51:41.314893 1 trace.go:205] Trace[1051692860]: "Reflector ListAndWatch" name:pkg/client/informers/externalversions/factory.go:117 (22-Aug-2023 10:51:11.313) (total time: 30001ms):
Trace[1051692860]: [30.001164853s] [30.001164853s] END
E0822 10:51:41.314912 1 reflector.go:138] pkg/client/informers/externalversions/factory.go:117: Failed to watch *v1.ClickHouseOperatorConfiguration: failed to list *v1.ClickHouseOperatorConfiguration: Get "https://<IP_ADDRESS>:443/apis/clickhouse.altinity.com/v1/namespaces/default/clickhouseoperatorconfigurations?resourceVersion=101836881": dial tcp <IP_ADDRESS>:443: i/o timeout
I0822 10:51:41.768145 1 trace.go:205] Trace[523315261]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (22-Aug-2023 10:51:11.767) (total time: 30000ms):
Trace[523315261]: [30.000906333s] [30.000906333s] END
E0822 10:51:41.768163 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/default/configmaps?resourceVersion=101837009": dial tcp <IP_ADDRESS>:443: i/o timeout
I0822 10:51:42.946368 1 trace.go:205] Trace[838394852]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (22-Aug-2023 10:51:12.945) (total time: 30000ms):
Trace[838394852]: [30.000528326s] [30.000528326s] END
E0822 10:51:42.946386 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Pod: failed to list *v1.Pod: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/default/pods?resourceVersion=101836916": dial tcp <IP_ADDRESS>:443: i/o timeout
looks like nothing happens 22 Aug
Could you share?
kubectl logs -n default -l app.kubernetes.io/component=operator -c clickhouse-operator
kubectl logs -n test -l app.kubernetes.io/component=operator -c clickhouse-operator
Sure, here are the logs:
$ kubectl logs -n default -l app.kubernetes.io/component=operator -c my-release-clickhouse-operator
I0823 18:03:06.328611 1 worker-reconciler.go:111] worker-reconciler.go:52:reconcileCHI():end:default/my-release-clickhouse
I0823 18:03:06.328632 1 worker.go:414] worker.go:380:updateCHI():end:default/my-release-clickhouse
E0823 18:03:49.676857 1 connection.go:190] Exec():FAILED Exec(http://clickhouse_operator:***@chi-my-release-clickhouse-cluster-0-0.default.svc.cluster.local:8123/) context deadline exceeded for SQL: SYSTEM DROP DNS CACHE
W0823 18:03:49.712800 1 retry.go:52] exec():chi-my-release-clickhouse-cluster-0-0.default.svc.cluster.local:FAILED single try. No retries will be made for Applying sqls
I0823 18:03:49.801384 1 worker.go:339] default/my-release-clickhouse/5d89fdfb-a136-430a-a8a5-310c07edef5b:IPs of the CHI-1 [<IP_ADDRESS>]
I0823 18:03:49.812084 1 worker.go:343] default/my-release-clickhouse/13b958a0-3cfd-46a6-a3ce-368a46cc97df:Update users IPS-1
I0823 18:03:49.821067 1 worker.go:1089] updateConfigMap():default/my-release-clickhouse/13b958a0-3cfd-46a6-a3ce-368a46cc97df:Update ConfigMap default/chi-my-release-clickhouse-common-usersd
I0823 18:03:49.903031 1 worker.go:339] default/my-release-clickhouse/88d992b4-a76a-4a2b-af19-2cc81d64548a:IPs of the CHI-1 [<IP_ADDRESS>]
I0823 18:03:49.910434 1 worker.go:343] default/my-release-clickhouse/5892dd12-16a8-4e6f-baf6-e5ec8776277d:Update users IPS-1
I0823 18:03:49.914824 1 worker.go:1089] updateConfigMap():default/my-release-clickhouse/5892dd12-16a8-4e6f-baf6-e5ec8776277d:Update ConfigMap default/chi-my-release-clickhouse-common-usersd
$ kubectl logs -n test -l app.kubernetes.io/component=operator -c my-release-clickhouse-operator
I0823 17:13:58.953558 1 worker-deleter.go:64] worker-deleter.go:64:dropReplicas():start:test/my-release-clickhouse/f3f569fa-3d23-4b03-a914-71c7bf07cf7d:drop replicas based on AP
I0823 17:13:58.953588 1 worker-deleter.go:81] worker-deleter.go:81:dropReplicas():end:test/my-release-clickhouse/f3f569fa-3d23-4b03-a914-71c7bf07cf7d:processed replicas: 0
I0823 17:13:58.953612 1 worker.go:573] includeStopped():test/my-release-clickhouse/f3f569fa-3d23-4b03-a914-71c7bf07cf7d:add CHI to monitoring
I0823 17:13:59.355282 1 controller.go:609] OK update watch (test/my-release-clickhouse): {"namespace":"test","name":"my-release-clickhouse","clusters":[{"name":"cluster","hosts":[{"name":"0-0","hostname":"chi-my-release-clickhouse-cluster-0-0.test.svc.cluster.local","tcpPort":9000,"httpPort":8123}]}]}
I0823 17:13:59.360534 1 worker.go:540] test/my-release-clickhouse:all IP addresses are in place
I0823 17:13:59.526393 1 worker.go:611] test/my-release-clickhouse/7de65ad0-12b1-4823-8657-ca1e4984bb8f:IPs of the CHI-2 [<IP_ADDRESS>]
I0823 17:13:59.534938 1 worker.go:615] test/my-release-clickhouse/789d4f97-3cee-499d-9e78-631399e02bea:Update users IPS-2
I0823 17:13:59.748918 1 worker.go:636] finalizeReconcileAndMarkCompleted():test/my-release-clickhouse/f3f569fa-3d23-4b03-a914-71c7bf07cf7d:reconcile completed successfully, task id: f3f569fa-3d23-4b03-a914-71c7bf07cf7d
I0823 17:14:00.146065 1 worker-reconciler.go:111] worker-reconciler.go:52:reconcileCHI():end:test/my-release-clickhouse
I0823 17:14:00.146085 1 worker.go:414] worker.go:380:updateCHI():end:test/my-release-clickhouse
=( this is not a full logs, this is only logs for 23 Aug
@Slach This is all that there is. :(
Closing it as it is not reproducible. Could be something related to Helm
@alex-zaitsev This is happening more often in recent time when trying to increase the ClickHouse PVC size in ClickHouse installation. Not able to pin point the root cause of the issue.
Operator logs:
I0729 03:49:26.917146 1 worker.go:1089] updateConfigMap():my-release/my-release-clickhouse/58053693-2b8a-4d1d-b5ed-e0432448d655:Update ConfigMap my-release/chi-my-release-clickhouse-deploy-confd-cluster-0-0
I0729 03:49:27.928385 1 cluster.go:84] Run query on: chi-my-release-clickhouse-cluster-0-0.my-release.svc.cluster.local of [chi-my-release-clickhouse-cluster-0-0.my-release.svc.cluster.local]
I0729 03:49:28.033592 1 worker-reconciler.go:292] reconcileHostStatefulSet():Reconcile host 0-0. ClickHouse version: <IP_ADDRESS>
I0729 03:49:28.033769 1 worker.go:134] shouldForceRestartHost():Force restart is not required. Host: 0-0
I0729 03:49:28.033802 1 worker-reconciler.go:304] reconcileHostStatefulSet():Reconcile host 0-0. Reconcile StatefulSet
I0729 03:49:28.033959 1 creator.go:589] getPodTemplate():my-release/my-release-clickhouse/58053693-2b8a-4d1d-b5ed-e0432448d655:statefulSet chi-my-release-clickhouse-cluster-0-0 use custom template: pod-template
W0729 03:49:28.034289 1 creator.go:994] containerAppendVolumeMount():my-release/my-release-clickhouse/58053693-2b8a-4d1d-b5ed-e0432448d655:container.Name:clickhouse volumeMount.Name:data-volumeclaim-template already used
I0729 03:49:28.036205 1 worker.go:1290] getStatefulSetStatus():my-release/chi-my-release-clickhouse-cluster-0-0:cur and new StatefulSets ARE DIFFERENT based on labels. StatefulSet reconcile is required for: my-release/chi-my-release-clickhouse-cluster-0-0
I0729 03:49:28.036374 1 worker.go:1431] updateStatefulSet():Update StatefulSet(my-release/chi-my-release-clickhouse-cluster-0-0) - started
I0729 03:49:28.348502 1 worker.go:1401] waitConfigMapPropagation():Wait for ConfigMap propagation for 9.003953254s 996.046746ms/10s
E0729 03:49:37.360361 1 creator.go:76] updateStatefulSet():StatefulSet update failed. err: StatefulSet.apps "chi-my-release-clickhouse-cluster-0-0" is invalid: spec: Forbidden: updates to statefulset spec for fields other than 'replicas', 'ordinals', 'template', 'updateStrategy', 'persistentVolumeClaimRetentionPolicy' and 'minReadySeconds' are forbidden
E0729 03:49:37.361923 1 creator.go:100] updateStatefulSet():NOT EQUAL: AP item start -------------------------
modified spec items: 24
ap item path [0]:'.Template.Spec.Volumes[1].VolumeSource.ConfigMap.DefaultMode'
ap item value[0]:'nil'
ap item path [1]:'.Template.Spec.Containers[0].LivenessProbe.SuccessThreshold'
ap item value[1]:'0'
ap item path [2]:'.Template.Spec.Containers[0].ReadinessProbe.Handler.HTTPGet.Scheme'
ap item value[2]:'""'
ap item path [3]:'.Template.Spec.SchedulerName'
ap item value[3]:'""'
ap item path [4]:'.Template.Spec.Containers[0].LivenessProbe.Handler.HTTPGet.Scheme'
--
I0729 03:49:42.476651 1 poller.go:213] pollStatefulSet():my-release/chi-my-release-clickhouse-cluster-0-0:OK :ObservedGeneration:2 Replicas:1 ReadyReplicas:1 CurrentReplicas:0 UpdatedReplicas:0 CurrentRevision:chi-my-release-clickhouse-cluster-0-0-8f9b7b45b UpdateRevision:chi-my-release-clickhouse-cluster-0-0-8f9b7b45b
W0729 03:50:27.487009 1 reflector.go:436] pkg/client/informers/externalversions/factory.go:117: watch of *v1.ClickHouseInstallation ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487049 1 reflector.go:436] k8s.io/client-go/informers/factory.go:134: watch of *v1.StatefulSet ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487052 1 reflector.go:436] k8s.io/client-go/informers/factory.go:134: watch of *v1.Endpoints ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487012 1 reflector.go:436] k8s.io/client-go/informers/factory.go:134: watch of *v1.Service ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487096 1 reflector.go:436] k8s.io/client-go/informers/factory.go:134: watch of *v1.ConfigMap ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487127 1 reflector.go:436] pkg/client/informers/externalversions/factory.go:117: watch of *v1.ClickHouseInstallationTemplate ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487145 1 reflector.go:436] k8s.io/client-go/informers/factory.go:134: watch of *v1.Pod ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
W0729 03:50:27.487294 1 reflector.go:436] pkg/client/informers/externalversions/factory.go:117: watch of *v1.ClickHouseOperatorConfiguration ended with: an error on the server ("unable to decode an event from the watch stream: http2: client connection lost") has prevented the request from succeeding
I0729 03:50:27.488331 1 labeler.go:292] deleteLabelReadyPod():FAIL get pod for host 'my-release/0-0' err: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/my-release/pods/chi-my-release-clickhouse-cluster-0-0-0": http2: client connection lost
E0729 03:50:57.489774 1 poller.go:237] pollStatefulSet():my-release/chi-my-release-clickhouse-cluster-0-0:my-release/chi-my-release-clickhouse-cluster-0-0 Get() FAILED
I0729 03:50:58.323884 1 trace.go:205] Trace[1632088119]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (29-Jul-2024 03:50:28.323) (total time: 30000ms):
Trace[1632088119]: [30.000791007s] [30.000791007s] END
E0729 03:50:58.323928 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Pod: failed to list *v1.Pod: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/my-release/pods?resourceVersion=409134742": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:58.387047 1 trace.go:205] Trace[1562350316]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (29-Jul-2024 03:50:28.386) (total time: 30000ms):
Trace[1562350316]: [30.000679957s] [30.000679957s] END
E0729 03:50:58.387074 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: Get "https://<IP_ADDRESS>:443/apis/apps/v1/namespaces/my-release/statefulsets?resourceVersion=409134694": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:58.482514 1 trace.go:205] Trace[1523333257]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (29-Jul-2024 03:50:28.481) (total time: 30000ms):
Trace[1523333257]: [30.000761918s] [30.000761918s] END
E0729 03:50:58.482587 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/my-release/configmaps?resourceVersion=409134621": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:58.627969 1 trace.go:205] Trace[328066567]: "Reflector ListAndWatch" name:pkg/client/informers/externalversions/factory.go:117 (29-Jul-2024 03:50:28.626) (total time: 30001ms):
Trace[328066567]: [30.001690277s] [30.001690277s] END
E0729 03:50:58.627999 1 reflector.go:138] pkg/client/informers/externalversions/factory.go:117: Failed to watch *v1.ClickHouseOperatorConfiguration: failed to list *v1.ClickHouseOperatorConfiguration: Get "https://<IP_ADDRESS>:443/apis/clickhouse.altinity.com/v1/namespaces/my-release/clickhouseoperatorconfigurations?resourceVersion=409134300": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:58.680316 1 trace.go:205] Trace[1867447287]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (29-Jul-2024 03:50:28.679) (total time: 30000ms):
Trace[1867447287]: [30.000683297s] [30.000683297s] END
E0729 03:50:58.680433 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/my-release/services?resourceVersion=409134696": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:58.870657 1 trace.go:205] Trace[321750039]: "Reflector ListAndWatch" name:pkg/client/informers/externalversions/factory.go:117 (29-Jul-2024 03:50:28.869) (total time: 30000ms):
Trace[321750039]: [30.000749817s] [30.000749817s] END
E0729 03:50:58.870684 1 reflector.go:138] pkg/client/informers/externalversions/factory.go:117: Failed to watch *v1.ClickHouseInstallation: failed to list *v1.ClickHouseInstallation: Get "https://<IP_ADDRESS>:443/apis/clickhouse.altinity.com/v1/namespaces/my-release/clickhouseinstallations?resourceVersion=409134690": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:59.038154 1 trace.go:205] Trace[345492881]: "Reflector ListAndWatch" name:pkg/client/informers/externalversions/factory.go:117 (29-Jul-2024 03:50:29.037) (total time: 30000ms):
Trace[345492881]: [30.000824067s] [30.000824067s] END
E0729 03:50:59.038264 1 reflector.go:138] pkg/client/informers/externalversions/factory.go:117: Failed to watch *v1.ClickHouseInstallationTemplate: failed to list *v1.ClickHouseInstallationTemplate: Get "https://<IP_ADDRESS>:443/apis/clickhouse.altinity.com/v1/namespaces/my-release/clickhouseinstallationtemplates?resourceVersion=409134495": dial tcp <IP_ADDRESS>:443: i/o timeout
I0729 03:50:59.057219 1 trace.go:205] Trace[1113844881]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:134 (29-Jul-2024 03:50:29.056) (total time: 30000ms):
Trace[1113844881]: [30.000738938s] [30.000738938s] END
E0729 03:50:59.057261 1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Endpoints: failed to list *v1.Endpoints: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/my-release/endpoints?resourceVersion=409134706": dial tcp <IP_ADDRESS>:443: i/o timeout
@alex-zaitsev @Slach can you please help here?
Failed to watch *v1.Endpoints: failed to list *v1.Endpoints: Get "https://<IP_ADDRESS>:443/api/v1/namespaces/my-release/endpoints?resourceVersion=409134706": dial tcp <IP_ADDRESS>:443: i/o timeout
means something wrong with your kubernetes API server
|
2025-04-01T06:36:41.880676
| 2023-04-18T13:42:27
|
1673164638
|
{
"authors": [
"borgethommesen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:146",
"repo": "Altinn/altinn-studio-docs",
"url": "https://github.com/Altinn/altinn-studio-docs/pull/896"
}
|
gharchive/pull-request
|
Update _index.en.md
Description
Minor corrections to text.
Your vs You.
Functionality misspelled.
Minor corrections to text.
|
2025-04-01T06:36:41.890260
| 2024-08-29T11:17:08
|
2494149417
|
{
"authors": [
"danielskovli"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:147",
"repo": "Altinn/app-lib-dotnet",
"url": "https://github.com/Altinn/app-lib-dotnet/pull/738"
}
|
gharchive/pull-request
|
chore: Move Telemetry class files to a separate /Telemetry folder
Description
The Telemetry partial class consists of 27 files, which were all located at root level in Core/Features/.
This change simply moves them to a /Telemetry subfolder, purely for organisational purposes. The namespace remains unchanged.
Related Issue(s)
N/A
Verification
[x] Your code builds clean without any errors or warnings
[x] Manual testing done (required)
[ ] Relevant automated test added (if you find this hard, leave it and we'll help out)
[x] All tests run green
Documentation
[ ] User documentation is updated with a separate linked PR in altinn-studio-docs. (if applicable)
Quality gate failed
😂
|
2025-04-01T06:36:41.900600
| 2017-08-09T21:50:06
|
249172551
|
{
"authors": [
"Arawn-Davies",
"valentinbreiz"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:148",
"repo": "Alve-OS/Alve-Operating-System",
"url": "https://github.com/Alve-OS/Alve-Operating-System/pull/15"
}
|
gharchive/pull-request
|
Small mistype (incorrect link) in README
Hi, I noticed the link to (this repository) was a hyperlink to another Github repo, which didn't contain the Alve OS source, so I corrected the link that is to be used :)
Thank you :P
|
2025-04-01T06:36:41.908712
| 2020-07-12T20:39:30
|
655472428
|
{
"authors": [
"helje5",
"meech-ward"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:149",
"repo": "AlwaysRightInstitute/swift-mac2arm-x-compile-toolchain",
"url": "https://github.com/AlwaysRightInstitute/swift-mac2arm-x-compile-toolchain/issues/5"
}
|
gharchive/issue
|
unknown argument: '-color-diagnostics'
When I run swift build --destination /tmp/cross-toolchain/arm64v8-ubuntu-bionic-destination.json
I get the following error message:
<unknown>:0: error: unknown argument: '-color-diagnostics'
Does anyone know how to fix this issue?
If I just run swift build then everything works correctly, so it's only an issue when trying to use the cross-toolchain.
I've tried this on two macs, one using Xcode 12 and the other using Xcode 11.
Try out this fork, it should be up2date: https://github.com/CSCIX65G/SwiftCrossCompilers
Thanks for the suggestion @helje5, but I get the same exact issue with SwiftCrossCompilers.
I'd still file it over there, the fork is AFAIK actively maintained. I guess the color diag option was probably new in some Swift 5.x, maybe they b0rked something else. Invoking it w/ -v is usually the way to start investigating the issue and see what breaks when and where.
Once there is more info, it might be also worth filing a bug against SPM at bugs.swift.org.
|
2025-04-01T06:36:41.912851
| 2023-04-13T08:03:20
|
1665924297
|
{
"authors": [
"FredrikNoren",
"droqen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:150",
"repo": "AmbientRun/Ambient",
"url": "https://github.com/AmbientRun/Ambient/issues/299"
}
|
gharchive/issue
|
Setting color and transparency_group for all entities in model
We currently have outline_recursive, which will apply the outline to all child entities. We could do something similar (color_recursive and transparency_group_recursive). Or we could try to figure out some more general solution to this.
@philpax Well the user would then have to somehow wait until the model is loaded and then run those.
Made a discussion about two-entity queries which could be the form of a possible 'more general solution': https://github.com/AmbientRun/Ambient/discussions/600
|
2025-04-01T06:36:41.920466
| 2024-07-12T18:17:26
|
2406119794
|
{
"authors": [
"AmirVahedix"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:151",
"repo": "AmirVahedix/weight-conversion",
"url": "https://github.com/AmirVahedix/weight-conversion/issues/4"
}
|
gharchive/issue
| |
2025-04-01T06:36:41.940930
| 2015-02-27T03:11:58
|
59183532
|
{
"authors": [
"ahdinosaur",
"cdaringe",
"doubleface",
"kahnjw",
"mst7555",
"pgilad"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:152",
"repo": "AmpersandJS/ampersand-view",
"url": "https://github.com/AmpersandJS/ampersand-view/issues/105"
}
|
gharchive/issue
|
unexpected that remove also modifies bindings
heya, :rabbit2:
i'm trying to render, remove, and re-render a view, however on the re-render the bindings don't work as expected. i realize now this is due to remove modifying the bindings, but i wonder if there is a better way to do this.
if we want to unlisten and delete bindings in remove, is it possible to provide a function that re-listens and re-creates the bindings? or another option is we move that unlisten and delete bindings to another function that maybe can be run by default in remove but there is an option to not run it.
cheers, :tea:
This is the updated ref: https://github.com/AmpersandJS/ampersand-view/blob/master/ampersand-view.js#L152-L168
I don't see any reason to leave the bindings if the element has been removed, and you stop listening to events.
If you have a problem with rendering a view that was removed, I'd love to see the use-case and a simple implementation (test-case) of it failing... If the usage is correct there must be some side effect we haven't seen
It looks like a problem I just found with view switcher ;-/
Switching a view calls view remove what also removes all related event and model bindings. I agree with @pgilad that this behaviour make sense so it should be performed.
The problem is that switching back to such previously removed view do not restore bindings - it just calls view render.
How can I restore all bindings (automatically) in such case?
This is likely a problem with ampersand-view-switcher using the view.rendered property to determine whether or not it needs to call view.render() on the new view.
The view.rendered property will be true if view.el is defined. view.el will be defined even after view.remove() is called. This means that when we are "re-rendering" a view with View Switcher, the view's element will be placed in the DOM, but view.render() will not be called again, so I think the events are torn down, and never set back up again.
Pretty sure this should be moved to ampersand-view-switcher.
Issue: https://github.com/AmpersandJS/ampersand-view-switcher/issues/25
@ahdinosaur @pgilad,
i agree, this causes issues. in ampersand-form-manager-view i cycle between forms, meaning I render, remove, and re-render them. i reinitialize bindings and subviews on each render manually. the fix may be as easy as pulling out the binding and subview initialization to a helper. then, on render, test if they are set. if not, re-init them.
I have the same problem as @ahdinosaur. I want be able to rerender a view which has been removed.
Here is my
I have the same problem as @ahdinosaur. I want be able to reuse a view which has been removed. I use the view switcher but I could have the same problem with any other way of switching views which calls the remove method of the view.
Here is an example
import View from "ampersand-view";
import ViewSwitcher from "ampersand-view-switcher";
var switcher = new ViewSwitcher(document.querySelector("main"));
var V1 = View.extend({
session: {
text: {
type: "string",
default: "bonjour"
}
},
bindings: {
text: ""
},
template: `<div>default text v1</div>`
});
var V2 = View.extend({
template: `<div>default text v2</div>`
});
window.views = [new V1(), new V2()];
var n = 1;
document.addEventListener("click", () => {
n++;
switcher.set(views[n % 2]);
});
In this example, the bindings do not work after one view switch.
A workaround for this could be to create a new instance of the view each time but I don't find this satisfying.
And there is already a good solution for the dom events : just call delegateEvents.
A simple createBindings function would also be enough for this need.
Don't you think?
|
2025-04-01T06:36:41.949092
| 2023-03-13T01:53:46
|
1620616190
|
{
"authors": [
"kevinpagtakhan",
"markfoo",
"markzegarelli"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:153",
"repo": "Amplitude-Developer-Docs/amplitude-dev-center",
"url": "https://github.com/Amplitude-Developer-Docs/amplitude-dev-center/pull/651"
}
|
gharchive/pull-request
|
Re-word manual session_id sentence
Reword last session under 'Sessions' to reflect usage of HTTP APIs for seeing session IDs manually
Amplitude Developer Docs PR
Description
Reword last session under 'Sessions' to reflect usage of HTTP APIs for seeing session IDs manually
Deadline
When do these changes need to be live on the site?
Change type
[ ] Doc bug fix. Fixes #[insert issue number]. Amplitude contributors include Jira issue number.
[X] Doc update.
[ ] New documentation.
[ ] Non-documentation related fix or update.
PR checklist:
[ ] My documentation follows the style guidelines of this project.
[ ] I previewed my documentation on a local server using mkdocs serve.
[ ] Running mkdocs serve didn't generate any failures.
[X] I have performed a self-review of my own documentation.
@amplitude-dev-docs
@markfoo do we still need this PR?
Closing this in preparation of migrating the repository.
|
2025-04-01T06:36:41.952644
| 2024-02-28T11:50:31
|
2158781517
|
{
"authors": [
"BeritJanssen",
"drikusroor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:154",
"repo": "Amsterdam-Music-Lab/MUSCLE",
"url": "https://github.com/Amsterdam-Music-Lab/MUSCLE/pull/810"
}
|
gharchive/pull-request
|
Add theme support to Experiment model and components
This pull request adds theme support per Experiment. It gives the admin user the ability to create and select a theme for an experiment, in which it can configure a logo image, background image, heading font, and body font).
Right now, this means the user can configure images and fonts by inputting image and google font urls. Later, we can expand this functionality by provoding upload features.
Resolves #805
Screenshots preview
Theme configuration add/edit
Themes admin overview
Select theme in Experiment
Should / could we also set the favicon through the admin interface?
Should / could we also set the favicon through the admin interface?
That sounds like a good idea. I prefer not to have scope creep in this PR so I've created issue #819 for it.
I like it! The only thing that's not clear to me: what would happen if a user set example.com/not-a-font.jpg in the font form fields? Silent fail?
It's possible to enter an incorrect font or image url that will silently fail in the frontend. However, thanks to the preview in the Django interface that shouldn't happen as you can see when it fails.
|
2025-04-01T06:36:41.969528
| 2021-06-04T06:33:23
|
911169931
|
{
"authors": [
"AnWeber",
"dragondove"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:155",
"repo": "AnWeber/vscode-httpyac",
"url": "https://github.com/AnWeber/vscode-httpyac/issues/40"
}
|
gharchive/issue
|
ReferenceError: sth is not defined
I've wrote sth like this:
### regenerate invoice
@invoiceId =<PHONE_NUMBER>2001061854
PUT {{host}}/v1/electronicInvoice/regenerate?invoiceId={{invoiceId}}
the 'host' variable is defined in a env file, and I defined invoiceId in current region. When I attempt to send a http request, I got errors. The trace followings:
ReferenceError: invoiceId is not defined
at Object.userJS (d:\apis\gtc\einvoice.http:12:92)
at Object.<anonymous> (c:\Users\jingc\.vscode-oss\extensions\anweber.vscode-httpyac-2.12.4\dist\extension.js:292:37177)
at Generator.next (<anonymous>)
at c:\Users\jingc\.vscode-oss\extensions\anweber.vscode-httpyac-2.12.4\dist\extension.js:292:36182
at new Promise (<anonymous>)
at s (c:\Users\jingc\.vscode-oss\extensions\anweber.vscode-httpyac-2.12.4\dist\extension.js:292:35927)
at Object.g (c:\Users\jingc\.vscode-oss\extensions\anweber.vscode-httpyac-2.12.4\dist\extension.js:292:36461)
at t.JavascriptVariableReplacer.<anonymous> (c:\Users\jingc\.vscode-oss\extensions\anweber.vscode-httpyac-2.12.4\dist\extension.js:292:140016)
at Generator.next (<anonymous>)
at s (c:\Users\jingc\.vscode-oss\extensions\anweber.vscode-httpyac-2.12.4\dist\extension.js:292:139204)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:93:5)
I could not reproduce the error right now. I guess the error also disappeared after reopening the file on your pc. Can you remember the order of the input. e.g. First RequestLine then Variable or the other way around?
reopen file or restart vscodium doesn't solve the problem. I tried rewrite code with input variable first, and failed again.
Then I try with a new http file, it worked.
Now I know how to reproduce the error:
### first req
GET http://sth/sth
### second
@sth = abc
GET http://sth/{{sth}}
I get the error now too. Interesting. I will have a look at it tonight.
Cause is the text in the delimiter `### second'. then the delimiter no longer works correctly
The following works
###
# @name second
The change can be tested with release 2.12.5.
It's still broken. Even when I remove all text after delimiter
###
@vara = 3
GET {{host}}/{{vara}} HTTP/1.1
###
@varb = 2
GET {{host}}/{{varb}} HTTP/1.1
the first variable vara works fine while varb does not.
Unfortunately, I won't get around to it today. I suspect it is the missing blank line between the first request and the ###. But that is just a guess.
error is as expected. The missing blank line between first request GET {{host}}/{{vara}} and the ### is causing this. The outline view of vscode give a hint to the error. @varb = 2 is added to request with vara.
Maybe I can optimize away the currently needed blank line. I will have a look at it
I have removed the required blank line. Now it is also possible to create the requests identically to Kibana. Nice:-)
It works :)
Again, thanks for your work and patience on this project, hoping this project can be discovered by more people.
Thx for using my extension:-)
|
2025-04-01T06:36:41.976788
| 2024-11-05T14:55:21
|
2635696506
|
{
"authors": [
"anaconda-pkg-build",
"anaobi"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:156",
"repo": "AnacondaRecipes/haystack-experimental-feedstock",
"url": "https://github.com/AnacondaRecipes/haystack-experimental-feedstock/pull/1"
}
|
gharchive/pull-request
|
first build
{package} {version} {:snowflake:}
Destination channel: {Snowflake | defaults}
Links
{ticket_number}
Upstream repository
Upstream changelog/diff
Relevant dependency PRs:
...
Explanation of changes:
...
Linter check found the following problems:
ERROR conda.cli.main_run:execute(125): `conda run conda-lint /tmp/abs_609kgkq8iw/clone` failed. (See above for error)
The following problems have been found:
===== WARNINGS =====
clone/recipe/meta.yaml:15: avoid_noarch: noarch: python packages should be avoided
clone/recipe/meta.yaml:39: missing_description: The recipe is missing a description
===== ERRORS =====
clone/recipe/meta.yaml:39: missing_dev_url: The recipe is missing a dev_url
clone/recipe/meta.yaml:39: missing_license_family: The recipe is missing the about/license_family key.
clone/recipe/meta.yaml:39: missing_documentation: The recipe is missing a doc_url or doc_source_url
===== Final Report: =====
3 Errors and 2 Warnings were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(125): `conda run conda-lint /tmp/abs_01z0cvojnf/clone` failed. (See above for error)
The following problems have been found:
===== WARNINGS =====
clone/recipe/meta.yaml:15: avoid_noarch: noarch: python packages should be avoided
clone/recipe/meta.yaml:41: missing_description: The recipe is missing a description
===== ERRORS =====
clone/recipe/meta.yaml:41: missing_license_family: The recipe is missing the about/license_family key.
===== Final Report: =====
1 Error and 2 Warnings were found
|
2025-04-01T06:36:42.008201
| 2024-04-08T22:50:02
|
2232254086
|
{
"authors": [
"anaconda-pkg-build",
"boldorider4",
"skupr-anaconda"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:157",
"repo": "AnacondaRecipes/polars-feedstock",
"url": "https://github.com/AnacondaRecipes/polars-feedstock/pull/3"
}
|
gharchive/pull-request
|
build 0.20.18 :snowflake:
polars 0.20.18 :snowflake:
Destination channel: Snowflake
Links
PKG-4133
Upstream repository
Upstream changelog/diff
Relevant dependency PRs:
rust-activation-feedstock
Explanation of changes:
used nightly version of rust to use experimental features/
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_aexmwhkhpb/clone` failed. (See above for error)
The following problems have been found:
===== WARNINGS =====
clone/recipe/meta.yaml:57: missing_description: The recipe is missing a description
===== ERRORS =====
clone/recipe/meta.yaml:15: patch_unnecessary: patch should not be in build when source/patches is not set
clone/recipe/meta.yaml:31: missing_wheel: For pypi packages, wheel should be present in the host section
clone/recipe/build.sh:19: pip_install_args: pip install should be run with --no-deps and --no-build-isolation.
===== Final Report: =====
3 Errors and 1 Warning were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_fe6pjnvth2/clone` failed. (See above for error)
The following problems have been found:
===== WARNINGS =====
clone/recipe/meta.yaml:55: missing_description: The recipe is missing a description
===== ERRORS =====
clone/recipe/meta.yaml:29: missing_wheel: For pypi packages, wheel should be present in the host section
clone/recipe/build.sh:19: pip_install_args: pip install should be run with --no-deps and --no-build-isolation.
===== Final Report: =====
2 Errors and 1 Warning were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_b3t1kudisi/clone` failed. (See above for error)
The following problems have been found:
===== ERRORS =====
clone/recipe/meta.yaml:30: missing_wheel: For pypi packages, wheel should be present in the host section
===== Final Report: =====
1 Error and 0 Warnings were found
Linter check found the following problems:
Traceback (most recent call last):
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/lint/__init__.py", line 834, in lint_file
recipe = _recipe.Recipe.from_file(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 243, in from_file
raise exc from exc
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 234, in from_file
recipe._load_from_string(text.read())
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 170, in _load_from_string
self.render()
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 290, in render
self.meta = renderer_utils.render(self.recipe_dir, self.dump(), self.selector_dict, self.renderer)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/_renderer.py", line 234, in render
raise JinjaRenderFailure(recipe_dir, message=exc.message, line=exc.lineno) from exc
percy.render.exceptions.JinjaRenderFailure: (PosixPath('/tmp/abs_4elg8j0fvo/clone/recipe'), "expected token 'end of statement block', got '+' (at line 42)")
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/run.py", line 159, in main
sys.exit(prime())
^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/run.py", line 138, in prime
result = linter.lint(recipes, subdir, args.variant_config_files, args.exclusive_config_files, args.fix)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/lint/init.py", line 735, in lint
msgs = self.lint_file(
^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/lint/init.py", line 851, in lint_file
recipe = _recipe.Recipe(recipe_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 107, in init
self.recipe_dir = recipe_file.parent
^^^^^^^^^^^^^^^^^^
AttributeError: 'str' object has no attribute 'parent'
ERROR conda.cli.main_run:execute(124): conda run conda-lint /tmp/abs_4elg8j0fvo/clone failed. (See above for error)
Linter check found the following problems:
Traceback (most recent call last):
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/lint/__init__.py", line 834, in lint_file
recipe = _recipe.Recipe.from_file(
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 243, in from_file
raise exc from exc
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 234, in from_file
recipe._load_from_string(text.read())
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 170, in _load_from_string
self.render()
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 290, in render
self.meta = renderer_utils.render(self.recipe_dir, self.dump(), self.selector_dict, self.renderer)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/_renderer.py", line 234, in render
raise JinjaRenderFailure(recipe_dir, message=exc.message, line=exc.lineno) from exc
percy.render.exceptions.JinjaRenderFailure: (PosixPath('/tmp/abs_34bssbgmb8/clone/recipe'), "expected token 'end of statement block', got '+' (at line 42)")
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/run.py", line 159, in main
sys.exit(prime())
^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/run.py", line 138, in prime
result = linter.lint(recipes, subdir, args.variant_config_files, args.exclusive_config_files, args.fix)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/lint/init.py", line 735, in lint
msgs = self.lint_file(
^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/anaconda_linter/lint/init.py", line 851, in lint_file
recipe = _recipe.Recipe(recipe_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/envs/linter/lib/python3.12/site-packages/percy/render/recipe.py", line 107, in init
self.recipe_dir = recipe_file.parent
^^^^^^^^^^^^^^^^^^
AttributeError: 'str' object has no attribute 'parent'
ERROR conda.cli.main_run:execute(124): conda run conda-lint /tmp/abs_34bssbgmb8/clone failed. (See above for error)
@boldorider4 You can skip these three tests on osx because they require missing *.avro files:
FAILED py-polars/tests/unit/io/test_iceberg.py::test_scan_iceberg_plain - pol...
FAILED py-polars/tests/unit/io/test_iceberg.py::test_scan_iceberg_filter_on_partition
FAILED py-polars/tests/unit/io/test_iceberg.py::test_scan_iceberg_filter_on_column
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_02bpls3vze/clone` failed. (See above for error)
The following problems have been found:
===== ERRORS =====
clone/recipe/meta.yaml:34: missing_wheel: For pypi packages, wheel should be present in the host section
===== Final Report: =====
1 Error and 0 Warnings were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_a7gl_ihyrn/clone` failed. (See above for error)
The following problems have been found:
===== ERRORS =====
clone/recipe/meta.yaml:33: missing_wheel: For pypi packages, wheel should be present in the host section
===== Final Report: =====
1 Error and 0 Warnings were found
@boldorider4 The only test that failed is
FAILED py-polars/tests/unit/test_polars_import.py::test_polars_import
> raise RuntimeError(msg)
E RuntimeError: measuring import timings failed
...
E Traceback (most recent call last):
E File "<string>", line 1, in <module>
E ModuleNotFoundError: No module named 'polars'
There was an upstream issue https://github.com/pola-rs/polars/issues/14442. You can skip the test or patch it. It's up to you
An error on osx:
File "/var/folders/sy/f16zz6x50xz3113nwtb9bvq00000gp/T/abs_92ywvsgg57/croot/polars_1713342610693/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placeho/lib/python3.10/zipfile.py", line 1378, in _RealGetContents
LookupError: unknown encoding: cp437
Is there something wrong with Python?
@boldorider4 The only test that failed is
FAILED py-polars/tests/unit/test_polars_import.py::test_polars_import
> raise RuntimeError(msg)
E RuntimeError: measuring import timings failed
...
E Traceback (most recent call last):
E File "<string>", line 1, in <module>
E ModuleNotFoundError: No module named 'polars'
There was an upstream issue pola-rs/polars#14442. You can skip the test or patch it. It's up to you
thank you @skupr-anaconda, I noticed it myself, I just need to skip it on all platforms.
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_41_loda37d/clone` failed. (See above for error)
The following problems have been found:
===== ERRORS =====
clone/recipe/meta.yaml:33: missing_wheel: For pypi packages, wheel should be present in the host section
===== Final Report: =====
1 Error and 0 Warnings were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_36ilziddxh/clone` failed. (See above for error)
The following problems have been found:
===== ERRORS =====
clone/recipe/meta.yaml:35: missing_wheel: For pypi packages, wheel should be present in the host section
===== Final Report: =====
1 Error and 0 Warnings were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_22w1rvzyd8/clone` failed. (See above for error)
The following problems have been found:
===== WARNINGS =====
clone/recipe/meta.yaml:33: host_section_needs_exact_pinnings: Linked libraries host should have exact version pinnings.
===== Final Report: =====
0 Errors and 1 Warning were found
Linter check found the following problems:
ERROR conda.cli.main_run:execute(124): `conda run conda-lint /tmp/abs_2b9me42f5u/clone` failed. (See above for error)
The following problems have been found:
===== WARNINGS =====
clone/recipe/meta.yaml:33: host_section_needs_exact_pinnings: Linked libraries host should have exact version pinnings.
===== Final Report: =====
0 Errors and 1 Warning were found
|
2025-04-01T06:36:42.013103
| 2022-03-06T08:49:56
|
1160554334
|
{
"authors": [
"fremartini",
"marfavi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:158",
"repo": "AnalogIO/coffeecard_app",
"url": "https://github.com/AnalogIO/coffeecard_app/issues/178"
}
|
gharchive/issue
|
Login page overflows
The login page has a small overflow when a presumably too large keyboard is present
Duplicate #151
|
2025-04-01T06:36:42.034426
| 2017-12-05T22:53:32
|
279560753
|
{
"authors": [
"DannyLebel",
"pjcozzi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:159",
"repo": "AnalyticalGraphicsInc/cesium",
"url": "https://github.com/AnalyticalGraphicsInc/cesium/issues/6032"
}
|
gharchive/issue
|
Camera.SetView() gives inaccurate orientation using direction and up vectors
Description
The view can be set with the method Camera.SetView() using direction and up vectors. Most of the time the behavior is correct. But in some angles, setView() does not behave as expected. The issue has been identified to occur when using a direction vector near to Cartesian3.UNIT_Y. In fact, this occurs when pointing a little under the middle of the earth (between 0 and ~-2.56radians). In the range of 0 to -2.56, the view is exactly the same as if the angle was positive. (a pitch rotation of -2.56 radians down gives a pitch rotation of 2.56 radians up). As soon as the rotation go to ~-2.565radians, the rotation goes down as expected.
Expected behavior
Using Camera.SetView with direction and up vectors points to the direction vector and the up vector is directly up.
Steps to reproduce
The issue is a little obscure. Here is a small sandcastle project to demonstrate the issue.
Use keys 1, 2, 3, 4 to switch from views
key 1 : set direction vector to vector a
key 2 : set direction vector to vector aPrime
key 3 : set direction vector to vector b
key 4 : set direction vector to vector bPrime
Alternate view by using key 1 and 2.
Notice that the view changes as expected and the view is mirrored since both vectors have only the z component opposite
Alternate view by using key 3 and 4.
Notice that the view does not change even if vectors b and bPrime have their z component opposed
//Start of example
var viewer = new Cesium.Viewer('cesiumContainer');
var scene = viewer.scene;
var canvas = viewer.canvas;
canvas.setAttribute('tabindex', '0');
canvas.onclick = function() {
canvas.focus();
};
var direction = Cesium.Cartesian3.UNIT_Y.clone();
var up = Cesium.Cartesian3.UNIT_Z.clone();
// initial position and orientation
viewer.camera.position = new Cesium.Cartesian3(0, -20000000, 0);
viewer.camera.setView(
{
orientation:
{
direction: direction,
up: up
}});
// setting view by switching from a to aPrime as direction changes view as expected
var a = new Cesium.Cartesian3(0, 0.9989980940754456, 0.044752743308394995);
var aPrime = new Cesium.Cartesian3(0, 0.9989980940754456, -0.044752743308394995);
// setting view by switching from b to bPrime does not change the view
var b = new Cesium.Cartesian3(0, 0.9993908270190958, 0.03489949670250097);
var bPrime = new Cesium.Cartesian3(0, 0.9993908270190958, -0.03489949670250097);
var ninetyDegreeRotation = Cesium.Matrix3.fromRotationX(Math.PI/2);
document.addEventListener('keydown', function(e) {
var camera = viewer.camera;
// setting direction vector to use
switch (e.keyCode)
{
case '1'.charCodeAt(0):
direction = a;
console.log("direction = a : " + direction);
break;
case '2'.charCodeAt(0):
direction = aPrime;
console.log("direction = aPrime : " + direction);
break;
case '3'.charCodeAt(0):
direction = b;
console.log("direction = b : " + direction);
break;
case '4'.charCodeAt(0):
direction = bPrime;
console.log("direction = bPrime : " + direction); // when using bPrime as direction, the view is the same as b.
break;
default:
break;
}
up = Cesium.Matrix3.multiplyByVector(ninetyDegreeRotation, direction, up);
camera.setView(
{orientation:
{
direction: direction,
up: up
}});
}, false);
// End of example
@DannyLebel thanks for the detailed report and code example. If there's anything you can do to narrow this down or contribute a fix before someone gets to it, we'd appreciate it.
Sorry, I do not have a fix. I thought the issue was in our project so I investigated it until I figured it was in Cesium itself. My report describes what I know of the issue. If you have any questions I will be glad to try to answer.
|
2025-04-01T06:36:42.039401
| 2018-02-13T20:31:27
|
296875256
|
{
"authors": [
"hpinkos",
"jmack2424"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:160",
"repo": "AnalyticalGraphicsInc/cesium",
"url": "https://github.com/AnalyticalGraphicsInc/cesium/issues/6213"
}
|
gharchive/issue
|
PointGraphics heightReference crash
When attempting to clamp a Cesium.CustomDataSource entity to ground:
{
id: id,
name: name,
position: cartesian,
description: desciption,
point: {
pixelSize: 6,
color: color,
outlineColor: Cesium.Color.WHITE,
outlineWidth: 2,
heightReference: 1,
},
label: {
text: name,
horizontalOrigin : Cesium.HorizontalOrigin.CENTER,
verticalOrigin : Cesium.VerticalOrigin.BOTTOM,
font: '12pt roboto',
style: Cesium.LabelStyle.FILL_AND_OUTLINE,
outlineWidth: 4,
pixelOffset: new Cesium.Cartesian2(0, -9),
disableDepthTestDistance: Number.POSITIVE_INFINITY,
heightReference: 1,
}
Including the "heightReference" on the point crashes Cesium. Binding to GeoJsonDataSource or CzmlDataSource works, but there is no way to clamp the custom PointGraphics object to the terrain without crashing Cesium.
Chrome Version 63.0.3239.132 (Official Build) (64-bit)
Cesium 1.37
@jmack2424 can you please paste a complete code example that reproduces the crash in Sandcastle? https://cesiumjs.org/Cesium/Build/Apps/Sandcastle/
Thanks
Hannah, we found our issue, and forcing it to work in the Sandcastle helped us to find it. =)
Our custom data points were being serialized, which broke on the load call, because the Cesium references in the custom object could not be resolved.
Thanks @jmack2424, glad you were able to figure it out =)
|
2025-04-01T06:36:42.041293
| 2019-07-09T09:19:23
|
465668048
|
{
"authors": [
"OmarShehata",
"TNMoOn"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:161",
"repo": "AnalyticalGraphicsInc/cesium",
"url": "https://github.com/AnalyticalGraphicsInc/cesium/issues/7988"
}
|
gharchive/issue
|
How can I set clipping planes to primitive object?
Thanks a lot!
There's a few code examples on Sandcastles showing this, here's one:
https://cesiumjs.org/Cesium/Apps/Sandcastle/index.html?src=3D Tiles Clipping Planes.html
Please keep general questions like this on the Cesium forum: https://groups.google.com/forum/#!forum/cesium-dev. Feel free to post a follow up question there!
|
2025-04-01T06:36:42.046327
| 2016-08-30T15:59:29
|
174058540
|
{
"authors": [
"lilleyse",
"pjcozzi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:162",
"repo": "AnalyticalGraphicsInc/cesium",
"url": "https://github.com/AnalyticalGraphicsInc/cesium/pull/4256"
}
|
gharchive/pull-request
|
3D Tiles Transform Tests
Tests for https://github.com/AnalyticalGraphicsInc/cesium/pull/4130
Sorry about the branch confusion. This is merging into pnts-updates instead of 3d-tiles-transform mainly because of all the file renaming that happened in #4228.
I reworked ModelInstanceCollection extensively to simplify it and fix some bugs related to its interaction with shadows and derived commands. The functionality is the same though.
Any typed array memory concerns with ModelInstanceCollection? No because everything is converted to a separate data structure?
Other than these comments, code and tests look good. Did you run coverage?
Any typed array memory concerns with ModelInstanceCollection? No because everything is converted to a separate data structure?
Yeah that's correct.
Did you run coverage?
Yeah coverage is mostly solid.
I'm curious what you think about the inline comments.
Updated. I squashed some commits because of the change to z-up and EAST_NORTH_UP.
Code looks OK.
I'm curious what you think about the inline comments.
What comments?
What comments?
The ones you already looked at before.
Thanks @lasalvavida for the updated tiles. @pjcozzi this is ready to merge now.
Tests and Sandcastle example are good!
|
2025-04-01T06:36:42.050908
| 2018-09-26T20:13:27
|
364191380
|
{
"authors": [
"OmarShehata",
"cesium-concierge",
"ggetz",
"lilleyse"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:163",
"repo": "AnalyticalGraphicsInc/cesium",
"url": "https://github.com/AnalyticalGraphicsInc/cesium/pull/7083"
}
|
gharchive/pull-request
|
Fix draco model in Edge
Fixes #7079
@hpinkos can you review?
Thanks for the pull request @ggetz!
:heavy_check_mark: Signed CLA found.
Reviewers, don't forget to make sure that:
[ ] Cesium Viewer works.
[ ] Works in 2D/CV.
[ ] Works (or fails gracefully) in IE11.
I am a bot who helps you make Cesium awesome! Contributions to my configuration are welcome.
:earth_africa: :earth_americas: :earth_asia:
Since this only happens in master, and not any released version, should we remove the mention from CHANGES.md ?
Looks good to me!
Travis is failing because of https://github.com/AnalyticalGraphicsInc/cesium/issues/7076, but tests are passing here and locally.
|
2025-04-01T06:36:42.052991
| 2023-10-01T10:08:40
|
1920681409
|
{
"authors": [
"Anandsg",
"LynxSumit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:164",
"repo": "Anandsg/Hungry-hero",
"url": "https://github.com/Anandsg/Hungry-hero/issues/16"
}
|
gharchive/issue
|
CORS Error is there while fetching data on Main Section.
I think this is a major issue cuz starters and newcomers wont be able to see anything.
@LynxSumit yes it is expected, have you installed CORS extension from chrome? please install and try sometime browser wont allow us to fetch API's from external sites
@LynxSumit Hope you're able to access the site now? are we good to close this issue
sure i will review once you made PR
|
2025-04-01T06:36:42.058677
| 2022-04-08T07:38:07
|
1196955812
|
{
"authors": [
"AndCake",
"rochacbruno"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:165",
"repo": "AndCake/micro-plugin-lsp",
"url": "https://github.com/AndCake/micro-plugin-lsp/issues/11"
}
|
gharchive/issue
|
Looking for Cargo.toml in a python project
HI, thanks for this amazing project
I tried following the instructions on the readme and opened a Python file.
message on status bar:
could not find `Cargo.toml` in `path` or any parent directory
It also asked me to add rls to my rustup toolchain, so looks like this is not considering the filetype.
Hi! I updated the messaging for this. Since we now have default values for the LSP configuration, it is unnecessary to show such messages in the status bar. It is now only logged.
|
2025-04-01T06:36:42.061944
| 2023-07-04T02:51:48
|
1787052495
|
{
"authors": [
"ifightcrime",
"joeylin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:166",
"repo": "Andarist/react-textarea-autosize",
"url": "https://github.com/Andarist/react-textarea-autosize/issues/380"
}
|
gharchive/issue
|
textarea-autosize conflict with custom key bind
I want to a function like this: when I use shift + enter, I can wrap lines, and the textarea can autoheight,
when I use enter key, I can send the message,
BUT, when I bind the enter key to textarea(with this plugin), I found enter key was used to wrapping lines default,
Does the plugin support this feature? I understand this is a more general scenario.
thanks very much
@joeylin in case it helps I think I figured this out. You can just do an e.preventDefault(). My resulting code looks something like this:
<textarea
onKeyDown={(e) => {
if (!e.shiftKey && e.key === 'Enter') {
e.preventDefault();
handleSendMessage();
}
}}
/>
|
2025-04-01T06:36:42.080085
| 2023-08-30T11:05:10
|
1873444882
|
{
"authors": [
"AndreRojasMartinsson",
"Scyfren"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:167",
"repo": "AndreRojasMartinsson/roduxutils",
"url": "https://github.com/AndreRojasMartinsson/roduxutils/pull/4"
}
|
gharchive/pull-request
|
Fix non-method error
I get this error when trying to use the prepare feature:
error TS roblox-ts: Attempted to assign non-method where method was expected.
18 prepare: () => {
~~~~~~~~~~~~~~~~
19 return {
~~~~~~~~~~~~~~~~~~~~~~~~
...
23 };
~~~~~~~~~~~~~~~~~~
24 },
I had to convert it to a method instead:
prepare() {
return {
payload: {
promptId: HttpService.GenerateGUID(),
},
};
},
But this compiles incorrectly:
prepare = function(self)
return {
payload = {
promptId = HttpService:GenerateGUID(),
},
}
end,
The self shouldn't be there.
This PR fixes that by changing some types to use a property instead of a method
Looks good to me.
|
2025-04-01T06:36:42.084784
| 2023-02-01T21:02:11
|
1566822768
|
{
"authors": [
"Okanda",
"jhpaques",
"salkin-mada"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:168",
"repo": "AndreasBackx/waycorner",
"url": "https://github.com/AndreasBackx/waycorner/issues/14"
}
|
gharchive/issue
|
core dumped
Arch Linux 6.1.8
Wayland 1.21.0
Sway 1.8
❯ waycorner
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
zsh: IOT instruction (core dumped) waycorner
It worked some weeks ago. Now no more :)
Something I am missing here?
Same error here
$ waycorner
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
Abandon (core dumped)
not really any info to find here:
~
❯ export RUST_LOG=trace
~
❯ waycorner
[2023-02-07T05:10:47Z DEBUG waycorner::config] Replacing ~/ with $HOME/
[2023-02-07T05:10:47Z INFO waycorner::config] Using config: /home/salkin/.config/waycorner/config.toml
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
zsh: IOT instruction (core dumped) waycorner
~
❯ export RUST_LOG=debug
~
❮ waycorner
[2023-02-07T05:11:07Z DEBUG waycorner::config] Replacing ~/ with $HOME/
[2023-02-07T05:11:07Z INFO waycorner::config] Using config: /home/salkin/.config/waycorner/config.toml
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
zsh: IOT instruction (core dumped) waycorner
~
❯ export RUST_LOG=info
~
❮ waycorner
[2023-02-07T05:11:19Z INFO waycorner::config] Using config: /home/salkin/.config/waycorner/config.toml
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
zsh: IOT instruction (core dumped) waycorner
~
❯ export RUST_LOG=warn
~
❮ waycorner
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
zsh: IOT instruction (core dumped) waycorner
~
❯ export RUST_LOG=error
~
❮ waycorner
[wayland-client error] Attempted to dispatch unknown opcode 0 for wl_shm, aborting.
zsh: IOT instruction (core dumped) waycorner
coredumpctl info
Is this a (g)libc versioning issue?
PID: 29687 (waycorner)
UID: 1000 (salkin)
GID: 1000 (salkin)
Signal: 6 (ABRT)
Timestamp: Tue 2023-02-07 06:11:44 CET (8min ago)
Command Line: waycorner
Executable: /usr/bin/waycorner
Control Group: /user.slice/user-1000.slice/session-1.scope
Unit: session-1.scope
Slice: user-1000.slice
Session: 1
Owner UID: 1000 (salkin)
Boot ID: ~~..ada3..~~
Machine ID: ~~..e5f8..~~
Hostname: ~~abcd~~
Storage: /var/lib/systemd/coredump/core.waycorner.1000.0f213c497c54456ea0db3b20e619ada3.29687.1675746704000000.zst (present)
Size on Disk: 115.8K
Message: Process 29687 (waycorner) of user 1000 dumped core.
Stack trace of thread 29687:
#0 0x00007f91d480464c n/a (libc.so.6 + 0x8864c)
#1 0x00007f91d47b4938 raise (libc.so.6 + 0x38938)
#2 0x00007f91d479e53d abort (libc.so.6 + 0x2253d)
#3 0x000055d41c8fb803 n/a (waycorner + 0x111803)
#4 0x00007f91d4ae7d65 n/a (libwayland-client.so + 0x7d65)
#5 0x00007f91d4ae7ffc wl_display_dispatch_queue_pending (libwayland-client.so + 0x7ffc)
#6 0x00007f91d4aeac10 wl_display_roundtrip_queue (libwayland-client.so + 0xac10)
#7 0x000055d41c85454f n/a (waycorner + 0x6a54f)
#8 0x000055d41c86560f n/a (waycorner + 0x7b60f)
#9 0x000055d41c87a6c2 n/a (waycorner + 0x906c2)
#10 0x000055d41c8684ed n/a (waycorner + 0x7e4ed)
#11 0x000055d41c86aa73 n/a (waycorner + 0x80a73)
#12 0x000055d41c8478b9 n/a (waycorner + 0x5d8b9)
#13 0x000055d41c9ae96f n/a (waycorner + 0x1c496f)
#14 0x000055d41c868f48 n/a (waycorner + 0x7ef48)
#15 0x00007f91d479f290 n/a (libc.so.6 + 0x23290)
#16 0x00007f91d479f34a __libc_start_main (libc.so.6 + 0x2334a)
#17 0x000055d41c832c55 n/a (waycorner + 0x48c55)
ELF object binary architecture: AMD x86-64
#15 fixes this.
The smithay-client-toolkit, wayland-client, and wayland-protocols crates were out of date.
|
2025-04-01T06:36:42.087328
| 2018-07-15T13:44:20
|
341322186
|
{
"authors": [
"AndreiMisiukevich",
"acaliaro"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:169",
"repo": "AndreiMisiukevich/ContextMenu",
"url": "https://github.com/AndreiMisiukevich/ContextMenu/issues/2"
}
|
gharchive/issue
|
Example with code
Hi @AndreiMisiukevich
could you translate your XAML example in CSHARP?
Thanks
Alessandro
Hi.
Sure, I will do it and let you know
Thanks for feedback
wow, you are fast. Thanks!
Readme is updated =)
Thanks I take a look
|
2025-04-01T06:36:42.088140
| 2024-11-26T21:50:02
|
2696212587
|
{
"authors": [
"noleakk"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:170",
"repo": "AndrejStojkovic/FishNet-Voice",
"url": "https://github.com/AndrejStojkovic/FishNet-Voice/pull/2"
}
|
gharchive/pull-request
|
Update VoiceChat.cs
Fixed stuttering issues with Voice Activation, optimized allocation, and threading.
Also added a function for getting the mic input volume
|
2025-04-01T06:36:42.116245
| 2014-04-17T21:59:27
|
31766685
|
{
"authors": [
"AndrewFromMelbourne",
"chandra50",
"matttheorbiter",
"michaelhanin"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:171",
"repo": "AndrewFromMelbourne/raspi2png",
"url": "https://github.com/AndrewFromMelbourne/raspi2png/issues/2"
}
|
gharchive/issue
|
Rotation
Hello,
This script work fine when no rotation, but with rotate=1 in kernel, it freeze and does not work .. any idea ? thx
Hi Andrew,
Revisiting the raspi2png on rotated screen. Were you able to get some
update on solving the rotated screen problem.
Thanks a lot.
Regards,
Chandra
On Sun, Oct 5, 2014 at 3:28 AM, Andrew Duncan<EMAIL_ADDRESS>wrote:
Hi Sorry for the delay. I have been busy on other things. No the change
only related to compiling in the math library. I will try to get back to
your issue shortly. Again sorry for the delay.
—
Reply to this email directly or view it on GitHub
https://github.com/AndrewFromMelbourne/raspi2png/issues/2#issuecomment-57932147
.
Hi Chandra,
Sorry for the very long delay. Other things came up and then I completely forgot. I am sorry!
As you will see above I have finally managed to reproduce the issue and I have raised an issue on the firmware.
The good news is I think I have a work around. Try using the -b option of omxplayer. The option adds a black layer underneath the playing video. As your video is full screen, you won't see it. However, from my experiments by adding the extra layer, the video is no longer distorted. So try
omxplay -b yourvideo.mpg
Hi Chandra,
I am sorry that this took so long (on my part). The Raspberry PI firmware has now been fixed (you will need to use rpi-update for now to get the latest firmware). Once the bug was reported on the Raspberry Pi firmware the fix was made very quickly. Once again sorry that it took so long for me to raise the issue. I hope that it now works for you.
Andrew
Hi Andrew,
Thanks a lot. It works perfectly. My sincere gratitude in you
addressing this problem.
One suggestion for raspi2png - a setting for resolution (to reduce the
file size).
Thanks again.
With kind regards,
Chandra
On Mon, Mar 23, 2015 at 2:58 AM, Andrew Duncan<EMAIL_ADDRESS>wrote:
Hi Chandra,
I am sorry that this took so long (on my part). The Raspberry PI firmware
has now been fixed (you will need to use rpi-update for now to get the
latest firmware). Once the bug was reported on the Raspberry Pi firmware
the fix was made very quickly. Once again sorry that it took so long for me
to raise the issue. I hope that it now works for you.
Andrew
—
Hi Chandra,
Thanks for your patience.
I will have a think about resolution ... so just to be clear are you wanting to reduce the file size or have a smaller (fewer pixel) image. Reducing the file size would be possible if I created a raspi2jpg program, which is possible. Reducing the number of pixels is pretty trivial and could implemented by specifying the desired width (and/or height) on the command line. Let me know.
With regards to this issue, are you happy to close it now?
Thanks,
Andrew
Hi Andrew,
We are delighted with your solution you have provided and you may close
the issue. I have tested it a few times.
Yes, we would like to have a reduced file size as an option to raspi2png.
Currently I am using a program to create a lower resolution jpg file (from
snapshot.png) to reduce the file size.
Thanks.
Regards,
Chandra
On Mon, Mar 23, 2015 at 11:31 AM, Andrew Duncan<EMAIL_ADDRESS>wrote:
Hi Chandra,
Thanks for your patience.
I will have a think about resolution ... so just to be clear are you
wanting to reduce the file size or have a smaller (fewer pixel) image.
Reducing the file size would be possible if I created a raspi2jpg program,
which is possible. Reducing the number of pixels is pretty trivial and
could implemented by specifying the desired width (and/or height) on the
command line. Let me know.
With regards to this issue, are you happy to close it now?
Thanks,
Andrew
—
Reply to this email directly or view it on GitHub
https://github.com/AndrewFromMelbourne/raspi2png/issues/2#issuecomment-84823612
.
Hi Andrew,
Thanks for all your help, I have been using raspi2png quite a bit.
Right now it has been painful to reduce the file size for copying the
image to the cloud - if you could kindly provide support for raspi2jpg or a
way to reduce the number of pixel would be great.
Thanks.
Regards,
Chandra
On Sun, Mar 22, 2015 at 11:01 PM, Andrew Duncan<EMAIL_ADDRESS>wrote:
Hi Chandra,
Thanks for your patience.
I will have a think about resolution ... so just to be clear are you
wanting to reduce the file size or have a smaller (fewer pixel) image.
Reducing the file size would be possible if I created a raspi2jpg program,
which is possible. Reducing the number of pixels is pretty trivial and
could implemented by specifying the desired width (and/or height) on the
command line. Let me know.
With regards to this issue, are you happy to close it now?
Thanks,
Andrew
Hi Chandra,
I am still struggling to understand what you require. You can reduce the number of pixels in the snapshot using either the --width or --height command line options
For reduced resolution use the --width and/or --height to specify the dimensions of the snapshot. If you just specify one the other is calculated from the the aspect ratio of the screen.
Unfortunately, at the moment that probably won't work for you as the is an issue open about this feature on rotated displays and related issue for the Raspberry Pi firmware.
Could you please explain why using the --width and/or --height options is not a solution for you.
Andrew
Hi Chandra.
The bug that prevented resizing on rotated screens is now addressed in the latest firmware release.
Andrew
Hi Andrew,
Thanks a lot. I will test it out this weekend.
With kind regards,
Chandra
On Wed, Jun 17, 2015 at 10:51 PM, Andrew Duncan<EMAIL_ADDRESS>wrote:
Hi Chandra.
The bug that prevented resizing on rotated screens is now addressed in the
latest firmware release.
Andrew
—
Reply to this email directly or view it on GitHub
https://github.com/AndrewFromMelbourne/raspi2png/issues/2#issuecomment-113044989
.
I have installed and reinstalled this on my Pi 2. I have successfully taken two screenshot. Then it always fails with raspi2png: vc_dispmanx_snapshot() failed. Anybody have any idea what's happening. Why it works one second, but not the next?
Followup. If I delete the directory. Then completely reinstall, sometimes it will work. Sometimes it won't. Maddening. Any help appreciated.
OK I will see if I can reproduce this. It would be good if you opened a new issue rather than adding to a closed and probably unrelated issue.
Sorry about that. I will. Just was glad to find some people talking about the program and got excited I guess. Thanks.
No problem. I am the author of the program.
Think I figured out the problem. When Scraping in RetroPie on a Raspberry Pi, while doing the final step, choosing the metadata, raspi2png always fails. Today I can’t get it to fail otherwise, so maybe it has something to do with the scrape utility. So far today raspi2png is working like a champ for me.
Thanks for the program, I will mention it in the credits of the book I’m writing, it’s been indispensable now that I’ve got it running.
Matt
Matt Smith, NBCT
<EMAIL_ADDRESS>AP 12 English/US History/Sociology
Valley High School
Albuquerque, New Mexico
On Jul 19, 2015, at 6:04 PM, Andrew Duncan<EMAIL_ADDRESS>wrote:
No problem. I am the author of the program.
—
Reply to this email directly or view it on GitHub https://github.com/AndrewFromMelbourne/raspi2png/issues/2#issuecomment-122719851.
|
2025-04-01T06:36:42.119137
| 2023-11-02T11:12:56
|
1974023083
|
{
"authors": [
"AndrewWalsh",
"colin6-work"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:172",
"repo": "AndrewWalsh/openapi-devtools",
"url": "https://github.com/AndrewWalsh/openapi-devtools/issues/9"
}
|
gharchive/issue
|
Feature Request: Session Continuation and OpenAPI Specification Import
Currently, closing the browser results in a loss of the generated OpenAPI specification data. To enhance usability and continuity, could we consider the following features?
Implementing a session save feature that allows users to pause and resume their recording sessions at a later time.
Providing an option to upload a JSON file of a previously downloaded OpenAPI specification, enabling users to continue recording from where they left off.
Great suggestion @colin6-work, I have things set up to make this possible. I'll implement this in the next version.
This is now in release v1.2.0.
|
2025-04-01T06:36:42.139698
| 2023-10-24T18:20:28
|
1959825265
|
{
"authors": [
"AndyTheFactory"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:173",
"repo": "AndyTheFactory/newspaper4k",
"url": "https://github.com/AndyTheFactory/newspaper4k/issues/449"
}
|
gharchive/issue
|
Newspaper not scraping full content or picking up random sentences in article.
Issue by shashank7596
Thu Apr 9 16:05:27 2020
Originally opened as https://github.com/codelucas/newspaper/issues/799
Hello everyone.
There's few urls especially for a particular site the newspaper is not scraping the full content for some reason. It randomly picks some sentences in the article. Can anyone help in fixing this issue. Newspaper scraper been very crucial in our project right now and for some websites it's not working. For different articles in the same domain it behaves differently.
Below are some example urls :
https://www.clinicaltrials.gov/ct2/show/study/NCT00034216?cond=breast+cancer&lupd_s=03%2F26%2F2020&lupd_d=14
https://www.clinicaltrials.gov/ct2/show/study/NCT04335006?cond=triple+negative+breast+cancer
https://www.clinicaltrials.gov/ct2/show/study/NCT04338269
https://www.clinicaltrials.gov/ct2/show/NCT04332653?cond=breast+cancer&lupd_s=03%2F25%2F2020&lupd_d=14
Newspaper is not suited for scraping research papers or studies
|
2025-04-01T06:36:42.142337
| 2019-06-18T16:00:56
|
457567305
|
{
"authors": [
"Andychochocho",
"oantila"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:174",
"repo": "Andychochocho/android-testing-tool",
"url": "https://github.com/Andychochocho/android-testing-tool/issues/6"
}
|
gharchive/issue
|
Save Logs button is functional even if device is not connected
The Save Logs button is functional even if no device is detected. Button should be disabled if there's no connected device, or an error message should pop up saying "Please connect your device first" when button is clicked. Let's implement which ever is the easiest/fastest to do.
Steps to reproduce:
Install build: https://www.dropbox.com/s/1l1hxinlwni6xhn/Android Test Tool.zip?dl=0
Launch app with no device connected
Click Save Logs button
I disabled click functionality when device is not connected. Implemented in 76b2610
|
2025-04-01T06:36:42.154228
| 2024-04-02T08:55:26
|
2219942155
|
{
"authors": [
"AngeloAvv",
"YukiAttano"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:175",
"repo": "AngeloAvv/flutter_flavorizr",
"url": "https://github.com/AngeloAvv/flutter_flavorizr/issues/251"
}
|
gharchive/issue
|
Add ignore linter rules to autogenerated files
Please add "ignore_for_file" statements to all autogenerated files that will always be overridden like the main_flavor.dart and flavors.dart.
For example:
"always_use_package_imports"
"prefer_relative_imports"
There are likely more that are of use.
Questions and bugs
If you need help with the use of the library or you just want to request new features, please use the Discussions section of the repository. Issues opened as questions will be automatically closed.
|
2025-04-01T06:36:42.160222
| 2021-02-28T22:33:17
|
818330467
|
{
"authors": [
"paranarimasu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:176",
"repo": "AnimeThemes/animethemes-server",
"url": "https://github.com/AnimeThemes/animethemes-server/issues/123"
}
|
gharchive/issue
|
API Date Filters
There should exist a Date Filter in the API Filter framework that uses the Range Filter strategy described in #121. All of our non-Pivot Models use the standard created_at and updated_at timestamps, which allow us to perform queries on objects created or updated within given ranges. Currently, this can only be accomplished by sorting results on these fields.
This has been implemented through filter conditions in commit f21f41e3a65d77f6a77cc92d9902fe13ba3f8528, date validation in commit eb0991a642ed12e3739679a74046c918bc034ca7, and greater precision timestamps in 6192b9e31ff8c139e99ff0bf45b4a701a15b4cf5.
Adopted date format (to match resource attributes output format)
YYYY-MM-DDTHH:mm:ss.SSSSSS
All date components except for year are optional.
|
2025-04-01T06:36:42.161682
| 2024-05-12T10:41:59
|
2291275646
|
{
"authors": [
"DARSHANITRIPATHI",
"Princegupta101"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:177",
"repo": "Anishkagupta04/RAPIDOC-HEALTHCARE-WEBSITE-",
"url": "https://github.com/Anishkagupta04/RAPIDOC-HEALTHCARE-WEBSITE-/issues/53"
}
|
gharchive/issue
|
Add Topics
In GSSoC'24, GitHub Topics will help the discoverability of your project.
I see that you already have great topics on your repository!
I would recommend adding the name of the company like the software you use to build like "vs-code, ghdesktop" to improve your discoverability.
If you are happy with the topics you have, feel free to close this issue. 👍
Please assign this issue to me.
|
2025-04-01T06:36:42.163128
| 2023-12-21T07:52:15
|
2051900790
|
{
"authors": [
"ChrisTG742",
"codingbeast"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:178",
"repo": "Anjok07/ultimatevocalremovergui",
"url": "https://github.com/Anjok07/ultimatevocalremovergui/issues/1042"
}
|
gharchive/issue
|
how to use command line
Hi , Can you please guide me how to use this by command line(I don't want gui ) ?
The command line feature was removed when the GUI version was published. I just made a fork to re-implement the ability to use command line parameters again.
Thankyou can you please share the link I also want to contribute.
@codingbeast I haven't commited anything yet, will share it as soon as I found some time to do so.
|
2025-04-01T06:36:42.197093
| 2018-08-26T00:55:35
|
354054884
|
{
"authors": [
"Anonymousdog",
"MechX2"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:180",
"repo": "Anonymousdog/displaycameras",
"url": "https://github.com/Anonymousdog/displaycameras/issues/4"
}
|
gharchive/issue
|
displaycameras.conf.1920x1200
Any way we can get a default configuration file for monitors with a resolution of 1920x1200? Maybe a 6 camera setup.
16x9 feeds won't fit cleanly on a 16x10 monitor. 16x10 windows will distort the feeds but leave no black bars on sides of each window. Which would you prefer (stretch or bars)?
You the man! Stretch.
https://github.com/Anonymousdog/displaycameras/blob/master/example_layouts/layout.conf.1920x1200.6on4x4 rotates six feeds through a 2x2 matrix. 3x2 or 2x3 would be very distorted. Recommend feeds no larger than the window size, 960x600.
|
2025-04-01T06:36:42.198139
| 2020-03-24T17:04:15
|
587124499
|
{
"authors": [
"epsilon-0"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:181",
"repo": "AnsiMail/AnsiMail",
"url": "https://github.com/AnsiMail/AnsiMail/issues/17"
}
|
gharchive/issue
|
Dummy bug for checking reports
will be closed shortly
Checking the addition of comment on an issue.
Checking closing an issue (final check hopefully)
|
2025-04-01T06:36:42.206171
| 2024-08-12T12:25:01
|
2460856033
|
{
"authors": [
"bclavie",
"khanzzirfan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:182",
"repo": "AnswerDotAI/rerankers",
"url": "https://github.com/AnswerDotAI/rerankers/issues/24"
}
|
gharchive/issue
|
[Offline Support] Pre-packaging Docker Image for Re-Ranker with Pre-Downloaded Libraries in strict proxy environments
Issue: Pre-packaging Docker Image for Re-Ranker with Pre-Downloaded Libraries
Description:
The current setup for the Re-Ranker application involves downloading libraries and packages at runtime when the Docker container is initiated. This approach introduces several challenges, particularly in environments where internet access is restricted or where applications are subject to strict network proxies.
In our deployment environment, all network traffic is routed through proxies, and internet access may be limited or heavily controlled. As a result, the application faces delays, failures, or interruptions when attempting to download required libraries and packages at runtime. This dependency on real-time downloading can lead to instability and unreliability, especially when running the application in environments with such network constraints.
Proposal:
To mitigate these issues, the Flash Rank Re-Ranker Docker image should be pre-packaged with all necessary libraries, dependencies, and language model (LLM) packages already downloaded and included. By doing so, the application will be fully self-contained and will not need to access external resources when the container is started. This ensures that the application can run smoothly and reliably, even in environments with strict network controls or limited internet access.
Benefits:
Improved Reliability: The application will no longer be dependent on external network conditions to function correctly.
Faster Startup Times: Since all dependencies are pre-downloaded, the container can start up more quickly without waiting for network downloads.
Network Independence: The application can run in offline environments or environments with restricted internet access.
Consistency: Ensures that the same versions of libraries and packages are used across different environments, reducing the risk of version conflicts or incompatibilities.
This approach will significantly enhance the robustness and reliability of the Flash Rank Re-Ranker application in production environments.
Hey!
I'm not quite sure that this issue is within scope of what we'd like to do here. rerankers is purposefully a very light, thin library so that it can be slotted in anywhere, and building docker images with other packages for various usecases isn't what we want to do for it at all, since it's simple enough for users to do so and only include the packages that they personally care about.
Is there any fundamental issue with the library that makes it difficult for you to build your own Docker images including it? If so, I'd be more than happy to take a look at what's causing the issues!
|
2025-04-01T06:36:42.321512
| 2021-10-03T19:09:22
|
1014496551
|
{
"authors": [
"AkhilChoubey",
"Anveshreddy18-collab"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:183",
"repo": "Anveshreddy18-collab/Anveshreddy18-collab.github.io",
"url": "https://github.com/Anveshreddy18-collab/Anveshreddy18-collab.github.io/pull/16"
}
|
gharchive/pull-request
|
UI Fixes - Font styles , Responsivness , Text align , Image resize, Aligned Arrow img perfectly at middle of the page for both mobile and desktop.
I have devoted 2 hrs. on this , please merge this request under hacktoberfest-accepted . Thank You!!
I have devoted 2 hrs. on this , please merge this request under hacktoberfest-accepted . Thank You!!
Thanks for your efforts.
|
2025-04-01T06:36:42.338122
| 2020-12-18T22:56:32
|
771220609
|
{
"authors": [
"Narek1994",
"RayJiang16"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:184",
"repo": "AnyImageProject/AnyImageKit",
"url": "https://github.com/AnyImageProject/AnyImageKit/issues/59"
}
|
gharchive/issue
|
Open ImageEditorController after image selection.
Hi Thnaks for this lib.
Is it possible to open editor controller right after image picked, just pushing it, and not close and after that in completion open ImageEditorController?
Thanks.
Run Example project and open Avatar Picker controller.
Is this what you expected?
similar to that, but because we are dismissing picker controller and after that show editor contrroller, even without animation, on some miliseconds we can see main view controller, it would be nice to have an option to push EditorController right after image selection, just simple flag, openEditorAfterSelection for example.
Thanks
And Also Here are 2 screenshots, in editor screen I can see chinese text, but my app has english localization, and also in picker screen Cancel button is not visible.
I been updated Avatar Picker code. Fixed the issue that you can see main view controller on some milliseconds. Please download newest code on master.
We will do this feature like that:
Add flag in PickerOptionsInfo default is false.
Tap photo in AssetPickerViewController, check disable rules, push editor if pass check.
User tap "Done" in Editor, will back to Picker and select this photo.
Picker will dismiss immediately if limit is 1.
User tap "Cancel" in Editor, will back to Picker and unselect this photo.
Show live camera in cell instead of static camera image, in picker collectionView.
We will not do this feature, because app will request photo, camera and microphone permission when user first open picker. We should request permission when user need this feature instead of request all permission at first time.
I been updated Avatar Picker code. Fixed the issue that you can see main view controller on some milliseconds. Please download newest code on master.
We will do this feature like that:
Add flag in PickerOptionsInfo default is false.
Tap photo in AssetPickerViewController, check disable rules, push editor if pass check.
User tap "Done" in Editor, will back to Picker and select this photo.
Picker will dismiss immediately if limit is 1.
User tap "Cancel" in Editor, will back to Picker and unselect this photo.
Yep, this would be great, I have checked example app as well, right now it shows withou any lags, but because I am using also camera open in picker, I cant open Editor right after image was picked, because maybe it was picked with camera and already edited, I will wait to this feature to be released, I have tried something similar to that in my fork, but faced few issues while openning Editor.
Thanks.
I am sorry, will it be available for upcoming 2 weeks, I will need to publish my app after few weeks, wanted to know will this option be available at that period, thanks
We will done this feature in the next 2 weeks, but we may not release a new version. You should use SPM as a dependency or fork this repo then release to Cocoapods.
Thanks a lot, I am using SPM
We have done this feature, but we will not release a new version, you can use SPM to download the latest code.
You can see the sample code on AvatarPickerController.
works as expected, Thanks a lot
|
2025-04-01T06:36:42.340579
| 2024-12-19T07:02:59
|
2749435308
|
{
"authors": [
"AnyaCoder",
"Igrium"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:185",
"repo": "AnyaCoder/fish-speech-gui",
"url": "https://github.com/AnyaCoder/fish-speech-gui/issues/8"
}
|
gharchive/issue
|
Virus
So, basically every security software on the planet flags the release as a virus. What's going on with this? I had to load this up in a VM to stop Windows Defender from automatically deleting it.
https://nuitka.net/user-documentation/common-issue-solutions.html#id3
I hope this can help!
|
2025-04-01T06:36:42.360495
| 2023-01-06T19:36:41
|
1523046849
|
{
"authors": [
"Ninjagod1251",
"sabotagebeats"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:186",
"repo": "ApeWorX/ape",
"url": "https://github.com/ApeWorX/ape/pull/1211"
}
|
gharchive/pull-request
|
feat: change address arg to addressType
What I did
changed the address arg to an AddressType
made a test for it
fixes: #Ape-368
How I did it
How to verify it
Checklist
[ ] All changes are completed
[ ] New test cases have been added
[ ] Documentation has been updated
This one seems to be failing tests, trying to figure out why now.
|
2025-04-01T06:36:42.365995
| 2024-09-30T23:09:54
|
2557805018
|
{
"authors": [
"Ninjagod1251"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:187",
"repo": "ApeWorX/hosted-compiler",
"url": "https://github.com/ApeWorX/hosted-compiler/pull/53"
}
|
gharchive/pull-request
|
feat: added draft of structured errors
What I did
Refactored the except CompilerError as e: in compuile_project a structured error response for failed tasks caused by CompilerError. This introduces a new Pydantic model to format the error and ensure it's compatible with frontend systems via FastAPI.
fixes: #50
How I did it
Added a CompilerErrorModel class using Pydantic, which includes fields like message, line, column, and errorType.
Updated the logic within the compile_project function to handle and format CompilerError in line with the new model.
How to verify it
Trigger a compilation task that raises a CompilerError by submitting a Vyper contract with a syntax error.
Check the response from /exceptions/{task_id} to verify that the error details are returned in the correct format, as described in the issue.
Ensure that the response structure matches the frontend expectations (JSON format with specific error fields).
Checklist
[] All changes are completed
[] New test cases have been added for error handling and proper formatting of CompilerError.
[ ] Documentation has been updated to reflect the changes in /exceptions/{task_id} endpoint behavior.
What still needs to check and verify
Refactored the /exceptions/{task_id} endpoint to check specifically for tasks that failed due to a CompilerError and return a structured JSON response for the error.
Im still using pdb to make sure the obj can parse the error.
but it format good so far?
and do we need to change the get_task_exceptions to handle the new format?
|
2025-04-01T06:36:42.392727
| 2022-08-23T19:58:47
|
1348478839
|
{
"authors": [
"josh-wende",
"qwetqwe",
"zhanglonggao"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:188",
"repo": "ApolloAuto/apollo",
"url": "https://github.com/ApolloAuto/apollo/issues/14568"
}
|
gharchive/issue
|
Planning very sensitive to vehicle vibrations, frequently fails.
Using Apollo 7.0:
When driving autonomously, the vehicle very frequently stops because a planning cycle fails due to the Piecewise Jerk Nonlinear Speed Optimizer determining that the QP problem is infeasible. I have determined that this happens when the vehicle shakes, presumably since the IMU velocity/acceleration values jump around during shaking. I can't figure out why I am having this issue now when I never did before, any insight into what might be causing this is greatly appreciated.
Maybe current initial acceleration exceed the maximum acceleration because of vibrations, which leads to Ipopt solve failed
Hello! I have encounterd the same issue, have you sloved it ?
Using Apollo 7.0:
When driving autonomously, the vehicle very frequently stops because a planning cycle fails due to the Piecewise Jerk Nonlinear Speed Optimizer determining that the QP problem is infeasible. I have determined that this happens when the vehicle shakes, presumably since the IMU velocity/acceleration values jump around during shaking. I can't figure out why I am having this issue now when I never did before, any insight into what might be causing this is greatly appreciated.
I added some functions that specifically checked for infeasible starting conditions (either initial acceleration/velocity out of range, or the initial velocity plus acceleration combined with the maximum jerk not being enough to slow down without going over the max speed) and modifies the inputs to what can be feasible if so. It's a bit hacky, but it works. I found that just setting the initial acceleration to always be 0 for the solver made it a lot better too, without really affecting the behavior too much.
Hope this helps.
well, thanks for your reply. I'd like to try the second way first. May I ask how to set the initial acceleration to always be 0 ?
At 2024-03-30 01:42:59, "Josh Wende" @.***> wrote:
I added some functions that specifically checked for infeasible starting conditions (either initial acceleration/velocity out of range, or the initial velocity plus acceleration combined with the maximum jerk not being enough to slow down without going over the max speed) and modifies the inputs to what can be feasible if so. It's a bit hacky, but it works. I found that just setting the initial acceleration to always be 0 for the solver made it a lot better too, without really affecting the behavior too much.
Hope this helps.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you commented.Message ID: @.***>
I believe I just set s_ddot_init_ to 0.
|
2025-04-01T06:36:42.394711
| 2018-07-23T04:56:14
|
343487585
|
{
"authors": [
"VigiZhang",
"lianglia-apollo"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:189",
"repo": "ApolloAuto/apollo",
"url": "https://github.com/ApolloAuto/apollo/issues/5093"
}
|
gharchive/issue
|
Failed on compiling velodyne
New added blocking_queue.h has problem in take function argument std::vector<T>* v which will lead to error:
./modules/common/util/blocking_queue.h:98:8: error: member reference base type 'std::vector *' is not a structure or union
We fixed this issue at https://github.com/ApolloAuto/apollo/commit/f12c3891fe29479e530860b6c6304515c752fc80.
|
2025-04-01T06:36:42.399030
| 2018-08-24T18:53:01
|
353891511
|
{
"authors": [
"hlchen1043",
"jilinzhou",
"lianglia-apollo",
"natashadsouza"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:190",
"repo": "ApolloAuto/apollo",
"url": "https://github.com/ApolloAuto/apollo/issues/5488"
}
|
gharchive/issue
|
utilize proto file in c++
To whom it may concern,
We are trying to utilize Apollo protobuf topic in C++. When including the header file for protobuf we used the following script:
#include <modules/canbus/proto/chassis.pb.h>
#include <modules/localization/proto/localization.pb.h>
#include <modules/perception/proto/perception_obstacle.pb.h>
#include <string>
#include <math.h>
When I run the cmake file for C++, it returns the following error:
hlchen@in_dev_docker:/apollo/catkin_ws/src/DSRC/build$ make
[ 1%] Building CXX object CMakeFiles/openAV_DSRC.dir/bsm/CID_BSM_Receiver.cpp.o
In file included from /apollo/catkin_ws/src/DSRC/bsm/CID_BSM_Receiver.cpp:23:0:
/apollo/catkin_ws/src/DSRC/bsm/dsrc_apollo.h:2:45: fatal error: modules/canbus/proto/chassis.pb.h: No such file or directory
#include <modules/canbus/proto/chassis.pb.h>
^
compilation terminated.
make[2]: *** [CMakeFiles/openAV_DSRC.dir/bsm/CID_BSM_Receiver.cpp.o] Error 1
make[1]: *** [CMakeFiles/openAV_DSRC.dir/all] Error 2
make: *** [all] Error 2
Which by theory I thought it shouldn't happen. According to the protobuf documentation, it says when using protobuf file with c++, just replace .proto with .pb.h and it should work.
Does anyone have any suggestions?
Thank you!
bazel build is the only method we currently support. For general questions on how to make protobuf, I suggest post on protobuf community for questions.
Hi lianglia,
Thanks for the reply! Does that mean that I need to build Apollo first then build my code?
Thank you!
My suggestion is to use bazel build your added files. You can find examples in other folders on how defined protos are used as dependent libraries, for example https://github.com/ApolloAuto/apollo/blob/master/modules/planning/BUILD.
My guess: The error msg shows that the path to the header files are not configured properly in your CMakeList.txt. Configure cmake build in verbose mode and check "-I" flag in the console output. You can figure it out.
Closing this issue as it appears to be resolved. Feel free to reopen if you have additional questions. Thanks!
|
2025-04-01T06:36:42.407762
| 2018-09-07T22:51:54
|
358234694
|
{
"authors": [
"HouYuu",
"XiranBai",
"muleisheng",
"sopXx",
"xinwf"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:191",
"repo": "ApolloAuto/apollo",
"url": "https://github.com/ApolloAuto/apollo/issues/5706"
}
|
gharchive/issue
|
Fail to open config.pb.txt when running yolo_camera_detector_test
System information
**OS Platform and Distribution **: Host: Linux Ubuntu 16.04/Docker: 14.04
**Apollo installed from **: docker
**Apollo version **: 3.0
I am trying to run the yolo_camera_detector_test in module/perception/obstacle/camera/detector/yolo_camera_detector after building it with Bazel to see my own image result, but I got the following error:
_[NVBLAS] NVBLAS_CONFIG_FILE environment variable is NOT set : relying on default config filename 'nvblas.conf'
[NVBLAS] Cannot open default config file 'nvblas.conf'
[NVBLAS] Config parsed
[NVBLAS] CPU Blas library need to be provided
Running main() from gmock_main.cc
[==========] Running 3 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 3 tests from YoloCameraDetectorTest
[ RUN ] YoloCameraDetectorTest.model_init_test
WARNING: Logging before InitGoogleLogging() is written to STDERR
E0907 15:41:44.330113 5446 file.h:97] Failed to open file modules/perception/model/camera/yolo_camera_detector_config.pb.txt in text mode.
E0907 15:41:44.330178 5446 file.h:140] Failed to open file modules/perception/model/camera/yolo_camera_detector_config.pb.txt in binary mode.
F0907 15:41:44.330260 5446 yolo_camera_detector.cc:41] Check failed: GetProtoFromFile(FLAGS_yolo_camera_detector_config, &config_)
*** Check failure stack trace: ***
Aborted (core dumped)_
It looks like there is problem with the yolo_camera_detector_config.pb.txt, but I checked that file and it
seems normal. Could somebody help me with this issue? Or if anyone know a better way to run the perception module with my own data, please share with me. Thanks!
You could set the path of nvblas.conf by using export NVBLAS_CONFIG_FILE=/usr/local/cuda.
You could not find the file in /usr/local/cuda. You may need to create one first.
Hi muleisheng, thanks for your reply. I created the nvblas.conf, now the first 3 lines of warning became:
_[NVBLAS] NVBLAS_CONFIG_FILE environment variable is set to '/usr/local/cuda'
[NVBLAS] Config parsed
[NVBLAS] CPU Blas library need to be provided_
I put
NVBLAS_CPU_BLAS_LIB /usr/lib/libopenblas.so
in the nvblas.conf, but it still says 'CPU Blas library need to be provided', and the rest of the error messages are still the same. Could you share with me if you have any idea what is going on? Thank you very much!
I am closing this as the rest of the problem seems to because the file path is wrong
@sopXx, hey, how do you fix this problem, I also meet the same problem as you, can you share your method here?
@xinwf export NVBLAS_CONFIG_FILE=/usr/local/cuda
@HouYuu, I have done this, it seems that works right, but another problem CPU Blas library need to be provided still exists, how to solve it?
I have never seen the problem like that,you should post your detailed operation steps and system version information
| |
houyu
Samsung Electronics -
Software Engineer
18189118087
<EMAIL_ADDRESS>|
Signature is customized by Netease Mail Master
On 11/21/2018 13:53, 辛文飞 wrote:
@HouYuu, I have done this, it seems that works right, but another problem CPU Blas library need to be provided still exists, how to solve it?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
detailed info is at this issue, #6013
if I set the NVBLAS_CONFIG_FILE=/usr/local/cuda,then the log is config is parsed, but the CPU BLAS library problem still exists, if I set the set the NVBLAS_CONFIG_FILE=/usr/local/cuda/nvblas.conf, it won't generate such log, no tips about the CPU, it looks like everything is ok, but when I search this problem in issues, they all answer this question with the reply setting the NVBLAS_CONFIG_FILE=/usr/local/cuda, so, I set it with this common setting, keeping consistent with all of you, I don't konw which is right.
@xinwf Have you tried to use the released version of Apollo 2.5 or 3.0?
@HouYuu, no, I always use develop version from 2.5 to 3.0
|
2025-04-01T06:36:42.443341
| 2017-08-29T10:29:51
|
253605159
|
{
"authors": [
"Rikkola",
"ederign",
"kie-ci"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:192",
"repo": "AppFormer/uberfire",
"url": "https://github.com/AppFormer/uberfire/pull/814"
}
|
gharchive/pull-request
|
Reviews welcome - GUVNOR-3193 Change Project from a folder in a repo to a repository
Please do not merge.
This PR does the following:
Introduces the Project and Module classes. The old Project that was the maven project is now called Module, just to avoid confusion with the new Project. The new Project is the active project that the workbench is presenting in the UI or the Project that we can later build ( these changes still only support building single Modules ). The Project has an active branch and branch changes require a recreation of the Project ( We can change this later, but after spending months with the code it is quite clear that keeping this step will save us from an extra month of code changes and bug fixes. )
Introduces a Branch class. This is again to make our code a bit easier and safer to use. We used to just store the branches as strings, when a branch is a location in a repository that has a name and a path.
Removes asset-mgmt, this is no longer used
Renames org.guvnor.common.services.project.model.Repository to org.guvnor.common.services.project.model.MavenRepository so it is easier to know the difference between MavenReposity = used where we have dependencies and Repository = GIT repository that stores code.
Changes the REST API. We now hide repositories and access projects directly.
Structural changes in RepositoryService and ConfiguredRepositories to fix fired events (See GUVNOR-3555)
Just using the build servers to run the tests since they take few hours locally.
https://issues.jboss.org/browse/AF-648
The related PRs:
https://github.com/kiegroup/kie-soup/pull/7
https://github.com/AppFormer/uberfire/pull/814
https://github.com/kiegroup/drools/pull/1402
https://github.com/kiegroup/kie-wb-common/pull/1015
https://github.com/kiegroup/drools-wb/pull/566
https://github.com/kiegroup/jbpm-designer/pull/643
https://github.com/kiegroup/jbpm-wb/pull/832
https://github.com/kiegroup/kie-wb-distributions/pull/573
Build finished. 3508 tests run, 7 skipped, 0 failed.
Can you please move this PR for https://github.com/kiegroup/appformer ?
|
2025-04-01T06:36:42.450641
| 2019-10-23T17:46:23
|
511476959
|
{
"authors": [
"sjones4"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:193",
"repo": "AppScale/appscale",
"url": "https://github.com/AppScale/appscale/pull/3194"
}
|
gharchive/pull-request
|
Remove cassandra backend for PullQueues (with master)
This pull request is #3177 with master merged and conflicts resolved.
Daily build no. 6926
Daily build no. 6982
Demo is that the Taskqueue-e2e-Test test passes and logs show postgres is used:
2019-11-06 19:43:15,718 INFO appscale_taskqueue.py:347 Starting TaskQueue server on port 50002
2019-11-06 19:43:15,720 INFO appscale_taskqueue.py:296 TaskQueue server registered at /appscale/tasks/servers/<IP_ADDRESS>:50002
2019-11-06 19:43:15,727 INFO queue_manager.py:59 Updating queues for test-project
2019-11-06 19:43:15,758 INFO pg_connection_wrapper.py:42 Establishing new connection to Postgres server
2019-11-06 19:43:15,773 INFO queue.py:769 Ensuring "appscale_test-project" schema is created
2019-11-06 19:43:15,776 INFO queue.py:782 Ensuring "appscale_test-project.queues" table is created
2019-11-06 19:43:15,801 INFO queue_manager.py:59 Updating queues for test-project
|
2025-04-01T06:36:42.453619
| 2021-04-14T08:43:47
|
857670274
|
{
"authors": [
"dolpheen",
"svrnm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:194",
"repo": "Appdynamics/flutter-plugin",
"url": "https://github.com/Appdynamics/flutter-plugin/issues/43"
}
|
gharchive/issue
|
adeum-config: open failed: ENOENT (No such file or directory)
There is an error on AD initialization for release and debug mode on Android
Plugin version is 1.0.0
adeum-config- open failed- ENOENT (No such file or directory).log
Hi @dolpheen ,
is this error causing any trouble? This error comes from the underlying android SDK and is not related to the flutter plugin, so I can not fix this from here.
Hi @dolpheen ,
is this error causing any trouble? This error comes from the underlying android SDK and is not related to the flutter plugin, so I can not fix this from here.
Hi, @svrnm
Looks like everything is OK))
I will try to check AD SDK issues.
|
2025-04-01T06:36:42.464525
| 2021-10-09T21:08:02
|
1021806058
|
{
"authors": [
"Motko222",
"Qwinn1",
"ariegrossman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:195",
"repo": "Apple-Network/apple-blockchain",
"url": "https://github.com/Apple-Network/apple-blockchain/issues/18"
}
|
gharchive/issue
|
Ubuntu install still failing on farmers in 1.2.90
The same "MAC check failed" error that is described in Issue #16 is still happening trying to run apple 1.2.90 on a farmer in Ubuntu. Installs fine on harvesters, but I still have to run 1.2.30 for my farmer.
Actually, wow, even 1.2.30 is failing now.
Same problem in Debian - "MAC check failed". 1.2.30 works ok.
Okay, I figured out what that was. Because the apple init is failing, the target addresses it set in the config.yaml were null. When I fixed that, I was able to start with 1.2.30.
I'm having the same issue installing 1.2.90 on ubuntu 20.04. Going to try to install 1.2.30 first.
|
2025-04-01T06:36:42.469343
| 2018-05-21T16:16:11
|
324974912
|
{
"authors": [
"davidADSP",
"xueeinstein"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:196",
"repo": "AppliedDataSciencePartners/WorldModels",
"url": "https://github.com/AppliedDataSciencePartners/WorldModels/issues/10"
}
|
gharchive/issue
|
Confusion on implementation of 'get_mixture_coef()'
https://github.com/AppliedDataSciencePartners/WorldModels/blob/36ebabf24783991dfe6f86fa5e25c2bee141db77/rnn/arch.py#L29
This line is from the function get_mixture_coef(). I wonder if pi should have shape [-1, rollout_length, GAUSSIAN_MIXTURES, 1] to match the definition of GMM in MDN. In this way, SketchRNN's implementation is a special case of this one when Z_DIM=2.
Am I right? I still cannot draw the final conclusion by myself. Hope to discuss with you.
The pis for each z dimension need to be independent I believe, hence the inclusion of Z_DIM in the final dimension
|
2025-04-01T06:36:42.473737
| 2023-05-29T19:26:32
|
1731160159
|
{
"authors": [
"FerPerales",
"sarahraqueld"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:197",
"repo": "ApprenticeshipStandardsDotOrg/ApprenticeshipStandardsDotOrg",
"url": "https://github.com/ApprenticeshipStandardsDotOrg/ApprenticeshipStandardsDotOrg/pull/253"
}
|
gharchive/pull-request
|
Setup Flipper
Asana ticket: None
Now that the site is live, we want to be careful rolling out new features.
Introduce Flipper, an open source library that provides simple feature flags for Ruby.
I'd like to use it on next PRs for this task: https://app.asana.com/0/1203289004376659/1203289004376689/f
@sarahraqueld Besides linter issues, looks good!
|
2025-04-01T06:36:42.518084
| 2022-08-04T08:32:30
|
1328263649
|
{
"authors": [
"MateiNenciu",
"pazlavi",
"sawaca96",
"undergroundcreative"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:198",
"repo": "AppsFlyerSDK/appsflyer-capacitor-plugin",
"url": "https://github.com/AppsFlyerSDK/appsflyer-capacitor-plugin/pull/30"
}
|
gharchive/pull-request
|
Migrate to capacitor v4
Upgraded plugin to use latest capacitor v4.0.1 core and bumped minimum sdk versions to match the new specs.
Hi, how can I use this in my project? I've just upgraded to Capacitor 4 but the main appsflyer-capacitor-plugin plugin is failing. I'd like to use this pull which I've installed using:
npm i --save-dev AppsFlyerSDK/appsflyer-capacitor-plugin#pull/30/head --force
It adds it in the devDependencies section of package.json so when I do npm run build, I'm getting:
Cannot find module 'appsflyer-capacitor-plugin' or its corresponding type declarations.
I get the same if I uninstall and use:
npm i --save AppsFlyerSDK/appsflyer-capacitor-plugin#pull/30/head --force
Please help! Stuck!
Thanks :)
I've also installed with:
npm i appsflyer-capacitor-plugin@github:Ejobs/appsflyer-capacitor-plugin
But, again, when I do npm run build I'm getting:
Cannot find module 'appsflyer-capacitor-plugin' or its corresponding type declarations.
Just to clarify, the files are downloaded in the node_modules folder.
Here's my app's package.json:
{
"name": "QuizSwipe",
"private": true,
"version": "1.0.18",
"description": "QuizSwipe",
"license": "MIT",
"files": [
"dist/"
],
"scripts": {
"build": "stencil build",
"start": "stencil build --dev --watch --serve",
"test": "stencil test --spec --e2e",
"test.watch": "stencil test --spec --e2e --watch",
"generate": "stencil generate",
"clean": "npx rimraf www"
},
"devDependencies": {
"@capacitor/cli": "^4.0.0",
"@ionic/core": "^5.0.7",
"@stencil/core": "2.10.0",
"@stencil/store": "^1.3.0",
"@types/node": "^18.11.2"
},
"dependencies": {
"@awesome-cordova-plugins/in-app-purchase-2": "^5.45.0",
"@capacitor-community/admob": "^3.3.0",
"@capacitor-community/firebase-analytics": "^1.0.1",
"@capacitor-community/native-audio": "^4.0.0-0",
"@capacitor/android": "^4.0.0",
"@capacitor/app": "^4.0.0",
"@capacitor/app-launcher": "^4.0.0",
"@capacitor/browser": "^4.0.0",
"@capacitor/clipboard": "^4.0.0",
"@capacitor/core": "^4.0.0",
"@capacitor/device": "^4.0.0",
"@capacitor/ios": "^4.0.0",
"@capacitor/local-notifications": "^4.0.0",
"@capacitor/network": "^4.0.0",
"@capacitor/preferences": "^4.0.1",
"@capacitor/push-notifications": "^4.0.0",
"@capacitor/share": "^4.0.0",
"@capacitor/splash-screen": "^4.0.0",
"@ionic-native/core": "^5.26.0",
"@ionic-native/screenshot": "^5.31.1",
"@sparkfabrik/capacitor-plugin-idfa": "github:AE1NS/capacitor-plugin-idfa",
"appsflyer-capacitor-plugin": "github:Ejobs/appsflyer-capacitor-plugin",
"capacitor-rate-app": "^3.0.0",
"com.darktalker.cordova.screenshot": "^0.1.6",
"cordova-plugin-device": "^2.0.3",
"cordova-plugin-purchase": "^11.0.0",
"cordova-support-android-plugin": "^1.0.2",
"cordova-support-google-services": "^1.4.0",
"es6-promise-plugin": "^4.2.2",
"jetifier": "^1.6.6",
"ts-md5": "^1.2.11",
"tslib": "^1.11.1",
"web-social-share": "^6.4.1"
},
"postinstall": "jetifier"
}
I've been struggling with this for days now, could someone help me out?
Thanks,
Sean
@MateiNenciu @sawaca96
How can I install this to use in my project?
See my package.json above.
There is no dist folder and I'm getting:
"Cannot find module 'appsflyer-capacitor-plugin' or its corresponding type declarations." when I do npm run build
Help! :)
I'm still use capacitor v3
@pazlavi Any update on this? Looks like it's not been released yet.
Why modifying / accepting an already opened MR from August, when you can have your clients waiting till December to do it yourself ^
Sorry for late reply, you can either build that project and generate the dist folder, or you could do something like
import {AppsFlyer} from 'appsflyer-capacitor-plugin/src' like I did, till the maintainer decides to care about users that are upgrading.
Sorry for late reply, you can either build that project and generate the dist folder, or you could do something like import {AppsFlyer} from 'appsflyer-capacitor-plugin/src' like I did, till the maintainer decides to care about users that are upgrading.
Thanks for the info.
I tried importing from the src folder as you suggested with the plugin installed as per my package.json as above but I was getting build errors:
Error: Unexpected token (Note that you need plugins to import files that are not JavaScript)
So I uninstalled the AppsFlyer plugin and checked your PR out as a submodule. Installed and built to get the dist folder, then installed from my local build. Same error:
[ ERROR ] Rollup: Parse Error: ./submodules/appsflyer-capacitor-plugin/src/index.ts:3:12
Unexpected token (Note that you need plugins to import files that
are not JavaScript)
L3: import type { AppsFlyerPlugin } from './definitions';
L4: const AppsFlyer = registerPlugin<AppsFlyerPlugin>('AppsFlyerPlugin', {
[11:17.2] build failed in 11.08 s
What am I doing wrong? ;)
Here's my latest package.json:
{
"name": "QuizSwipe",
"private": true,
"version": "1.0.18",
"description": "QuizSwipe",
"license": "MIT",
"files": [
"dist/"
],
"scripts": {
"build": "stencil build",
"start": "stencil build --dev --watch --serve",
"test": "stencil test --spec --e2e",
"test.watch": "stencil test --spec --e2e --watch",
"generate": "stencil generate",
"clean": "npx rimraf www"
},
"devDependencies": {
"@capacitor/cli": "^4.0.0",
"@ionic/core": "^5.0.7",
"@stencil/core": "2.10.0",
"@stencil/store": "^1.3.0",
"@types/node": "^18.11.2"
},
"dependencies": {
"@awesome-cordova-plugins/in-app-purchase-2": "^5.45.0",
"@capacitor-community/admob": "4.0.0",
"@capacitor-community/firebase-analytics": "^1.0.1",
"@capacitor-community/native-audio": "^4.0.0-0",
"@capacitor/android": "^4.0.0",
"@capacitor/app": "^4.0.0",
"@capacitor/app-launcher": "^4.0.0",
"@capacitor/browser": "^4.0.0",
"@capacitor/clipboard": "^4.0.0",
"@capacitor/core": "^4.0.0",
"@capacitor/device": "^4.0.0",
"@capacitor/ios": "^4.0.0",
"@capacitor/local-notifications": "^4.0.0",
"@capacitor/network": "^4.0.0",
"@capacitor/preferences": "^4.0.1",
"@capacitor/push-notifications": "^4.0.0",
"@capacitor/share": "^4.0.0",
"@capacitor/splash-screen": "^4.0.0",
"@ionic-native/core": "^5.26.0",
"@ionic-native/screenshot": "^5.31.1",
"@sparkfabrik/capacitor-plugin-idfa": "github:AE1NS/capacitor-plugin-idfa",
"appsflyer-capacitor-plugin": "file:submodules/appsflyer-capacitor-plugin",
"capacitor-plugin-android-post-notifications-permission": "file:submodules/capacitor-plugin-android-post-notifications-permission",
"capacitor-rate-app": "^3.0.0",
"com.darktalker.cordova.screenshot": "^0.1.6",
"cordova-plugin-device": "^2.0.3",
"cordova-plugin-purchase": "^11.0.0",
"cordova-support-android-plugin": "^1.0.2",
"cordova-support-google-services": "^1.4.0",
"es6-promise-plugin": "^4.2.2",
"jetifier": "^1.6.6",
"ts-md5": "^1.2.11",
"tslib": "^1.11.1",
"web-social-share": "^6.4.1"
},
"postinstall": "jetifier"
}
We just released v6.9.2, which supports Capacitor v4.
If you still wish to use Capacitor v3, please check this page .
Thanks for the update :)
I installed it and sync-ed but now my app won't build :(
I'm getting:
[ ERROR ] Rollup: Parse Error: ./node_modules/appsflyer-capacitor-plugin/src/index.ts:3:12
Unexpected token (Note that you need plugins to import files that are not JavaScript)
L3: import type { AppsFlyerPlugin } from './definitions';
L4: const AppsFlyer = registerPlugin<AppsFlyerPlugin>('AppsFlyerPlugin', {
[03:04.1] build failed in 12.42 s
Here's my package.json:
{
"name": "QuizSwipe",
"private": true,
"version": "1.0.18",
"description": "QuizSwipe",
"license": "MIT",
"files": [
"dist/"
],
"scripts": {
"build": "stencil build",
"start": "stencil build --dev --watch --serve",
"test": "stencil test --spec --e2e",
"test.watch": "stencil test --spec --e2e --watch",
"generate": "stencil generate",
"clean": "npx rimraf www"
},
"devDependencies": {
"@capacitor/cli": "^4.0.0",
"@ionic/core": "^5.0.7",
"@stencil/core": "2.10.0",
"@stencil/store": "^1.3.0",
"@types/node": "^18.11.2"
},
"dependencies": {
"@awesome-cordova-plugins/in-app-purchase-2": "^5.45.0",
"@capacitor-community/admob": "4.0.0",
"@capacitor-community/firebase-analytics": "^1.0.1",
"@capacitor-community/native-audio": "^4.0.0-0",
"@capacitor/android": "^4.0.0",
"@capacitor/app": "^4.0.0",
"@capacitor/app-launcher": "^4.0.0",
"@capacitor/browser": "^4.0.0",
"@capacitor/clipboard": "^4.0.0",
"@capacitor/core": "^4.0.0",
"@capacitor/device": "^4.0.0",
"@capacitor/ios": "^4.0.0",
"@capacitor/local-notifications": "^4.0.0",
"@capacitor/network": "^4.0.0",
"@capacitor/preferences": "^4.0.1",
"@capacitor/push-notifications": "^4.0.0",
"@capacitor/share": "^4.0.0",
"@capacitor/splash-screen": "^4.0.0",
"@ionic-native/core": "^5.26.0",
"@ionic-native/screenshot": "^5.31.1",
"@sparkfabrik/capacitor-plugin-idfa": "github:AE1NS/capacitor-plugin-idfa",
"appsflyer-capacitor-plugin": "^6.9.2",
"capacitor-plugin-android-post-notifications-permission": "file:submodules/capacitor-plugin-android-post-notifications-permission",
"capacitor-rate-app": "^3.0.0",
"com.darktalker.cordova.screenshot": "^0.1.6",
"cordova-plugin-device": "^2.0.3",
"cordova-plugin-purchase": "^11.0.0",
"cordova-support-android-plugin": "^1.0.2",
"cordova-support-google-services": "^1.4.0",
"es6-promise-plugin": "^4.2.2",
"jetifier": "^1.6.6",
"ts-md5": "^1.2.11",
"tslib": "^1.11.1",
"web-social-share": "^6.4.1"
},
"postinstall": "jetifier"
}
To get my version of TypeScript I did:
$ npx tsc -v
Which gives:
Version 4.1.6
Ionic:
Ionic CLI : 6.19.0 (/Users/seanwilson/.nvm/versions/node/v15.1.0/lib/node_modules/@ionic/cli)
Capacitor:
Capacitor CLI : 4.4.0
@capacitor/android : 4.4.0
@capacitor/core : 4.4.0
@capacitor/ios : 4.4.0
Utility:
cordova-res (update available: 0.15.4) : 0.15.3
native-run : 1.7.1
System:
NodeJS : v15.1.0 (/Users/seanwilson/.nvm/versions/node/v15.1.0/bin/node)
npm : 7.5.2
OS : macOS Catalina
|
2025-04-01T06:36:42.548511
| 2024-10-13T21:22:13
|
2584279805
|
{
"authors": [
"April-Gras",
"Panthon13",
"thompa2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:199",
"repo": "April-Gras/obsidian-auto-timelines",
"url": "https://github.com/April-Gras/obsidian-auto-timelines/issues/190"
}
|
gharchive/issue
|
Colon space (: ) inside title or event body breaks entire plugin
Describe the bug
If you type a colon followed by a space in any event variable, the entire plugin breaks and ceases to display any timeline.
To Reproduce
This works:
%%aat-inline-event
aat-event-start-date: 54
aat-event-end-date: true
aat-render-enabled: true
aat-event-title: This Works.
timelines: [inline-events]
%%
This breaks the plugin:
%%aat-inline-event
aat-event-start-date: 54
aat-event-end-date: true
aat-render-enabled: true
aat-event-title: This: Does Not.
timelines: [inline-events]
%%
If you have a colon followed by a space in aat-event-title, or aat-event-body, the whole plugin ceases to display any timelines, not just the timeline with the offending character combination.
I don't know if this is something that can be accounted for, but sometimes it's nice to have a colon in a note title. This may be a limitation of how things get parsed in Obsidian, but I wanted to point it out. Took me hours to figure out why my timelines completely disappeared on me! Lol.
Did you try:
%%aat-inline-event
aat-event-start-date: 54
aat-event-end-date: true
aat-render-enabled: true
aat-event-title: "This: Does Not."
timelines: [inline-events]
%%
Did you try:
%%aat-inline-event
aat-event-start-date: 54
aat-event-end-date: true
aat-render-enabled: true
aat-event-title: "This: Does Not."
timelines: [inline-events]
%%
Not OP, but I just tested your solution and still get the broken timeline. It seems to be specifically when the colon has a space immediately after it. Putting a colon anywhere else, including with a space before it, works fine.
Hi again, @Panthon13 @thompa2
The wrapping with " seems to work on my end, :thinking: feel free to provide a reproduction vault and post it in .zip here
https://github.com/user-attachments/assets/07f296c9-4b65-40da-81fb-d4651f4edba1
|
2025-04-01T06:36:42.575437
| 2023-07-17T15:14:45
|
1808011659
|
{
"authors": [
"153957",
"aMahanna"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:200",
"repo": "ArangoDB-Community/python-arango",
"url": "https://github.com/ArangoDB-Community/python-arango/issues/261"
}
|
gharchive/issue
|
Unable to disable request timeout
Previously it was possible to disable the timeout for the HTTP client by setting request_timeout to None, but after: https://github.com/ArangoDB-Community/python-arango/pull/257 it 'has' to be an int of float.
So the timeout can no longer be easily disabled (can of course be set super high, but that means having to set it to something instead of clearly nothing/unlimited).
Please allow None as a value for the request_timeout.
Hey @153957, #265 should fix the mypy issue you are seeing
|
2025-04-01T06:36:42.576946
| 2022-12-05T16:04:34
|
1476886986
|
{
"authors": [
"Aratramba",
"benmawla"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:201",
"repo": "Aratramba/sanity-plugin-media-library",
"url": "https://github.com/Aratramba/sanity-plugin-media-library/pull/43"
}
|
gharchive/pull-request
|
Add the ability to search by most used asset
I think this would be a beneficial addition.
On some sites I re-use the same asset multiple times. This filter provides quick access to the most commonly used assets (see screenshot for example):
Looks great, thanks! ✨
Released in 2.1.2
|
2025-04-01T06:36:42.580257
| 2023-09-19T14:47:46
|
1903162469
|
{
"authors": [
"Arcader717"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:202",
"repo": "Arcader717/Async-DisOAuth2",
"url": "https://github.com/Arcader717/Async-DisOAuth2/issues/2"
}
|
gharchive/issue
|
[General] Rebuild the async version
Are you in the correct repository?
[x] Yes
Request
Rebuild the async version, as there are too many errors that I found during testing. I will spend a few days changing the code and fixing it.
I realized that the rebuilding didn't take as long as I expected, so I changed the target date to today
Nevermind, it still needs to be rebuilt
I feel really dumb, I forgot to use await in my testing program
|
2025-04-01T06:36:42.588669
| 2024-12-23T19:35:52
|
2756619139
|
{
"authors": [
"AdmilZhao",
"ArchLeaders"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:203",
"repo": "ArchLeaders/HavokActorTool",
"url": "https://github.com/ArchLeaders/HavokActorTool/issues/3"
}
|
gharchive/issue
|
System.IO.DirectoryNotFoundException: Could not find a part of the path 'C:\Actor\Pack\
Whenever I try to use the tool it gives me this error. I don't know what I am doing wrong
Please help
Make sure you have set your game path in the settings.
Now it's giving me this error
Sorry for the trouble
Ah, now that may be an error on my end, I'll look into it.
Okay thank you
Ah, I see the issue, change the BaseActorName to nothing. That field is for a vanilla actor name to use when building the custom one.
Hey I'm really sorry to keep bothering you but that didn't seem to work. I'm still getting the same error but now it's just saying tis:
I'm really sorry for the trouble
I must have made a mistake when checking the field. Please try restarting the app and avoid filling out that field.
Tried that and it just keeps defaulting to "TwnObj_HunterHouseBed_A_01" whenever I run it for some reason
I am very sorry for the trouble
That is expected behaviour, however, if you update to 2.0.1 you can clear it without an issue.
It's still defaulting but now it's saying this
"System.IO.FileNotFoundException: The actor 'TwnObj_HunterHouseBed_A_01' is not a vanilla actor. Please change the 'Base Actor' field."
Even when I clear it and run it again it just defaults back again
It should default to that because it automatically looks for the best match. However, that actor does exist, so the issue is with your game dump. Possibly the path is wrong, or else the dump is incomplete.
I have the actor in my game directory. What should the path look like so I can verify if I have it right?
Sorry again for the trouble
I have the actor in my game directory. What should the path look like so I can verify if I have it right?
Sorry again for the trouble
It depends on where it's stored on your computer, but it should end in content (or the folder containing Actor, but not the Actor folder itself).
There we go! Finally! I had the game directory folder wrong I think.
Thank you very much for the help. Sorry it ended up being so complicated
No worries, I'm glad you got it sorted.
|
2025-04-01T06:36:42.615660
| 2017-06-27T03:25:17
|
238727545
|
{
"authors": [
"CB49",
"chopraaa",
"donguyenha",
"threathgfx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:204",
"repo": "ArdiArtani/Google-Drive-Player-Script",
"url": "https://github.com/ArdiArtani/Google-Drive-Player-Script/issues/17"
}
|
gharchive/issue
|
it does not generate the redirector.googlevideo.com..
i check the cached files in cache folder and the link is
https://doc-08-7c-docs.googleusercontent.com/docs/securesc/ha0ro937gcuc7l7deffksulhg5h7mbp1/kkkkq0kcmslfojlfrlojiq058bpmf1e1/1498528800000/1340883420501142032
it is not redirector.googlevideo.com/videoplayback
how to fix it
@donguyenha, it really doesnt generate redirector.googlevideo.com. It only focuses to the direct download using uc=
@threathgfx how to generate the link with redirector.googlevideo,com,
ex:
https://redirector.googlevideo.com/videoplayback?id=49ce579a8ac5005b&itag=22&source=webdrive&requiressl=yes&ttl=transient&pl=33&ei=OnNTWeaMJpPIqAWgx5WoBg&mime=video/mp4&lmt=1498589208203180&ip=2600:3c01::f03c:91ff:fe60:c14e&ipbits=0&expire=1498655610&sparams=api,ei,expire,id,ip,ipbits,itag,lmt,mime,mm,mn,ms,mv,pl,requiressl,source,ttl&signature=0E1E8B5519026E7AC7BDCB89C55F2A305B6D50B7.7657FCDC57659A4F06FB0C9A03E502539C79074C&api=B4D9C084A100269C74C2292DDF010&cms_redirect=yes&mm=31&mn=sn-n4v7sn7y&ms=au&mt=1498641098&mv=m&key=cms1&app=storage
That's currently being discussed as we don't know yet. Nobody has made it public if they do know.
except they have. refer to youtube-dl. its public
|
2025-04-01T06:36:42.636660
| 2018-03-29T13:49:11
|
309762006
|
{
"authors": [
"ArduCAM",
"coldstart01"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:205",
"repo": "ArduCAM/Arduino",
"url": "https://github.com/ArduCAM/Arduino/issues/313"
}
|
gharchive/issue
|
opencv files, version question
Hi,
Does the example that you sent me last time (Linux) require a client-side opencv installation, or are all the necessary opencv files included? I am getting the following error when executing:
./Test: error while loading shared libraries: libopencv_core.so.2.4: cannot open shared object file: No such file or directory
If are to install OpenCV do we do that with 3.0+ or match the 2.4, as above.
you need to install the opencv binary library by sudo apt-get install libopencv-dev
I did that and manually checked /usr/lib and /usr/local/lib and no libopencv_core.so.2.4.
I did system search as well for libopencv_core.so.2.4 as well, with negative results.
The system did install
a/usr/lib/x86_64-linux-gnu/libopencv_core.so.3.1.0
/usr/lib/x86_64-linux-gnu/libopencv_core.so
/usr/lib/x86_64-linux-gnu/libopencv_core.so.3.1
What next?
Should I uninstall the 3.1.0 and replace with the older distro opencv Unix 2.4?
I really could use the help here, I really want to move forward here.
I think you need to recompile the source code to run the ./Test example. If you link the source code with opencv 3, it won't require 2.4 version.
Awesome, that makes sense.
What compiler/linker did you use? I plan o using LinuxGCC toolchain unless you think it better to use something else.
There is comments on top of the cpp file about how to compile the code, something like this: g++ MT9V022_demo.cpp -o Test pkg-config --cflags --libs opencv -lArduCamLib -lpthread -lusb-1.0 -L. -I. -std=gnu++11
So you need g++ to compile the code.
Things going ok recompiling source code. It appears opencv library files were not installed wioth the orignal sudo command:
you need to install the opencv binary library by sudo apt-get install libopencv-dev
I can link to header files but not link to library files - they don't exist. I' ve done a thorough search--and manually checked /usr
Should I reinstall opencv?
The documentation is quite good, thanks for checking anyway. My only problem rightn ow is my opencv libraries have literally disappeared. If I can find them I can move on.
Hi Lee, me again. A lot of progress. Can you please tell me what to fix ?
directory>./Test
`Device found = 1
ArduCam_open successful
create capture thread successfully
create display thread successfully
frame available
No protocol specified
Unable to init server: Could not connect: Connection refused
(Display Image:1991): Gtk-WARNING **: cannot open display: :0
`
I have never seen "No protocol specified" and "Unable to init server: Could not connect: Connection refused" errors, did you modified our example ?
|
2025-04-01T06:36:42.725265
| 2020-03-17T21:24:55
|
583308285
|
{
"authors": [
"kibaekkim",
"yim0331"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:206",
"repo": "Argonne-National-Laboratory/MaximinOPF.jl",
"url": "https://github.com/Argonne-National-Laboratory/MaximinOPF.jl/issues/22"
}
|
gharchive/issue
|
Publishing the package
We need to publish this package as a public Julia package to finalize this development cycle.
Done!
|
2025-04-01T06:36:42.771637
| 2018-02-05T11:43:37
|
294364109
|
{
"authors": [
"fix",
"j-a-m-l"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:207",
"repo": "ArkEcosystem/ark-desktop",
"url": "https://github.com/ArkEcosystem/ark-desktop/pull/555"
}
|
gharchive/pull-request
|
Add a message on the sending dialog when using a Ledger
To avoid problems such as https://cointelegraph.com/news/newly-discovered-vulnerability-in-all-ledger-hardware-wallets-puts-user-funds-at-risk
there is another issue, is that the receiving address might be redacted. So we need to display on device the receiving address to be sure it is not redacted.
If i remember well, the ledger app already allows this, you need to put some switch in the request to display the requested address on device.
https://github.com/ArkEcosystem/ark-desktop/blob/master/LedgerArk.js#L11
@alexbarnsley might need to confirm with ledger on their slack
see here https://github.com/ArkEcosystem/ark-ledger/blob/master/src/main.c#L752
|
2025-04-01T06:36:42.788582
| 2023-09-13T20:01:58
|
1895183391
|
{
"authors": [
"NULLlnull",
"ZhaoZuohong"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:208",
"repo": "ArkMowers/arknights-mower",
"url": "https://github.com/ArkMowers/arknights-mower/issues/284"
}
|
gharchive/issue
|
软件打开报错
Traceback (most recent call last):
File "webview_ui.py", line 4, in
File "", line 991, in _find_and_load
File "", line 975, in _find_and_load_unlocked
File "", line 671, in _load_unlocked
File "PyInstaller\loader\pyimod02_importers.py", line 385, in exec_module
File "server.py", line 3, in
File "", line 991, in _find_and_load
File "", line 975, in _find_and_load_unlocked
File "", line 671, in load_unlocked
File "PyInstaller\loader\pyimod02_importers.py", line 385, in exec_module
File "arknights_mower_init.py", line 9, in
File "pathlib.py", line 1181, in resolve
File "pathlib.py", line 206, in resolve
OSError: [WinError 1] 函数不正确。: 'R:\Temp\_MEI60362\arknights_mower\init'
我使用了内存盘。系统的缓存地址是指向内存盘R盘的
你运行的是哪个版本
|
2025-04-01T06:36:42.790448
| 2024-08-07T05:47:59
|
2452527042
|
{
"authors": [
"Arka-Bhowmik",
"MYZyuzhoumu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:209",
"repo": "Arka-Bhowmik/mri_triage_normal",
"url": "https://github.com/Arka-Bhowmik/mri_triage_normal/issues/1"
}
|
gharchive/issue
|
The link for sharing the network weights is invalid. Could the author please share it again?
The link for sharing the network weights is invalid. Could the author please share it again?
The link for sharing the network weights is invalid. Could the author please share it again?
Hi myzyuzhoumu,
The google drive link for network weights is correct. I am able to download the weights without issue. Here, are the weights.
Can you share the error message or screeshot?
https://github.com/Arka-Bhowmik/mri_triage_normal/tree/main/output
|
2025-04-01T06:36:42.800606
| 2018-02-13T13:13:01
|
296732004
|
{
"authors": [
"Arlen22",
"garu57"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:210",
"repo": "Arlen22/TiddlyServer",
"url": "https://github.com/Arlen22/TiddlyServer/issues/35"
}
|
gharchive/issue
|
Make backupDirectory unique for each served single file TiddlyWiki
Given that i'm completely new both to node.js and TiddlyWiki pls forgive me if i'm telling some oddity.
I'm still experimenting to see what kind of beast i've got (syntax, macros, widgets and mostly its philosophy that i don't yet completely grasp) and i found really useful making one or more copy of a master wiki file and apply different changes to the different copies just to see the differences, then apply what i like most to the master copy.
Then i delete the copies and make them new from the master with the same names (so i don't have to change settings.json) and restart the cycle.
The problem is that when i delete a copy, the unique backupDirectory remain polluted with backups that are no longer usable and must be deleted manually.
Would it be possible to have the backupDirectory relative to the tree so that each single file instance can have its own?
This would help at least in two ways, one, with this usage pattern, i could delete with a single action both the instance and the backups, second having the each backupDirectory together with its file instance will allow to backup easily them together.
Thanks,
Gabriele
It is possible that in the future I could allow a backup folder to be
specified for each tree item. I will keep that in mind. Currently it is not
possible.
I would like to mention that you can specify a folder in the tree instead
of just a file, as it seems you have done, which then allows you to serve
anything in that folder.
Thanks, i tried using folder but i see there the behavior is completely different and as far as i understand backupDirectory is not applicable there.
I plan to use TiddlyWiki as a companion notebook of Jabref and Zotero for literary research and it will end containing a huge amount of data scattered on many different wikis, so before filling it with real data i want to be sure to have explored every single facet of it's behavior.
Now i'm stressing single file wikis, then will do the same on folder wikis. I need to be sure that's rock solid and i don't risk to loose data in any circumstances.
Anyhow thanks for your attention,
Gabriele
Actually, I'm referring to folders containing multiple single file wikis. If you specify a folder path instead of a file path in settings.json it will serve all the single file wikis AND data folders found in that folder. So you can have multiple single file wikis in the folder and just specify the whole folder instead of each individual wiki. You can also have multiple data folders (any folder containing a tiddlywiki.info file) in the folder. Of course, you can have both in the same folder.
Wow! That's powerful! I didn't got it.
As i said in my first post, i don't yet really grasp the "philosophy" behind this object. All the wikis i tried are php and a lot more rigid in structuring data, but this, as Ruston said, is something completely different. What really fascinated me from the beginning is the apparent simplicity by which once you have your data, you can build multiple "knowledge" paths through them and i'm looking for the magic "click" on this.
Anyhow, in the meanwhile, i solved another small problem. I don't like products that mix installation files with runtime byproducts, but fortunately with a rapid look at the code i saw that settings.json can be passed as an argument to the server, so now i have a wiki root dir with settings.json using only relative paths, backupDirectory and all the wiki folders and now i can move the server installation to a read only zone.
Thanks,
Gabriele
A backup folder may now be specified for each tree item in version 2.1. The documentation I am writing will include instructions for this.
|
2025-04-01T06:36:42.837321
| 2017-02-23T19:53:52
|
209865910
|
{
"authors": [
"Artanicus"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:212",
"repo": "Artanicus/python-cozify",
"url": "https://github.com/Artanicus/python-cozify/issues/2"
}
|
gharchive/issue
|
Multisensor may not work with all sensors
Multisensor has only been tested on Proove sensors, needs to be validated by someone else with sensors from other manufacturers.
v0.2.10 brings a new devices interface, hub.devices() that supports filtering by capabilities such as TEMPERATURE. It's still crude and doesn't support really the way this was built at the start so this issue may become moot as we move towards a more granular way of operating on a set of devices. So in fact the whole legacy multisensor module may go away in favor of more generalized tools.
Also, got myself some Develco moisture sensors that also do temperature sensing so now I can actually test pulling data by capabilities across different manufacturers.
Multisensor is being deprecated with v0.3 in favor of capability based device access. The capabilities are already there (and improved in current devel branch) so nothing stopping from using them already now.
|
2025-04-01T06:36:42.846266
| 2021-02-01T02:18:33
|
797891681
|
{
"authors": [
"ArtexJay",
"Jexle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:213",
"repo": "ArtexJay/Obsidian-CyberGlow",
"url": "https://github.com/ArtexJay/Obsidian-CyberGlow/issues/2"
}
|
gharchive/issue
|
Add .signature class to the <cite> element
While the signature class works on <cite> elements as far as applying basic text styles, it doesn't render on the right like it does with <div>s. Using <cite> for quotes seems to fall more in-line with how many in the Obsidian community appear to use them, and at the moment, their positioning is a bit intrusive for the writing it's associated with.
Expected behavior:
Actual behavior:
I haven't tried anything yet due to a lack of time, though I think setting display: block; should allow for it to be positioned more easily.
I didn't even know about the option. I'll play around with it.
Oh, the "incorrect" font isn't because of issues with the theme; the screenshots were just to show the positioning. I changed the font using a second class with .signature; if I take it off, it defaults back to Edwardian.
Will be coming in v6, already have it "" working on my working build
|
2025-04-01T06:36:42.851475
| 2023-08-04T12:43:14
|
1836680853
|
{
"authors": [
"Arthur151",
"Dipankar1997161"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:214",
"repo": "Arthur151/ROMP",
"url": "https://github.com/Arthur151/ROMP/issues/477"
}
|
gharchive/issue
|
Quick clarification of projected 2d points of smpl
hello @Arthur151, I had a quick query to be clarified.
Now that we have the 3d smpl joints, I want to project it on the resp. 2d image.
I am aware that we have pj2d, but I wish to do the conversion myself.
So, I checked the repo and saw a perspective projective function
Q1. is this the function used for the conversion of 3d smpl joints to 2d ?? or there's something else.
def perspective_projection_ROMP(points, rotation, translation, focal_length,
camera_center):
"""
This function computes the perspective projection of a set of points.
Input:
points (bs, N, 3): 3D points
rotation (bs, 3, 3): Camera rotation
translation (bs, 3): Camera translation
focal_length (bs,) or scalar: Focal length
camera_center (bs, 2): Camera center
"""
batch_size = points.shape[0]
K = torch.zeros([batch_size, 3, 3], device=points.device)
K[:, 0, 0] = focal_length
K[:, 1, 1] = focal_length
K[:, 2, 2] = 1.0
K[:, :-1, -1] = camera_center
# Transform points
points = torch.einsum("bij,bkj->bki", rotation, points)
points = points + translation.unsqueeze(1)
# Apply perspective distortion
projected_points = points / points[:, :, -1].unsqueeze(-1)
# Apply camera intrinsics
projected_points = torch.einsum("bij,bkj->bki", K, projected_points)
return projected_points[:, :, :-1]
Q2. We only require the 'smpl_joints' for such plotting right? or do we need the 'smpl_poses'?
Q3. For plotting the skeleton, can I use the following function after I have generated the 2d smpl
https://github.com/Arthur151/ROMP/blob/5cf8068297e8700701748c58d98428d8b6bcea91/trace/lib/utils/vis_utils.py#L246
and this is the skeleton_tree for smpl I suppose
smpl24_connMat = np.array([0,1, 0,2, 0,3, 1,4,4,7,7,10, 2,5,5,8,8,11, 3,6,6,9,9,12,12,15, 12,13,13,16,16,18,18,20,20,22, 12,14,14,17,17,19,19,21,21,23]).reshape(-1, 2)
Q4. Lastly, for plotting the 2d joints on the image(not the skeleton), do you recommend any function in ROMP to check or can we just use matplotlib scatterplot to plot the joints on the IMAGES.
I really need to know these. Thank you @Arthur151
@Dipankar1997161
Sorry for the late reply!
Q1. Yes, this function is written to perform the 2D projection of SMPL's 3D joints.
Q2. Actually, it just implements a normal perspective projection, so the 3D joints are required. SMPL theta poses are not needed.
Q3. Yes, this could be very convenient.
Q4. Any functions you like to plot the 2D joints should be fine.
|
2025-04-01T06:36:42.853079
| 2017-09-15T07:44:12
|
257953896
|
{
"authors": [
"PBascones",
"Smurf-IV",
"mikeobrien",
"sadiqkhoja",
"vishalmane"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:215",
"repo": "ArthurHub/HTML-Renderer",
"url": "https://github.com/ArthurHub/HTML-Renderer/issues/103"
}
|
gharchive/issue
|
Does not work!
Colouring a Row in a table does not work.
Same here...
Hi any fix for this?
Same here, it seems that this library is no longer maintained :/
background-color doesn't work on td either
|
2025-04-01T06:36:42.874262
| 2023-03-29T16:30:26
|
1646150094
|
{
"authors": [
"amkhlv",
"genusistimelord"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:217",
"repo": "AscendingCreations/AxumCSRF",
"url": "https://github.com/AscendingCreations/AxumCSRF/issues/7"
}
|
gharchive/issue
|
example does not compile
error[E0599]: no method named into_make_service found for struct Router<CsrfConfig, _> in the current scope
ahh Ill update the Example to the newest Axum.
It seems that into_make_service() is only defined on Router<(), _> while here we have Router<CsrfConfig, _>
hmm it should just take CSRFConfig just fine. I am wondering if the error is from somewhere else. as sometimes Axum errors dont always Tell you what the real issue is.
You can try to make a State Struct instead like
use axum::extract::FromRef;
#[derive(Clone)]
pub struct SystemState {
pub odbc: axum_odbc::ODBCConnectionManager,
pub flash_config: axum_flash::Config,
pub csrf: axum_csrf::CsrfConfig,
}
impl SystemState {
pub fn new(
odbc: axum_odbc::ODBCConnectionManager,
flash_config: axum_flash::Config,
csrf: axum_csrf::CsrfConfig,
) -> Self {
Self {
odbc,
flash_config,
csrf,
}
}
}
impl FromRef<SystemState> for axum_csrf::CsrfConfig {
fn from_ref(input: &SystemState) -> Self {
input.csrf.clone()
}
}
It seems that into_make_service() is only defined on Router<(), _> while here we have Router<CsrfConfig, _>
I got it. The thing is, .with_state(...) should go the last in the chains, after all .route(...)s . Precisely as in your example. I thought it would not matter... Now it all works. Thank you !
|
2025-04-01T06:36:43.422518
| 2016-09-12T07:15:59
|
176308682
|
{
"authors": [
"kevin-hirsch",
"mattem86"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:218",
"repo": "AssistoLab/KVNProgress",
"url": "https://github.com/AssistoLab/KVNProgress/issues/88"
}
|
gharchive/issue
|
even import only (from pods) crashes OpenGL view
I have a OpenGL ES 2.0 view, which I wanted to overlay with the KVNProgress. Every single time, I opened the GLView twice, the app crashes with gpus_ReturnGuiltyForHardwareRestart ... Even if only imported the KVNProgress module via CocoaPods and no showing of the KVNProgress itself...
Hello @mattem86, thanks for using this library!
Are you sure this crash is from KVNProgress itself? Do the crash disappear when you remove the KVNProgress import?
Yep, the crash disappears, if I do not import KVNProgress within the ViewController which contains the GLView. I didn't believe that to, so I double-checked :) But I will try it once again tomorrow...
So the view of your view controller is a GLView?
Not exactly, the View in my View Controller has a View, which is holds an GLKView inside and is a GLKViewDelegate
Am 12. September 2016 um 16:30:49, Kevin Hirsch<EMAIL_ADDRESS>schrieb:
So the view of your view controller is a GLView?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
So if you remove the "import KVNProgress" statement, the app does not crash anymore?
Which version of CocoaPods are you using? (pod --version)
Sorry, I was busy the last week ... I use pod 1.0.0 ... but, something changed in the meantime, the app crashes no more when I import KVNProgress on that particular ViewController... maybe some other pod were interfering? I have to try again...!
Am 13. September 2016 um 16:22:19, Kevin Hirsch<EMAIL_ADDRESS>schrieb:
So if you remove the "import KVNProgress" statement, the app does not crash anymore?
Which version of CocoaPods are you using? (pod --version)
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
Ha! It is the Viewcontroller which I segue off from to reach the new Viewcontroller with the GLView. So, if I import (just import!) KVNProgress in my ViewController A and I segue off to my ViewController B (which contains my GLView), then the crash appears at the second time I segue. Weird.
Am 20. September 2016 um 11:37:08, Matthias Temmen<EMAIL_ADDRESS>schrieb:
Sorry, I was busy the last week ... I use pod 1.0.0 ... but, something changed in the meantime, the app crashes no more when I import KVNProgress on that particular ViewController... maybe some other pod were interfering? I have to try again...!
Am 13. September 2016 um 16:22:19, Kevin Hirsch<EMAIL_ADDRESS>schrieb:
So if you remove the "import KVNProgress" statement, the app does not crash anymore?
Which version of CocoaPods are you using? (pod --version)
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
Ok, seems to be a bug in the GLView Framework I was using ... I was trying again yesterday and everything went well, after updating the specific framework!
|
2025-04-01T06:36:43.424714
| 2018-10-16T07:43:14
|
370479418
|
{
"authors": [
"Astn",
"jdengitw"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:219",
"repo": "Astn/JSON-RPC.NET",
"url": "https://github.com/Astn/JSON-RPC.NET/issues/99"
}
|
gharchive/issue
|
Received parameter in client is not equal to that sent in server
If I send a server function continually, sometimes the client will lose one. Then when I send it one more time, the client will get the last losed one.
Is this a potential issue?
Any chance you could post some code showing what issue you are experiencing?
I think I made a mistake in my own code.
I added 2 server functions with the same name but different number of parameters, these 2 functions are triggered at the same time. I only handled one function in my client code. It works well after deleting one function in server.
|
2025-04-01T06:36:43.427705
| 2018-08-21T12:37:18
|
352519301
|
{
"authors": [
"OrangeFlag",
"gonchik"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:220",
"repo": "AstroMatt/atlassian-python-api",
"url": "https://github.com/AstroMatt/atlassian-python-api/pull/67"
}
|
gharchive/pull-request
|
Make more comfortable get_page_id behaviour
If get_pade_id exec with wrong id parameter, don't raise 'AttributeError: 'NoneType' object has no attribute 'get''
Thanks!
|
2025-04-01T06:36:43.434524
| 2021-02-02T15:23:18
|
799379441
|
{
"authors": [
"AstroViking",
"mayou36"
],
"license": "WTFPL",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:221",
"repo": "AstroViking/tf-kde",
"url": "https://github.com/AstroViking/tf-kde/issues/2"
}
|
gharchive/issue
|
Integration into zfit
I was thinking on what is possibly the best way of moving this inside of zfit so that it can be used there.
Since there are a few, I would suggest that maybe the pure zfit PDFs, we can directly move into zfit. While the TFP ones, we can ask whether this is actually of interest (but this may needs some more work on them, e.g. added tests etc.)
I would propose that either you or me is gonna copy them to zfit, add docs, tests. If you're short on time, I can do that, or parts, and you're very welcome to add things. If you're interested and want to contribute directly - it's your work and it should fit without a problem - please go ahead and add it yourself to zfit and make a PR, that is also very welcome!
Do you have a preference on how to proceed?
Or would you want to promote this to a real package? I think it would just need more polish, unittests, docs, then for sure...
Thanks for following up with this!
I do not have plans to make this into a real (and maintained) package currently.
There are actually four PDF classes of interest (all located in tf_kde/distribution/kernel_density_estimation_zfit.py).
KernelDensityEstimation, which is the only one that uses TFP Distributions and has some overlap with GaussianKDE1DimV1 in zfit, although it supports multiple kernels
KernelDensityEstimationFFT, which has some overlap with FFTConvPDFV1 in zfit
KernelDensityEstimationISJ, which is based on the Improved Sheather Jones algorithm
KernelDensityEstimationHofmeyr, which I think is not suitable to port to zfit right now, since its TensorFlow implementation is awfully slow compared to the other three methods and its CPP based (a custom TensorFlow Op) implementation is not stable and has to be compiled specifically for different architectures
I started to move the three suitable methods into zfit in the following pull request here: https://github.com/zfit/zfit/pull/285
Since I currently do not have the time to extend this with tests and docs all by myself, your help is very welcome. Also I am not entirely sure about the architecture of zfit and where to put things like utility classes, so please advise me, where I made a suboptimal choice.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.