added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:40:30.640007
| 2015-04-06T11:43:10
|
66588261
|
{
"authors": [
"voxpelli",
"zertosh"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11015",
"repo": "substack/node-detective",
"url": "https://github.com/substack/node-detective/pull/43"
}
|
gharchive/pull-request
|
Update Travis CI tests to use newer Node.js versions
As the ^1.4.1 syntax used for currently including the escodegen module was first introduced in npm 1.3, the npm 1.2 verison included with node 0.8.x fails at installing the dependency.
The escodegen module itself also in version 1.3.3 changed from requiring node >=0.4.0 to requiring node >=0.10.0, so it explicitly doesn't support 0.8.x anymore either.
The pull request to update the required acorn dependency ran into this issue and since that very same PR preferably should use the caret operator for the dependency it would be better to stop testing in node 0.8.x and start testing in newer node versions instead.
Related PR: https://github.com/substack/node-detective/pull/41
Thanks! Fixed by https://github.com/substack/node-detective/pull/46
|
2025-04-01T06:40:30.646806
| 2018-09-20T16:54:59
|
362276149
|
{
"authors": [
"michael",
"oliver----"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11016",
"repo": "substance/texture",
"url": "https://github.com/substance/texture/issues/912"
}
|
gharchive/issue
|
Find and Replace: Select All in one of the inputs should not select all in the document.
At the moment, keyboard events bubble up to the editor which handles certain combos, such as CtrlOrCommand-a, having the effect that Select All is applied to the area where the internal selection is.
We should either prevent all events from bubbling up, or consume certain combos.
Yeah I saw that too. I wonder if it would make sense to use Substance TextInput instead of the native input fields. Maybe then we have things better under control? (also enabling commands cmd+g etc.)
Command-G is more relevant as a short-cut for next, when the cursor is inside the doc.
Have you noticed, that you can travers with ENTER? Shift-Enter is not handled yet, though.
Don't think that Substance TextInput is solving this problem for us.
In this case it is really important to bind a keydown handler for CtrlOrCommand+a on the input. The SelectAllCommand is working only for Surfaces.
Maybe I didn't understand your suggestion... It would make sense to have a native input component, which is having all the common handlers. That's true.
If Substance TextInput is doing this already, I'd suggest to move it into the kit.
But don't use it for a model! :)
Surfaces are bound to models. So the inputs in the FaRD(ialog) ;) are not candidates for Surfaces.
The would need a detached (temporary model) of course. Anyways.. as said it was just a thought, since now this is the only place where we have <input> elements. :)
That is an overkill. We are perfectly able to create a native input with enough handlers attached to be compliant with the rest of the Substance actors.
Still it makes sense to have such a thing, so we do not need to reinvent it each time.
now this is the only place where we have elements.
The key difference is that FaRD is app chrome, not content.
Fair enough. Let's stick with what we have then.
Yeah. Still I am in favor of creating a Component around Input which behaves nicely. :)
Fixed via https://github.com/substance/texture/commit/64860304b46f5df232771efb81ddc059f934d298
|
2025-04-01T06:40:30.650771
| 2021-03-14T07:21:25
|
831086983
|
{
"authors": [
"koppen",
"timdiggins"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11017",
"repo": "substancelab/route_downcaser",
"url": "https://github.com/substancelab/route_downcaser/pull/47"
}
|
gharchive/pull-request
|
add rubocop to travis
Seems like a shame to have defined rubocop config (and implemented rubocop changes) and not enforce it.
I've pinned the version of rubocop to the latest version to prevent arbitrary correctness changes.
Thank you so much for contributing.
We run Rubocop via Code Climate, though, (see https://codeclimate.com/github/substancelab/route_downcaser/pull/47 for an example). I wonder if it is worth running it on Travis as well?
Ah ok, I didn't see that. But it would be helpful to make it a development dependency pinned to the version being used. (just to make it easier to develop on). Shall I adjust the PR to remove travis and keep the dev dependency? Or if you're not convinced, you can just close it :)
Have made this change in case helpful
I am toying with the idea of switching to https://github.com/testdouble/standard instead of raw Rubocop. It would require us to not rely on Code Climate, but that's not a dealbreaker for me.
What do you think of this instead?
I don't see a problem with testdouble/standard -- the standards seem more sensible than rubocop defaults.
And in terms of moving the rubocop from codeclimate to travis -- I would generally prefer this as I feel like a codeclimate fail is advisory (and needs informed judgement) whereas rubocop fail (like test suite fail) tend to be more absolute.
We've switched to standard by now.
|
2025-04-01T06:40:30.652909
| 2017-04-19T09:51:47
|
222682008
|
{
"authors": [
"Dilshat"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11018",
"repo": "subutai-io/base",
"url": "https://github.com/subutai-io/base/issues/1817"
}
|
gharchive/issue
|
Use programmaitc SS API instead of REST API from Hub Manager module
Currently there are some classes in hub manager module that address SS API via REST. However we have programmatic code API that could be used. REST API is for external clients
Done
|
2025-04-01T06:40:30.655493
| 2016-03-07T19:40:47
|
139073456
|
{
"authors": [
"NikolaiB",
"akarasulu"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11019",
"repo": "subutai-io/base",
"url": "https://github.com/subutai-io/base/issues/189"
}
|
gharchive/issue
|
[UI] [Monitoring] Too many values impacts usability.
For replay this improvement you need:
Test case
Go to SS and login under any account.
Go to Monitoring page.
Improvement: On the Monitoring graphs (i.e. disk) usability is impacted by excessive values. The popup displayed is out of control. Let's condense some of the values into rows where it makes sense. For example here:
On this popup let's condense the "avialable", "used" and "total" metrics in a nice way onto a single row instead of have each one of these on its own row.
Tested.
|
2025-04-01T06:40:30.662646
| 2022-12-16T16:59:56
|
1500555444
|
{
"authors": [
"Dellyare",
"suchmememanyskill"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11020",
"repo": "suchmememanyskill/CssLoader-ThemeDb",
"url": "https://github.com/suchmememanyskill/CssLoader-ThemeDb/pull/140"
}
|
gharchive/pull-request
|
Add DellyBoard theme
DellyBoard
https://github.com/Dellyare/DellyThemes
Theme for keyboard in Apple style
Checklist
Failure to complete this checklist or deleting boxes from the checklist will result in the closing of your pull request unless this is a theme update. Please write any comments regarding this checklist at the bottom of your pull request.
Check every box.
[x] I am the original author of this theme or have permission from the original author to make this pull request.
[x] All copyright of this theme's contents belong to the listed author or is cited in the repository linked above.
[x] This theme's target has been marked appropriately and only styles said target.
[x] This theme works properly on the latest versions of SteamOS for Steam Deck, decky-loader and SDH-CssLoader.
[x] This theme only uses * or !important if absolutely necessary.
[x] This theme is under 4MB in size and uses the least disk space possible.
[x] This theme's preview image does not include text unless it is necessary to describe changes that can be made.
[x] This theme is safe for work and does not contain any sexual, drug-related, or profane content.
[x] This theme prefixes any CSS variables with a unique identifier.
Check one box.
[x] I am not bundling a part of another theme with this theme to encourage mixing and matching themes.
[ ] Themes included with this theme are toggleable using a patch.
Check one box.
[x] This is a keyboard theme applied to the default keyboard.
[ ] This is a system-wide theme applied to the default keyboard. The keyboard is toggleable using a patch.
[ ] This theme does not target the keyboard.
Hi, would you mind joining this discord? We are planning to migrate all themes to a custom built site soon, and want to discuss how that should happen.
For the PR, I'll review it tomorrow
|
2025-04-01T06:40:30.668289
| 2019-08-20T21:37:42
|
483095576
|
{
"authors": [
"ofsound",
"sudara"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11021",
"repo": "sudara/alonetone",
"url": "https://github.com/sudara/alonetone/issues/684"
}
|
gharchive/issue
|
"Community Guidelines" footer link clicked on mobile loads that forum page scrolled down rather than at its top
Same issue that once existed for the main forum link, turbo-links whitelist or something?
Closed by https://github.com/sudara/alonetone/commit/b2a0979f7075e6b56652e80eb7606e5bdf0b1999
|
2025-04-01T06:40:30.758261
| 2018-11-07T13:16:38
|
378283075
|
{
"authors": [
"justinlittman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11022",
"repo": "sul-dlss/rialto-etl",
"url": "https://github.com/sul-dlss/rialto-etl/issues/285"
}
|
gharchive/issue
|
WOS client should log non-200 responses
It currently does not check the response code or log anything. As a result, the user has no way of knowing if there was an error.
|
2025-04-01T06:40:30.819789
| 2021-03-25T05:23:03
|
840569449
|
{
"authors": [
"sumneko",
"winterwolf"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11025",
"repo": "sumneko/lua-language-server",
"url": "https://github.com/sumneko/lua-language-server/issues/465"
}
|
gharchive/issue
|
"Lua.workspace.ignoreDir" doesn't ignore my file
I use custom format to store my settings, similar to json but in lua notation. I called it "luat" and when I added "*.luat" pattern to "Lua.workspace.ignoreDir" it is still highlighted red:
The language service currently has a setting. If a lua file is ignored, but you actively open the file, the language service will temporarily assume that the file has not been ignored (in order to provide functions such as automatic completion and floating prompts). But this does cause inconvenience in some cases, and I will consider how to improve it.
|
2025-04-01T06:40:30.854423
| 2016-05-19T20:22:32
|
155826301
|
{
"authors": [
"sungkim3"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11026",
"repo": "sungkim3/portfolio",
"url": "https://github.com/sungkim3/portfolio/issues/18"
}
|
gharchive/issue
|
About me and Contact sections are not clearly delineated on Home page
On home navigation, about me and contacts do not have distinctive sections.
Created headers to separate these items and display them with very clear sections.
|
2025-04-01T06:40:30.916668
| 2020-05-01T12:36:40
|
610730522
|
{
"authors": [
"Cadair",
"dstansby"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11027",
"repo": "sunpy/radiospectra",
"url": "https://github.com/sunpy/radiospectra/issues/36"
}
|
gharchive/issue
|
Hard to find docs
Do docs exist for this project? If so some sort of link on the github page (maybe in the Readme) would be helpful, as I can't obviously find one.
In case you haven't found them already they are here: https://docs.sunpy.org/projects/radiospectra/en/latest/ :rofl:
|
2025-04-01T06:40:30.932566
| 2019-08-12T15:54:29
|
479723389
|
{
"authors": [
"roblabla"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11028",
"repo": "sunriseos/SunriseOS",
"url": "https://github.com/sunriseos/SunriseOS/issues/414"
}
|
gharchive/issue
|
Update to official latest mergeable
Our mergeable is full of bugs. We should update our config and move to the official hosted version.
Well, I thought mergeable had implemented my feature, but it's actually still sitting in the issue queue. Fuck.
https://github.com/mergeability/mergeable/issues/145
https://github.com/mergeability/mergeable/issues/62
Tagging as upstream. I tried clobbering something up for the proposed "unless" operator, but it's a bit non-trivial.
See #434 .
Blocked on https://github.com/mergeability/mergeable/pull/227
|
2025-04-01T06:40:31.406733
| 2020-01-09T07:26:15
|
547299335
|
{
"authors": [
"DraqueT",
"apatrida",
"dota17",
"edu-de",
"jjenkov",
"kbilsted",
"patric-r",
"shawnweeks"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11029",
"repo": "super-csv/super-csv",
"url": "https://github.com/super-csv/super-csv/issues/176"
}
|
gharchive/issue
|
v2.50 release
What features and issues are more urgent for version 2.5.0? And what is your plan for 2.5.0?
@dota17 @jjenkov @ChristianWilkie
I like you guys feedback on this. In my mind it is
get jenkins up and running with code coverage
its been a while since last release, so get a release out soon
maybe we should upgrade to java 8 ,just like @jjenkov says java 6 or java 7 is too old #169
I would say, that the highest priority should be to just get a release out. As far as I can see, v. 2.4.0 is 3 years old. I would go for a minimum of functionality in 2.5.0, just to bring SuperCSV "back to life" .
If upgraded to J8, would SuperCSV work correctly under J9+? My application uses SCSV, and a problem that I run into is its incompatibility with J9+ just generally. It would be a big upgrade for me if that issue was resolved. (as always, thank you for maintaining this!)
@DraqueT Is the approach to make a J9 specific release? or will the https://github.com/super-csv/super-csv/pull/152 suffice?
Hi, @kbilsted
I don`t think super-csv version 2.5.0 make a J9 specific release. I think it is enough that ensuring super-csv 2.5.0 can be imported and used in java9 module application(#138 ).
Because java9 modules is important and influential. When we upgrade Java 9, we need to design reasonable modules for super-csv,
And
Hi, @kbilsted
Are you plan to upgrade to java8 in 2.5?
If so, we can incorporate super CSV java8 into super CSV and address #152. that will ensuring super-csv can be imported and used in java9 module application.
Adding support for escaping quotes using \" instead of leaving this as unhandled seems reasonable, very hard to recover from errors based on this otherwise.
@dota17 I'd vote for caution. Can you make things work in a side branch for JDK 9 for now and get experince with it with some users?
isn't java8-specifc jar a part of the existing mvn build?
I can probably help out on getting this to compile on Java 8. This is used on one of the Apache projects I'm working on and there are some bug fixes I could really stand to have. Glad to see active development again.
The master branch does build just fine on Java 8 though. Not sure if what some of the other issues actually were.
@shawnweeks thanks. I'll install jdk8 and retry
@shawnweeks thanks. I'll install jdk8 and retry
I just ran "mvn clean install" using adoptopenjdk on macos. All the test cases ran and everything. Jars appear to be there in the build not sure what else needs to be tested but it should at least build until we move the compile versions forward.
Hi, @kbilsted
I am sorry. Can you change the name of the branch from java9 to java8 :).
sorry yes. I have been away for some days. Had some business I had to do.
Any chance for a 2.5.0 release soon?
What issues/things are still open to do so?
Any chance for a 2.5.0 release soon?
What issues/things are still open to do so?
Same question here, do you have a 2.5.0 release soon or is anything open here ?
|
2025-04-01T06:40:31.411279
| 2015-12-16T11:45:31
|
122486429
|
{
"authors": [
"ZioberMichal",
"kbilsted",
"nagraj321"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11030",
"repo": "super-csv/super-csv",
"url": "https://github.com/super-csv/super-csv/issues/87"
}
|
gharchive/issue
|
Delete File
Just want to know whether we can delete a file using supercsv.
First created the ICsvBeanWriter
Write the data using ICsvBeanWriter
Close the ICsvBeanWriter
even if the data is not available in the step 2 we are creating the file and their is not way to delete the file.
any idea how can we achieve this?
Thanks,
Nagaraju
HI i'm not fully sure i understand the problem. Try submitting some Java code for how you think you could use the framework. Why not simply delete the file yourself? why start wrtiting to a file before you check if you have data to write?
ICsvBeanWriter is just an abstraction for writing any bean as CSV row. CsvBeanWriter class pushes CSV String to Writer object which is given in constructor. You probably uses FileWriter class which creates file at the beginning.
You have two solutions:
You can use StringWriter which writes data to memory. At the end, after writing, you can check the length of the String and save it to file if needed. Consumes time and memory but doesn't create file if it is not needed.
You can close the stream and remove file if it is empty.
Summary: does not look like an issue in super-csv library. I vote to close this.
+1 @ZioberMichal
|
2025-04-01T06:40:31.414352
| 2015-07-12T17:05:25
|
94581798
|
{
"authors": [
"ZioberMichal",
"ludovicofischer"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11031",
"repo": "super-csv/super-csv",
"url": "https://github.com/super-csv/super-csv/pull/60"
}
|
gharchive/pull-request
|
Add java.time parsers and formatters
Fixes #53 . I have tried to match the structure of the java.time package, which diverges from joda.time. I also tried to design locale-independent tests that also use the current locale.
Hi. It looks really good! I've noticed that you are using different imports order than it is in project.
Project's order:
static
java.*
org.*
In your commit:
org.*
java.*
static
I think that we should use the same order.
I’ll see what I can do about the imports. Importing the eclipse configuration does not seem to solve all format issues :-)
This is a reason why I've crated this issue #61 :smirk:
|
2025-04-01T06:40:31.510532
| 2022-10-04T18:50:07
|
1396704368
|
{
"authors": [
"mdfaisalahmed057",
"surajondev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11033",
"repo": "surajondev/JavaScript",
"url": "https://github.com/surajondev/JavaScript/pull/60"
}
|
gharchive/pull-request
|
blackjackgame
this will help to learn the javascript concept by making this game
Thanks for PR. Merged.
|
2025-04-01T06:40:31.513283
| 2017-06-14T12:02:15
|
235856143
|
{
"authors": [
"kadel",
"surajssd"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11034",
"repo": "surajssd/opencomposition",
"url": "https://github.com/surajssd/opencomposition/issues/23"
}
|
gharchive/issue
|
Only health at container level
Right now pod spec allows you to do both livenessProbe and readinessProbe and since both the fields represent the same data strcuture behind the scene which is called probes.
So a container will look like the following:
containers:
- image: wordpress:4
health:
httpGet:
path: /
port: 80
initialDelaySeconds: 120
timeoutSeconds: 5
the pod that is generated out of this will look like this:
containers:
- image: wordpress:4
livenessProbe:
httpGet:
path: /
port: 80
initialDelaySeconds: 120
timeoutSeconds: 5
readinessProbe:
httpGet:
path: /
port: 80
initialDelaySeconds: 120
timeoutSeconds: 5
So essentially you just specify health field which is probe type and then it is populated into readinessProbe and livenessProbe in the output. But if user provides the fields readinessProbe or livenessProbe explicitly then health field is ignored! Those fields are used directly.
agreed, with one exception. lets fail if there is health field with readinessProbe or livenessProbe defined
|
2025-04-01T06:40:31.563391
| 2024-04-11T12:16:41
|
2237591398
|
{
"authors": [
"codecov-commenter",
"internetova"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11035",
"repo": "surfstudio/surf-flutter-app-template",
"url": "https://github.com/surfstudio/surf-flutter-app-template/pull/114"
}
|
gharchive/pull-request
|
feat: Added a service for displaying snacks
Checklist
A service has been written that will help you use flutter_easy_dialogs to display snacks "out of the box".
https://github.com/surfstudio/surf-flutter-app-template/assets/39735343/142d53c3-d05f-4d2c-b4de-421ab6411766
[x] Have you added an explanation of what your changes do and why you'd like us to include them?
[x] Is there an existing issue for this PR?
link issue here (use keywords like fix, close, resolve etc. if necessary)
[x] Have the files been linted and formatted?
[ ] Have the docs been updated to match the changes in the PR?
[ ] Have the tests been updated to match the changes in the PR?
[x] Attached videos/screenshots demonstrating the fix/feature.
[ ] Have you run the tests locally to confirm they pass?
New Features
What new capabilities does this PR add?
What docs changes are needed to explain this?
Codecov Report
Attention: Patch coverage is 0.77519% with 128 lines in your changes are missing coverage. Please review.
Project coverage is 23.25%. Comparing base (339c48c) to head (59786b9).
Report is 17 commits behind head on main.
Files
Patch %
Lines
...tures/snack_queue/presentation/snack_queue_wm.dart
0.00%
61 Missing :warning:
...k_queue/presentation/default_snack_controller.dart
4.16%
23 Missing :warning:
...tures/debug/presentation/ui_kit/ui_kit_screen.dart
0.00%
18 Missing :warning:
.../features/debug/presentation/ui_kit/ui_kit_wm.dart
0.00%
8 Missing :warning:
...s/snack_queue/presentation/snack_queue_widget.dart
0.00%
5 Missing :warning:
...snack_queue/presentation/snack_queue_provider.dart
0.00%
4 Missing :warning:
...atures/snack_queue/presentation/top_snack_bar.dart
0.00%
4 Missing :warning:
lib/l10n/app_localizations_en.g.dart
0.00%
2 Missing :warning:
lib/l10n/app_localizations_ru.g.dart
0.00%
2 Missing :warning:
...es/snack_queue/presentation/snack_queue_model.dart
0.00%
1 Missing :warning:
Additional details and impacted files
@@ Coverage Diff @@
## main #114 +/- ##
==========================================
- Coverage 26.39% 23.25% -3.14%
==========================================
Files 55 61 +6
Lines 917 1045 +128
==========================================
+ Hits 242 243 +1
- Misses 675 802 +127
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
2025-04-01T06:40:31.571924
| 2024-02-09T09:53:53
|
2126803933
|
{
"authors": [
"Evgenia-bit",
"samir-a-ts"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11036",
"repo": "surfstudio/surf-flutter-faded-text",
"url": "https://github.com/surfstudio/surf-flutter-faded-text/pull/2"
}
|
gharchive/pull-request
|
feat: Changed readme file
Changes
updated readme file
Checklist for self-check
[x] Commits and PRs have been filed according to the rules on the project.
[x] The author is marked as an assigne and assigned mandatory reviewers.
[x] Required labels marked
[ ] Specified related tasks and/or related PRs.
[x] Specified Changes.
[ ] Attached videos/screenshots demonstrating the fix/feature.
[x] All unspecified fields in the PR description deleted.
[ ] New code covered by tests.
Checklist for reviewers
[ ] CI passed successfully (with a green check mark).
[ ] PR is atomic, by volume no more than 400 (+-) corrected lines (not including codogen).
Design:
[ ] System design corresponds to the agreements on structure and architecture on the project.
[ ] The code is decomposed into necessary and sufficient components.
Functionality:
[ ] The code solves the problem.
[ ] Any changes to the user interface are reasonable and look good.
Complexity:
[ ] The code is clear, easy to read, functions are small, no more than 50 lines.
[ ] The logic is not overcomplicated, there is no overengineering (no code sections that may be needed in the future, but no one knows about it).
Tests:
[ ] Updated or added tests for mandatory components.
[ ] The tests are correct, helpful, and well designed/developed.
Naming:
[ ] The naming of variables, methods, classes and other components is understandable.
Comments:
[ ] The comments are understandable and helpful.
Documentation:
[ ] All labels are correct
[ ] Technical documentation updated (after approval, updates last reviewer).
@Evgenia-bit I do not have permission for creating a PR or commiting to a branch. Please, contact SA and lease me with access.
|
2025-04-01T06:40:31.704565
| 2017-12-07T14:53:32
|
280158499
|
{
"authors": [
"dmitrykurmanov",
"tjtaojiang",
"tsv2013"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11037",
"repo": "surveyjs/widgets",
"url": "https://github.com/surveyjs/widgets/issues/12"
}
|
gharchive/issue
|
widget request "lokesh pretty checkbox"
https://lokesh-coder.github.io/pretty-checkbox/#basic-checkbox
from the https://github.com/surveyjs/editor/issues/232
I've added pretty-checkbox custom widget basic implementation. It works for checkboxes and radiobuttons with limited set of options. The widget inner markup highly depends on the certain pretty-checkbox options. This makes common implementation of this widget too complex IMO.
@tsv2013, I can't make your version of the widget to work for my case. I am not sure what I miss.
I attached a simplified version of my code, "suyveyjs-form.html" and "surveyjs-form.js" here: https://github.com/surveyjs/editor/issues/232, to show how it works for my case.
Please advise.
@tjtaojiang Sorry for delay. I've added example how to integrate pretty-checkbox custom widget. It resides on our site - https://surveyjs.io/Examples/Library/?id=custom-widget-pretty-checkbox&platform=Angular2
As for your implementation of the custom widget - I cann't find binding between question value and the widget control.
I suggest you try to integrate our custom widgets pack into your application.
We'll provide you with the necessary support. Please feel free to contact us in case of any difficulties.
@tsv2013, binding by: row.value and row.text. In my simplified code, you can run it and see the result.
@tsv2013, I noticed one thing, your alter the look and feel for the pretty checkbox. Check https://lokesh-coder.github.io/pretty-checkbox/, and look for "Svg" and "Recurring" checkbox, it supposed to be curved.
@tsv2013, I copy your sample code from your link above and click "Edit in Plunker", for some reason, I still can't make your code to work.
I attached a simplified version from my implementation here for you to see how it work for me:
surveyjs_widget_for_pretty_checkbox.zip.
To answer your previous concern, check row.value and row.text in the js, which I used for data binding.
I think my way has full control in term of custom widget. I just do not know if it will break some surveyjs process , like required filed property, etc. I will keep testing.
Any advise is much appreciated.
@tsv2013, I noticed a very big bug in your recent changes.
All other types are blocked from addCustomWidget, except "radiogroup" and "checkbox". In other words, in "isFit:", you can only see "radiogroup" and "checkbox" type, other types like "text" will be blocked.
I think you need to fix this issue.
@tjtaojiang On this page (https://surveyjs.io/Examples/Library/?id=custom-widget-pretty-checkbox&platform=Angular2) you can find the "Edit in Plunker" button. In this case our site creates a new "ready to run" plunk containing the current sample code.
E.g. after I clicked this button I've got the following plunk with the pretty-checkbox example for angular platform:
https://plnkr.co/edit/chHf8x?p=preview
It works for me ok.
Does this plunk work for you?
@tjtaojiang In the sampe plunk - https://plnkr.co/edit/chHf8x?p=preview
the following lines from the index.html files are responsible for adding the "pretty-checkbox" custom widget functionality:
<link rel="stylesheet"<EMAIL_ADDRESS><script src="https://unpkg.com/surveyjs-widgets"></script>
because of the code of our "pretty-checkbox" custom widget implementation is resides in the "https://unpkg.com/surveyjs-widgets" script.
@tjtaojiang I've also created a plunk with the code you sent me in the "surveyjs_widget_for_pretty_checkbox.zip" archive.
This is the link to the plunk with your code - https://plnkr.co/edit/e3uqCK?p=preview
We can use this plunk as a starting point for the further discussion.
Note that I've commented reference to our custom widgets package (in the index.html file) because you are using your own implementation of the "pretty-checkbox" custom widget:
<!--<script<EMAIL_ADDRESS>type="text/javascript"></script>-->
Could you check this plunk also?
My thoughts about the pretty-checkbox cutom widget implementation
"pretty-checkbox" is the CSS library and requires different markup for different features.
This makes us to drop our default-generated markup of the "checkbox", "radiogroup" (and that much more worse the "matrix") questions and replace it with the markup required to make "pretty-checkbox" work properly.
If we remove our default markup, we become responsible for
the two-way binding of the custom widget: subscribe on the question value changes and reflect them in the UI, and subscribe the UI control input and pass entered values to the question;
the whole SurveyJS question functionality as read-only behaviour, "other item", "column count" and other options
Our basic implementation of the "pretty-checkbox" custom widget at this moment contains basic options and binding functionality and supports "checkbox" and "radiogroup" quesions only.
The full functionality can be added to the certain custom widget implementation by the united efforts of the SurveyJS team and the community. I hope on your understanding and support.
@tsv2013, for some reason, this Plunker version does not work when I copy them into my local: https://plnkr.co/edit/?p=preview. I still see the old radio and checkbox.
@tsv2013, for some reason, this Plunker version does not work when I copy them into my local: https://plnkr.co/edit/chHf8x?p=preview. I still see the old radio and checkbox.
@tjtaojiang Could you tell me whether the plunker https://plnkr.co/edit/chHf8x?p=preview works for you or not?
@tjtaojiang hello! Could you please describe your use case? Does the example in our site looks good? Do your problem with integrate widget in your project or with widget functionality?
@dmitrykurmanov, as I mentioned above, I copy three files in https://plnkr.co/edit/chHf8x?p=preview into my local, because some websites are blocked by my company, I need to use https://go.weblife.io to download all blocked css and js into my local. I attach zip file contain all three I modified and used in my local. You can see duplicated due to the websites are block and I have downloaded. After I put all three in one directory and click html, I still only see the default (old fashion) radio and checkbox.
@dmitrykurmanov, as I mentioned above, I copy three files in https://plnkr.co/edit/chHf8x?p=preview into my local, because some websites are blocked by my company, I need to use https://go.weblife.io to download all blocked css and js into my local. I attach zip file contain all three I modified and used in my local. You can see duplicated due to the websites are block and I have downloaded. After I put all three in one directory and click html, I still only see the default (old fashion) radio and checkbox.
test-surveyjs-form.zip
Hi, @dmitrykurmanov, @tsv2013,
After getting the sample code from https://plnkr.co/edit/chHf8x?p=preview, I do see the default checkbox and radiogroup are changed to pretty checkbox.
But at least I noticed those issues:
(1). You have modified the default checkbox to a square shaped pretty checkbox. If I need other shapes like "Svg" -> "Recurring", a curve shape checkbox, on https://lokesh-coder.github.io/pretty-checkbox/, or if I need to use the old square checkbox, I still need to create a custom widget, which is actually what users will need: how to create custom widget to cope their requests.
(2). Your new implementation deactivate colCount. Check your sample case.
(3). https://unpkg.com/surveyjs-widgets, or actually<EMAIL_ADDRESS>has disabled any custom widget implementation. Uncomment addCustomWidgetForDatePicker() in test-surveyjs-forms.js in my attachment:
test surveyjs widget.zip
, you will see isFit: never gets called.
(4). For lengthy label radio or checkbox, the current pretty checkbox cuts the total description. Check my attachment above.
@tjtaojiang Answering your questions:
(1) - I've implemented "addOn" feature for the "pretty-checkbox" custom widget. It allows to insert any additional markup before the "label" tag
For "SVG" mode you will need to add the following code:
var prettyCheckbox = Survey.CustomWidgetCollection.Instance.getCustomWidgetByName(
"pretty-checkbox"
);
var checkboxSettings = prettyCheckbox.widgetJson.settings.checkbox;
checkboxSettings.rootClass = "pretty p-svg p-curve";
checkboxSettings.addOn =
"<!-- svg path -->" +
'<svg class="svg svg-icon" viewBox="0 0 20 20">' +
'<path d="M7.629,14.566c0.125,0.125,0.291,0.188,0.456,0.188c0.164,0,0.329-0.062,0.456-0.188l8.219-8.221c0.252-0.252,0.252-0.659,0-0.911c-0.252-0.252-0.659-0.252-0.911,0l-7.764,7.763L4.152,9.267c-0.252-0.251-0.66-0.251-0.911,0c-0.252,0.252-0.252,0.66,0,0.911L7.629,14.566z" style="stroke: white;fill:white;">' +
"</path>" +
"</svg>";
(2) I've implemented the "colCount" feature
(3) I've checked different custom widgets (including the "pretty-checkbox") on the same page of a survey - they all works fine for me.
I've checked the last code you sent us and found that you commented your custom widget initialization function:
function onAngularComponentInit() {
// addCustomWidgetForDatePicker();
Survey.SurveyNG.render("surveyElement", {model: survey});
}
After I've uncommented this call, everything works ok for me.
(4) I've checked your attachment, and see that long item descriptions are not trimmed. At this point I'd like to clarify - what behaviour for the long item descriptions is suitable for you? Could you share a screenshot or drawing with desired view.
@tjtaojiang The changes mentioned above will be available since 0.98.6. For more details you can check the 0d7f9b9195ee6a07cc578981e4ec54cb48b26341 commit.
Hi @tsv2013,
(1). Due to the (2) - (3), I can't test this: "If I need other shapes like "Svg" -> "Recurring", a curve shape checkbox, on https://lokesh-coder.github.io/pretty-checkbox/, or if I need to use the old square checkbox..."
(2). colCount is still not working. I attach my code here for reference
test surveyjs widget.zip. You can check the attached image file, test-surveyjs-form.jpg .
(3). Check my attachment above, https://unpkg.com/surveyjs-widgets, or actually now<EMAIL_ADDRESS>still disable radiogroup and checkbox. Much worse, in my real code (not attached), it disables any custom widget implementation.
(4). Check the attached image file, lengthy label still stretch out and trimmed.
@tjtaojiang
As I wrote in this comment (https://github.com/surveyjs/widgets/issues/12#issuecomment-353289703) - "The changes mentioned above will be available since 0.98.6". After 0.98.6 will be released, you will be able to try "addOn" feature and "colCount" option. I'll also add corresponding demo on our site.
As for (3) - I'm not able to reproduce this issue on my side. Can you prepare an example in plunker or any other sandbox or project or github repo? So I will be able to reproduce the issue and fix it.
(4) - I do understand that things are not so good as they can be. But could you give your opinion - how it should be fixed?
@tjtaojiang I've checked your code and screenshot you sent. And I can say that datePickerWidget is applied on your page. You see an empty space instead of date picker control because you disable default rendering via setting the isDefaultRender: false option and nothing created in the afterRender: function (question, el) { } function as I can see from your custom datePickerWidget code from your file:
function addCustomWidgetForDatePicker() {
var datePickerWidget = {
isFit: function (question) {
alert("question = |" + question + "|\n"
+ "question.getType() = |" + question.getType() + "|\n"
+ "question.inputType = |" + question.inputType + "|\n"
+ "question.class = |" + question.class + "|\n"
+ "question.name = |" + question.name + "|\n"
+ "question.title = |" + question.title + "|\n"
);
return (question.getType() === 'text') && (question.inputType === 'date');
},
isDefaultRender: false,
afterRender: function (question, el) {
}
};
Survey.CustomWidgetCollection.Instance.addCustomWidget(datePickerWidget);
}
@tjtaojiang In order to use custom widgets, their code sould be loaded after the core SurveyJS library:
<script src="https://surveyjs.azureedge.net/0.98.5/survey.angular.js" type="text/javascript"></script>
<script<EMAIL_ADDRESS>type="text/javascript"></script>
@tsv2013,
Refer to your: "As for (3) - I'm not able to reproduce this issue on my side. Can you prepare an example in plunker or any other sandbox or project or github repo? So I will be able to reproduce the issue and fix it."
Please use my previous zip file, test surveyjs widget.zip, the 4th post above this one, click test-surveyjs-form.html after you unzip. You can see radio and checkbox element never got passed into isFit: in addCustomWidgetForDatePicker() .
I will post another post to illustrate how do I cope my issues (10 - (4) next.
Hi @tsv2013 and @dmitrykurmanov,
Here is how I solve my issues from (1) - (4). I need to comment out , otherwise it will disable all custom widget I created, as I mentioned before.
Here is the attachement for it:
test surveyjs widget-3.zip
click test-surveyjs-form.html after you unzip it.
I am not sure my way will introduce any other issues, which may break some calculation and dependency, like required property etc.
Please offer advise.
Hi @tsv2013 and @dmitrykurmanov,
Here is how I solve my issues from (1) - (4). You need to comment out<EMAIL_ADDRESS>as in test-surveyjs-form.html, otherwise it will disable all custom widget I created, as I mentioned before.
Here is the attachement for it:
test surveyjs widget-3.zip
, click test-surveyjs-form.html after you unzip it.
I am not sure if my code will introduce any other issues, which may break some surveyjs calculation and dependency, like required property etc.
Please offer advise.
@tjtaojiang The "radio and checkbox element never got passed into isFit in addCustomWidgetForDatePicker()" because they are already fit for another custom widget. It is hard to apply two custom widgets on the same question. This can lead to unexpected side effects.
But until now our pretty-checkbox custom widget was activated always by question type if pretty-checkbox css styles were loaded on the page. I've changed this behaviour to activation by property: renderAs that should be set to "prettycheckbox" value.
As for your implementation of the pretty-checkbox custom widgets. Your certain implementation can better be suitable for solving your certain tasks. I don't see any problems in your implementation for now except of absence of two-way bindings: UI will not updated in case of model has been changed and choosen in UI values are not passed in the model. The last problem is crucial because if you choose some items and press Complete button you will see validation complains "Please answer the question." because the values you have choosen haven't been stored in the question.
@tsv2013, I will look into the two-way bindings issue you mentioned.
As far as "radio and checkbox" are already fit for another custom widget, then user will not be able to create a widget for their requirement? This may not be a good idea. Can we remove the limit?
@tsv2013, I will look into the two-way bindings issue you mentioned. Do you have any recommendation how should I modify my implementation to cope with the issue?
As far as "radio and checkbox" are already fit for another custom widget, then user will not be able to create a widget for their requirement? This may not be a good idea. Can we remove the limit?
Hi @tsv2013 and @dmitrykurmanov,
Here is my current version which will fix two-way bindings issue. It also contains a custom widget for a date picker (pikaday object):
test surveyjs widget-4.zip. If you wish, I can move the date picker (pikaday object) custom widget to a separate thread.
But there is a minor issue which I have not figured out. For some reason, the checkbox does not show the white check sign inside the checkbox. Do you know why?
I will be on vacation. I won't reply to you until I come back. Thanks.
Hi @tsv2013 and @dmitrykurmanov,
Here is my current version which will fix two-way bindings issue. It also contains a custom widget for a date picker (pikaday object):
test surveyjs widget-4.zip
. If you prefer, I can move the date picker (pikaday object) custom widget to a separate thread.
But there is a minor issue which I have not figured out. For some reason, the checkbox does not show the white check sign inside the checkbox. Do you know why?
I will be on vacation. I won't be able to reply to you until I come back. Thanks.
|
2025-04-01T06:40:31.725696
| 2019-10-12T06:31:57
|
506141399
|
{
"authors": [
"Inframind",
"shadabms"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11038",
"repo": "sushant-j/my-first-pull-request",
"url": "https://github.com/sushant-j/my-first-pull-request/pull/152"
}
|
gharchive/pull-request
|
Shadab M S
I have added my code here please accept the pull request.
Hey @shadabms ,
Could you please change the font size?
|
2025-04-01T06:40:31.789170
| 2020-08-20T20:18:32
|
683069915
|
{
"authors": [
"LucianoPolit",
"svdo"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11039",
"repo": "svdo/ReRxSwift",
"url": "https://github.com/svdo/ReRxSwift/issues/14"
}
|
gharchive/issue
|
New version
New version should be released in order to start using SPM. I would suggest doing a minor bump to 2.3.0.
Yes indeed, that also occurred to me while I was drinking my beer this evening 😀 Will do tomorrow!
Thanks a lot and enjoy that beer 💪
Done! Since nothing else changed, I labeled the version 2.2.2. Thanks again!
|
2025-04-01T06:40:31.792560
| 2023-09-03T09:16:27
|
1879005804
|
{
"authors": [
"JayeshVP24",
"cogoo",
"t4top"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11040",
"repo": "svelte-on-solana/wallet-adapter",
"url": "https://github.com/svelte-on-solana/wallet-adapter/pull/57"
}
|
gharchive/pull-request
|
Wallet Standard support, bug fixes and upgrades
Wallet Standard support
Dependencies upgraded to recent versions
Accessibility warnings fixed
If you can't wait for the merge of this PR, consider using aztemi/svelte-on-solana-wallet-adapter repo instead. It is a fork and already has these commits with additional features integrated.
Thanks!
Will try this out now
That'll great @cogoo. Thanks.
@cogoo
have you also tried
https://github.com/portalpayments/svelte-on-solana-wallet-adapter
do u know if these support solana mobile adapters?
which one do you recommend?
|
2025-04-01T06:40:31.800143
| 2024-11-07T20:55:14
|
2642176520
|
{
"authors": [
"AdrianGonz97",
"manuel3108"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11041",
"repo": "sveltejs/cli",
"url": "https://github.com/sveltejs/cli/pull/275"
}
|
gharchive/pull-request
|
feat: simple testing library
Alternative to #32
This takes more of a lightweight approach where we provide a bare minimum api for testing rather than controlling the whole testing pipeline. This should allow for better flexibility as we're no longer tied to any particular testing framework.
Notes:
Had to rename the directory of the vitest addon as it was somehow interfering with vitest or tinypool during runtime, causing shit to randomly break
Storybook has been pinned to v8.3.x as v8.4.x flat out doesn't work. I haven't had a moment to fully investigate the issue yet.
TODO:
Snapshots?
look into using vitest workspaces
Should we continue to use degit to clone the vite templates? If so, should we pin it to a specific commit hash (see here)? Or should this be something we manage fully (perhaps as part of create)?
Ohhh and there is another thing I just noticed: How are the directory names inside .test-output/_tests generated? Here are my directories inside drizzle
-1930824653_0
-1930824653_1
-1930824653_2
-1930824653_3
-1930824653_4
-1930824653_5
-1930824653_6
-1930824653_7
They are semi-useful at least. In #32 I generated them based on the options passed for each test, so that if a test failed you could identify it by directory name and run it by hand.
How are the directory names inside .test-output/_tests generated? Here are my directories inside drizzle
they're based on the task id provided by vitest. Each test gets their own meta.json (similar to how we had it in svelte-add) that documents the options that were passed in.
They are semi-useful at least. In #32 I generated them based on the options passed for each test, so that if a test failed you could identify it by directory name and run it by hand.
not sure if that's a scalable solution. imagine having multiple test cases that share the same addon options but they evaluate different things
|
2025-04-01T06:40:31.807026
| 2022-04-12T18:28:50
|
1202238620
|
{
"authors": [
"Conduitry",
"benmccann",
"cowboycodr",
"rocketinventor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11042",
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/4598"
}
|
gharchive/issue
|
Support for subdomains in SvelteKit
Describe the problem
I do want to start this off with the acknowledgment that subdomains with SvelteKit is possible. However, there is no clear documentation nor have I found any clear and concise examples on how to do so. The only reason I even assume it is possible is because of kit.svelte.dev.
If I am wrong, and there is a clear and documented way of adding support for subdomains to SvelteKit websites then please do let me know and I will close this issue.
Describe the proposed solution
The proposed solution is that of the title.
Alternatives considered
Subdirectories are fine, just not ideal.
Importance
nice to have
Additional Information
No response
I don't know what this is asking for. You have access to the hostname via the event that's sent to endpoints and the $page store that's available to pages.
kit.svelte.dev is its own app/site that is completely separate from the app/site at svelte.dev.
Is the request to be able to have one app serve multiple hostnames, and have the hostname be part of what distinguishes routes, instead of just the path?
I understand I have access to the site's hostname. Even then there is still no clear documentation.
I am using cloudflare pages and see no clear way of using multiple subdomains with one or even multiple SvelteKit projects.
subdomains are something that you configure with DNS (not with your frontend framework), so there's nothing that SvelteKit would do here
If you'd like help or are confused about something, please ask on Discord
Hi!
I think that the question being asked is if someone wants to use dynamic subdomain routing can they do that in Svelte Kit as part of their routing scheme:
I.e.
http://mysubdomain.example.com => '/subdomain/mysubdomain/'
http://myexcellentsubdom.example.com/homepage => '/subdomain/myexcellentsubdom/homepage'
http://first.second.example.com => '/subdomain/first/second/'
http://first.second.example.com/another/page => '/subdomain/first/second/another/page'
In express this is possible using express-subdomain-handler
In other words: Just like you can use src/routes/about/+page.svelte and src/routes/blog/[slug]/+page.svelte to make /blog/foo and about pages on mysite.com, so too you should be able to create a route for a subdomain such as /src/routes/[subdomains]/login/+page.svelte or /routes/search.site/+page.svelte to make a login.mysite.com or search.mysite.com subdomain.
This seems like a reasonable and understandable request, so I think that this issue should maybe be reopened.
|
2025-04-01T06:40:31.814495
| 2022-09-07T01:05:58
|
1363961052
|
{
"authors": [
"Rich-Harris",
"papiro"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11043",
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/6626"
}
|
gharchive/issue
|
1.0.0-next.455 causing broken Netlify response header behavior
Describe the bug
With a simple upgrade to @sveltejs/kit from 1.0.0-next.454 to 1.0.0-next.455 with no code changes except the ones needed for the upgrade, the response headers of the pages as served from Netlify are different. I suspect that something changed with the build to where now the netlify.toml is not being read (even though it seems to be detected when looking at the Netlify build logs for the branch).
The response headers (truncated) for an initial html page are changed as follows:
- cache-control: public, max-age=0, must-revalidate
- content-type: text/html; charset=UTF-8
- cross-origin-embedder-policy: require-corp
- cross-origin-opener-policy: same-origin
- strict-transport-security: max-age=31536000; includeSubDomains; preload
+ cache-control: no-cache
+ content-type: text/html
+ link: <./_app/immutable/assets/_layout-7e2cdc5f.css>; rel="preload";as="style"; nopush, <./_app/immutable/assets/modalProvider-2c7dffb9.css>; rel="preload";as="style"; nopush, <./_app/immutable/assets/outputLoadingMessage-6ab61537.css>; rel="preload";as="style"; nopush, <./_app/immutable/assets/tooltip-0a1c217e.css>; rel="preload";as="style"; nopush, <./_app/immutable/start-9e2d9bc6.js>; rel="modulepreload"; nopush, <./_app/immutable/chunks/index-f9b80742.js>; rel="modulepreload"; nopush, ...
The only custom ones set in the netlify.toml are the cross-origin headers you see there (COOP and COEP).
Reproduction
Upgrade from 1.0.0-next.454 to 1.0.0-next.455 and follow the simple changes described here
Logs
No response
System Info
The build image used is the default one in Netlify Ubuntu Focal 20.04.
Severity
blocking an upgrade
Additional Information
No response
Please provide a repro as described in the issue template
https://github.com/papiro/svelte-kit-netlify-issue
netlify: https://master--aquamarine-palmier-db1503.netlify.app/
I haven't been able to get the old behaviour working, despite countless permutations of kit and adapter-netlify versions, and the provided repro won't build on Netlify at all, so unfortunately I have no idea what changed. The https://master--aquamarine-palmier-db1503.netlify.app/ deployment doesn't appear to be active.
I can't spend any more time on this, so if it's not possible to create a reliable repro then I'm afraid we'll have to close this. For all we can tell, it could be a change to Netlify itself. It seems quite unlikely that a change to Kit would result in this behaviour — far more likely to be a change to adapter-netlify.
Hmmmm, i left a clean git commit history so you could see what changes i made. Barely any. I apologize i took the site down. Got overzealous cleaning stuff up. I can redeploy it, but it sounds like you're confident this isn't a Svelte-Kit issue anyways. Also, I moved past the issue by creating a +page.ts so it's not an issue which is bothering me any longer so we can close it if you'd like. I'd be okay with that.
|
2025-04-01T06:40:31.828517
| 2023-01-17T00:38:50
|
1535632866
|
{
"authors": [
"aakash14goplani",
"benmccann",
"dummdidumm",
"hartwm",
"izznatsir",
"leoj3n",
"lettucebowler"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11044",
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/8561"
}
|
gharchive/issue
|
Adapter Vercel Image Optimizations
Describe the problem
Would be really awesome if adapter-vercel package accepted images option for using Vercel's native image optimization. It only requires simple addition of images object to build config file .vercel/output/config.json. I would submit a pull request but I am sure the options, naming conventions, etc would not be as the team would desire. Given it is such a straightforward task it probably wouldn't prove helpful.
Describe the proposed solution
Here are the options
type ImageFormat = 'image/avif' | 'image/webp';
type RemotePattern = {
protocol?: 'http' | 'https';
hostname: string;
port?: string;
pathname?: string;
};
type ImagesConfig = {
sizes: number[];
domains: string[];
remotePatterns?: RemotePattern[];
minimumCacheTTL?: number; // seconds
formats?: ImageFormat[];
dangerouslyAllowSVG?: boolean;
contentSecurityPolicy?: string;
};
https://vercel.com/docs/concepts/image-optimization
https://vercel.com/docs/build-output-api/v3#build-output-configuration/supported-properties/images
Alternatives considered
No response
Importance
nice to have
Additional Information
No response
I'm not sure just adding the options is enough, there's also work involved to create a specific image component I guess? Related to #241
I'm not sure just adding the options is enough, there's also work involved to create a specific image component I guess? Related to #241
you can leave that to the community, because thats more subjective and that would be separate from an adapter package anyway, but without the ability to put these options in the build config then you can't use it, so this shouldn't be the breakpoint of using features
Also all the solutions I have seen are for local images, not remote images.
I'm not sure just adding the options is enough, there's also work involved to create a specific image component I guess? Related to #241
Next and Nuxt have supported image components, but from my understanding once enabled you could also just update your image srcs manually from eg /public/images/logo.png to /_vercel/image?url=/public/image/logo.png?w=320&q=80. A fancy image component could come later if the community wants it, but just enabling this would be a great enhancement IMO
Next and Nuxt have supported image components, but from my understanding once enabled you could also just update your image srcs manually from eg /public/images/logo.png to /_vercel/image?url=/public/image/logo.png?w=320&q=80. A fancy image component could come later if the community wants it, but just enabling this would be a great enhancement IMO
Exactly
@lettucebowler has made a PR for this #8667.
Someone shared an example on Discord (https://discord.com/channels/457912077277855764/1068354495240741005/1070047413898850424)
// Image.svelte
<script lang="ts">
import { dev } from '$app/environment';
let className = ""
export { className as class };
export let image:WpImage
export let showTitle = false
export let lazy = true
export let height = 1280
export let quality = 70
export let width = 720
export let unoptimized = false
const sizeArr = [640, 768, 1024, 1280, 1536, 2048]
const vercelImg = (sourceUrl:string,size:number) =>
(`/_vercel/image?url=${encodeURIComponent(sourceUrl)}&w=${size}&q=${quality} ${size}w`)
$:({ srcSet, sourceUrl, title, altText} = image)
$: vercelSrcSet = sizeArr.map(size=>vercelImg(image?.sourceUrl,size)).join(',')
// dyanmic srcset only for optimized (w/vercel)
$: imageSrcSet = (unoptimized || dev) ? srcSet : vercelSrcSet
let element
</script>
<img
{width}
{height}
title={showTitle ? title : undefined}
srcset={imageSrcSet}
src={sourceUrl}
alt={altText ?? title}
class={className ?? undefined}
loading={lazy ? 'lazy' : 'eager'}
{...$$restProps}
bind:this={element}
/>
// scripts/vercel-images.js
import fs from 'node:fs';
import dotenv from 'dotenv'
dotenv.config()
// sizes should be shared variable with Image component
import {sizes} from './theme.cjs'
const file = '.vercel/output/config.json';
const config = {
...JSON.parse(fs.readFileSync(file, 'utf-8')),
images: {
"sizes": sizes,
"domains": [],
"minimumCacheTTL": 60,
"formats": ["image/avif", "image/webp"],
"remotePatterns": [
{
"hostname": "yourcdn\\.cloudfront\\.net$"
},
]
}
};
fs.writeFileSync(file, JSON.stringify(config, null, 2));
// package.json
"build": "vite build && node scripts/vercel-images",
@benmccann do you have working example hoisted anywhere (e.g. GitHub, Code sandbox etc. ). I wanted to see the configuration details. Thanks.
There's a draft PR for this: https://github.com/sveltejs/kit/pull/10323
Hi @hartwm, thanks so much for providing these examples! I do have one question I wonder if anyone knows the answer to.
Vite lets you import images like this:
<script>
import logo from '$images/example.png';
</script>
That means that a hash gets put in their name so that they can be cached forever. Is it possible to use such an image with the Vercel image solution?
When I go to https://vercel-images-sveltekit.vercel.app/ I notice that the local image can't be cached as strongly as the Vite processed images we have today. If you hit the page a second time, it returns a 304 response as opposed to avoiding the request entirely. I wonder if it's possible to use an imported image so that Vite processes the request before Vercel takes over possibly allowing us to skip the request entirely still. If it's possible, it'd be a great example to add to https://github.com/hartwm/vercel-images-sveltekit
@benmccann You can set Cache-Control headers with Vercel...
https://github.com/leoj3n/svelte-vercel-optimized-images?tab=readme-ov-file#avoiding-304-network-requests
It looks like you merged https://github.com/sveltejs/kit/pull/8667#pullrequestreview-1730990777 which relates to the OP... Considering the ability to set Cache-Control with Vercel, your remaining question may be answered, and so you might now be able to close this open issue.
@benmccann Regarding "Vite processes the request before Vercel takes over"...
It kind of defeats the purpose of using Vercel to send the correct image type (avif/webp) based on headers sent from the browser, as well as the (pre-defined) dynamic image sizes. If you circumvent Vercel you would have to generate all those permutations ahead of time yourself which could be prohibitive for user uploaded images, as mentioned.
I do believe Next.js implements their own code that emulates what the Vercel image optimization endpoint does, so it may be possible to add such functionality to SvelteKit itself, but you probably will have a very similar approach.
|
2025-04-01T06:40:31.841533
| 2021-04-06T17:01:49
|
851621816
|
{
"authors": [
"CaptainCodeman",
"evan-thurston",
"rmunn"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11045",
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/906"
}
|
gharchive/issue
|
Fetching works on initial page load, not on page reload
Describe the bug
I'm fetching data from an API in my route's load function, but I can't find a way to make it work on subsequential loads.
Logs
using ctx.fetch: TypeError: Failed to execute 'fetch' on 'Window': Illegal invocation at load
using fetch: ReferenceError: fetch is not defined at load
To Reproduce
install sveltekit
create a component that loads data in the <script> tag, as follows:
export async function load(ctx) { // Call our fetch let postsData; postsData = await fetch( 'https://jsonplaceholder.typicode.com/posts' ).then((x) => x.json()); console.log(postsData); }
Expected behavior
I expect the response to be logged to the console on each pageload.
Using fetch, it works on initial load, not reloads.
Using ctx.fetch, it works on reloads, but not inital load.
Stacktraces
Unfamiliar with stacktraces. This is my first time publishing an issue to a repo, so any information about this/how I can better format my issues would be much appreciated. I have more bugs/issues I've found that I'm too scared to post.
Information about your SvelteKit Installation:
Diagnostics
The output of npx envinfo --system --npmPackages svelte,@sveltejs/kit,vite --binaries --browsers
System:
OS: Windows 10 10.0.19042
CPU: (8) x64 Intel(R) Core(TM) i7-6700K CPU @ 4.00GHz
Memory: 8.92 GB / 15.89 GB
Binaries:
Node: 15.11.0 - C:\Program Files\nodejs\node.EXE
Yarn: 1.22.10 - ~\AppData\Roaming\npm\yarn.CMD
npm: 7.7.5 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: Spartan (44.19041.423.0), Chromium (89.0.774.68)
Internet Explorer: 11.0.19041.1
npmPackages:
@sveltejs/kit: next => 1.0.0-next.69
svelte: ^3.29.0 => 3.37.0
vite: ^2.1.0 => 2.1.5
Your browser
(Using NPM & Brave Web Browser)
Your adapter (e.g. Node, static, Vercel, Begin, etc...)
Node adapter
Severity
How severe an issue is this bug to you? Is this annoying, blocking some users, blocking an upgrade or blocking your usage of SvelteKit entirely?
I guess I could store content statically, but this is a big problem for me as I'm trying to use Sveltekit to refactor some of my projects that use APIs everywhere. Axios might work as a workaround, but I'm trying to use as many Svelte/Sveltekit features as possible because I'm a big believer in what you guys are doing. If I could figure this out & some other problems I'm having, I'm ready to start refactoring some of my projects to use Svelte/Sveltekit instead of react & related helped frameworks
Additional context
I love what you guys are doing & would love to help if I can. Svelte is a breath of fresh air after getting annoyed & leaving the development community years ago. Thank you so much for everything youre doing, even if you can't help me out
One thing to remember with the load function is that it can run on both the server OR the client.
So when you say you see it in the console, do you mean the terminal window where you are running npm run dev or the browser console?
With SSR enabled (the default) the load function is called on the server and the response embedded into the page so that it doesn't need to re-run on the client (but in dev, it does re-run). One the app has been loaded then navigating between routes will call the appropriate load functions for them but only on the client at that point. So, be sure to check both.
Minor tip: as you already have the function marked as async it's easier to use await than ye-olde promise chaining, so you could write:
export async function load(ctx) {
const resp = await fetch( 'https://jsonplaceholder.typicode.com/posts' )
const post = await resp.json()
return {
props: post
}
}
(assuming the route component has a property named post)
@evan-thurston You're forgetting to use the fetch parameter passed to your load function. Because load can run on the server or the client, but the built-in window.fetch function is only available on the client, Svelte-Kit includes a fetch function that will work in either environment, and passes it to your load function as a parameter. You're supposed to use that function rather than the built-in fetch. So change your load function to use ctx.fetch instead of fetch and it should work. Or rewrite it the way the documentation recommends (and the way CaptainCodeman showed in his example) using destructuring to get the fetch parameter, i.e. export async function load({ fetch }) and then you can just write await fetch('...') and you'll be using Svelte-Kit's supplied function.
|
2025-04-01T06:40:31.848147
| 2023-05-19T09:22:37
|
1716926676
|
{
"authors": [
"Destaq",
"geoffrich"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11046",
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/9984"
}
|
gharchive/issue
|
Component styles flicker/blink with SSR mode on
Describe the bug
When applying stylings to components or pages, on SSR mode, first server stylings get rendered, and then client ones.
In some cases, when components / pages use more 'dynamic' data like slots, $$restProps with class names, lists, etc., what oddly happens is that there is a 'flicker', where server stylings do not seem to match client ones.
Below is a video with two examples of when this happens:
https://github.com/sveltejs/kit/assets/61620873/e5688342-1c23-4994-a4ba-807fc1247e1b
Reproduction
About twenty lines: https://www.sveltelab.dev/vg1t4x2cb8n84f4
Logs
No response
System Info
I've spoken to a number of people so I don't think that it is a device-specific thing. But anyway:
System:
OS: macOS 13.3.1
CPU: (8) arm64 Apple M1
Memory: 83.23 MB / 16.00 GB
Shell: 5.9 - /bin/zsh
Binaries:
Node: 19.6.0 - /opt/homebrew/bin/node
Yarn: 1.22.11 - /opt/homebrew/bin/yarn
npm: 9.4.0 - /opt/homebrew/bin/npm
Browsers:
Brave Browser: <IP_ADDRESS>
Safari: 16.4
npmPackages:
@sveltejs/adapter-auto: ^1.0.0-next.91 => 1.0.0-next.91
@sveltejs/kit: next => 1.0.0-next.589
svelte: ^3.55.1 => 3.55.1
vite: ^4.0.0 => 4.1.1
Severity
annoyance
Additional Information
No response
The same issue happens even when setting export const csr = false. Interestingly, the only workaround seems to be wrapping for loops in a flexbox.
<div style="display:flex;">
{#each item as item}
<!-- represent the item details in a tailwind card -->
<div class="w-64 inline-block bg-pink">
<div>
<a href="/media/{item}">Page</a>
</div>
</div>
{/each}
</div>
But I would suggest this approach be avoided. When setting e.g. 5 divs (without the for loop), there are no spaces. Only through the above do they have spaces on the server-rendered HTML, and then revert to no spaces / gaps as expected (see repro). So this seems to be more of a workaround than expected behavior.
This issue is not caused by when styles are applied, but differences in the whitespace present in the SSR markup and the client-side rendered elements.
The SSR markup inserts a text node with "\n " between each child div:
But when client-side rendered, those nodes are removed:
There are multiple whitespace issues open in the Svelte repo, so not sure if this is a duplicate or not (or what workarounds would be). Transferring over there for now.
|
2025-04-01T06:40:31.935679
| 2022-12-08T16:33:33
|
1484994739
|
{
"authors": [
"ToP29",
"dominikg"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11047",
"repo": "sveltejs/vite-plugin-svelte",
"url": "https://github.com/sveltejs/vite-plugin-svelte/issues/531"
}
|
gharchive/issue
|
Inspector - open parent component
Describe the problem
If I have a button and i open it with inspector, it opens the button component, but usually I dont care about the button component, but about component where the button is used.
Describe the proposed solution
If i click on component with inspector, i could get list of parents for this component and i could click on one of them to open them instead. It would open parent component on line, where selected component is used.
Inspector could have current behavior on left click and this more advanced one could be on right click.
Alternatives considered
Importance
would make my life easier
you can use the up arrow key to select the parent component
@ToP29 does arrow selection help with your usecase?
Yes, it is ok. Thank you
|
2025-04-01T06:40:31.942119
| 2023-11-13T17:14:42
|
1991097478
|
{
"authors": [
"dominikg",
"sortofsleepy"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11048",
"repo": "sveltejs/vite-plugin-svelte",
"url": "https://github.com/sveltejs/vite-plugin-svelte/issues/792"
}
|
gharchive/issue
|
Svelte 5 - css nesting broke
Describe the bug
Not sure if this is the right spot to report, but since the error window mentions this plugin specifically, figured I'd start here.
Anyways, not sure if I'm doing something wrong but in the current stable version of Svelte, the following component compiles and functions as expected
<h1>Welcome to SvelteKit</h1>
<p>Visit <a href="https://kit.svelte.dev">kit.svelte.dev</a> to read the documentation</p>
<style>
h1 {
&:hover {
color:red;
}
}
</style>
However in the current Svelte 5 beta, an error is triggered saying Expected a valid CSS identifier. This is tested from a fresh install of SvelteKit running
npm create svelte@latest
afterwards running
npm install svelte@next
to install Svelte 5
Not sure how support was implemented in Svelte 4 but it'd be great to have this working in the final release without having to download another package.
Reproduction URL
https://github.com/sortofsleepy/svelte5-vite-css-error
Reproduction
Svelte 5 should already be setup in package.json so error should pop up immediately
Rerun npm i svelte@latest to compare and contrast.
Logs
[plugin:vite-plugin-svelte] C:/Users/Joseph/Documents/projects/tests/svelte-test/src/routes/+page.svelte:9:1 Expected a valid CSS identifier
C:/Users/Joseph/Documents/projects/tests/svelte-test/src/routes/+page.svelte:9:1
7 | color:red;
8 | }
9 | }
| ^
10 | </style>
CompileError: Expected a valid CSS identifier
at error (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/errors.js:568:8)
at read_identifier (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:491:3)
at read_selector (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:289:11)
at read_selector_list (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:157:17)
at read_rule (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:139:12)
at read_body (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:64:18)
at read_style (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:25:19)
at tag (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/state/element.js:264:20)
at new Parser (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/index.js:65:12)
at parse (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/ph
System Info
System:
OS: Windows 10 10.0.19045
CPU: (24) x64 AMD Ryzen 9 5900X 12-Core Processor
Memory: 16.00 GB / 31.90 GB
Binaries:
Node: 20.8.0 - C:\Program Files\nodejs\node.EXE
npm: 10.1.0 - C:\Program Files\nodejs\npm.CMD
pnpm: 6.32.3 - ~\AppData\Roaming\npm\pnpm.CMD
Browsers:
Edge: Spartan (44.19041.3570.0), Chromium (119.0.2151.58)
Internet Explorer: 11.0.19041.3570
native css nesting isn't supported in svelte4 by itself, some parts of it work but thats because svelte4 uses css-tree. svelte5 does not and hasn't implemented native css nesting yet.
To work around this, use a preprocessor that converts nested syntax to unnested. This can be done with postcss or lightningcss (experimental option on vite)
see https://github.com/sveltejs/svelte/issues/8587 and https://github.com/sveltejs/svelte/pull/9343
closing here as it is not caused by vite-plugin-svelte. Please follow the issues on svelte repo for updates.
|
2025-04-01T06:40:31.963607
| 2024-08-14T13:32:01
|
2465850589
|
{
"authors": [
"GregorySchwartz",
"svenssonaxel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11049",
"repo": "svenssonaxel/pdf-sign",
"url": "https://github.com/svenssonaxel/pdf-sign/issues/12"
}
|
gharchive/issue
|
Non-zero exit status 3
Using pdf-sign file.pdf, I receive the error below. I haven't run into this before the new version (maybe when pdftk was used?). Seems to be for only some files.
WARNING: file.pdf, object 52 0 at offset 395659: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 55 0 at offset 387012: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 58 0 at offset 378362: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf (object 10 0): object has offset 0
WARNING: file.pdf (object 15 0): object has offset 0
WARNING: file.pdf (object 414 0): object has offset 0
WARNING: file.pdf, object 52 0 at offset 395659: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 55 0 at offset 387012: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 58 0 at offset 378362: this widget annotation is not reachable from /AcroForm in the document catalog
qpdf: operation succeeded with warnings; resulting file may have some problems
Traceback (most recent call last):
File "/nix/store/ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj-pdf-sign-git/bin/.pdf-sign-wrapped", line 755, in <module>
main(parser.parse_args(sys.argv[1:] or ['-h']))
File "/nix/store/ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj-pdf-sign-git/bin/.pdf-sign-wrapped", line 38, in main
qpdfOrPdftk([
File "/nix/store/ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj-pdf-sign-git/bin/.pdf-sign-wrapped", line 487, in qpdfOrPdftk
subprocess.run(cmd, check=True)
File "/nix/store/7hnr99nxrd2aw6lghybqdmkckq60j6l9-python3-3.11.9/lib/python3.11/subprocess.py", line 571, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['qpdf', '--flatten-annotations=all', '--generate-appearances', 'file.pdf', '/tmp/tmpmu_7ot87/input.pdf']' returned non-zero exit status 3.
@GregorySchwartz Would be great if you could provide 1) An example problematic file without sensitive content and 2) The nix flake producing ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj or versions for pdf-sign and qpdf.
@GregorySchwartz Bump. An example file might be enough.
Unfortunately the only files I could reproduce this with have sensitive information...
@GregorySchwartz Could you please test branch dev?
Executing as before, you should see the same warnings, but no traceback. pdf-sign should print qpdf exited with code 3, indicating warnings. Use --continue-on-warnings to ignore. and exit with code 1.
Adding --continue-on-warnings to the invocation should cause pdf-sign to function normally, except the warnings should still be printed.
|
2025-04-01T06:40:31.999245
| 2016-01-07T01:07:16
|
125299736
|
{
"authors": [
"Ismael-VC",
"swadey"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11050",
"repo": "swadey/Lisp.jl",
"url": "https://github.com/swadey/Lisp.jl/issues/2"
}
|
gharchive/issue
|
MethodError: |(::Pattern, ::SubstitutionString{ASCIIString})
_
_ _ _(_)_ | A fresh approach to technical computing
(_) | (_) (_) | Documentation: http://docs.julialang.org
_ _ _| |_ __ _ | Type "?help" for help.
| | | | | | |/ _` | |
| | |_| | | | (_| | | Version 0.4.2 (2015-12-06 21:47 UTC)
_/ |\__'_|_|_|\__'_| | Official http://julialang.org/ release
|__/ | x86_64-w64-mingw32
julia> using Lisp
ERROR: LoadError: LoadError: MethodError: `|` has no method matching |(::ParserCombinator.Pattern, ::Base.SubstitutionString{ASCIIString})
Closest candidates are:
|(::Any, ::Any, ::Any, ::Any...)
|(::ParserCombinator.Matcher, ::ParserCombinator.Alt)
|(::ParserCombinator.Matcher, ::ParserCombinator.Matcher)
in include at boot.jl:261
in include_from_node1 at loading.jl:304
in include at boot.jl:261
in include_from_node1 at loading.jl:304
in require at loading.jl:243
while loading C:\Users\Peter\.julia\v0.4\Lisp\src\parser.jl, in expression starting on line 9
while loading C:\Users\Peter\.julia\v0.4\Lisp\src\Lisp.jl, in expression starting on line 2
julia>
This is the line: https://github.com/swadey/Lisp.jl/blob/master/src/parser.jl#L9
This was due to a change in ParserCombinator's API. Should be fixed now. I also updated everything to support 0.4 and higher.
Thank you very much!
|
2025-04-01T06:40:32.039279
| 2017-12-12T06:56:38
|
281277106
|
{
"authors": [
"duerrm",
"frantuma"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11051",
"repo": "swagger-api/swagger-core",
"url": "https://github.com/swagger-api/swagger-core/issues/2579"
}
|
gharchive/issue
|
"oneOf" property in @Schema annotation ignored in resource response
Q
A
Bug or feature request?
Unknown
Which Swagger-Core version?
v2.0.0-rc2
Which Java version?
1.8.0_131
Which JAX-RS framework & version?
2.0.1
I try to annotate my jaxrs class in order to generate an open api 3.0.0 spec that supports inheritance / polymorphism. For some reason I'm not able to generate a responses object that reflects the semantic of "oneOf" in order to show the user that different "subtypes" of the response type may be returned.
In detail, I want to achieve the following result
{
"openapi": "3.0.0",
"paths": {
"/test/bean": {
"get": {
"summary": "Test inheritance / polymorphism",
"operationId": "getBean",
"parameters": [
{
"name": "number",
"in": "query",
"description": "Test inheritance / polymorphism",
"required": true,
"schema": {
"type": "integer",
"format": "int32"
},
"example": 1
}
],
"responses": {
"200": {
"description": "bean answer",
"content": {
"application/json": {
"schema": {
"oneOf": [
{
"$ref": "#/components/schemas/MultipleSub1Bean"
},
{
"$ref": "#/components/schemas/MultipleSub2Bean"
}
]
}
}
}
}
}
}
}
},
"components": {
"schemas": {
"MultipleSub2Bean": {
"type": "object",
"properties": {
"d": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub2Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
},
"MultipleBaseBean": {
"type": "object",
"properties": {
"beanType": {
"type": "string"
},
"a": {
"type": "integer",
"format": "int32"
},
"b": {
"type": "string"
}
},
"description": "MultipleBaseBean",
"example": {
"a": 0,
"b": "test"
}
},
"MultipleSub1Bean": {
"type": "object",
"properties": {
"c": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub1Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
}
}
}
}
At present I use the following java code:
@Path("/test")
public class TestSwaggerApi {
@GET
@Path("/bean")
@Operation(summary = "Test inheritance / polymorphism",
responses = {
@ApiResponse(description = "bean answer",
responseCode = "200",
content = @Content(
mediaType = MediaType.APPLICATION_JSON,
schema = @Schema(
oneOf = { MultipleSub1Bean.class, MultipleSub2Bean.class }
)
)
)
})
@Produces({ MediaType.APPLICATION_JSON })
public MultipleBaseBean getBean(
@Context HttpServletRequest req,
@Parameter(description = "Test inheritance / polymorphism",
required = true,
example = "1")
@QueryParam("number") final int beanNumber) {
// do something...
...
}
}
@io.swagger.v3.oas.annotations.media.Schema(
description = "MultipleBaseBean",
subTypes = { MultipleSub1Bean.class, MultipleSub2Bean.class }
)
public class MultipleBaseBean {
public String beanType;
public int a;
public String b;
}
@io.swagger.v3.oas.annotations.media.Schema(
description = "MultipleSub1Bean"
)
public class MultipleSub1Bean extends MultipleBaseBean {
public int c;
}
@io.swagger.v3.oas.annotations.media.Schema(
description = "MultipleSub2Bean"
)
public class MultipleSub2Bean extends MultipleBaseBean {
public int d;
}
But the generated api holds not the expected responses object:
{
"openapi": "3.0.0",
"paths": {
"/test/bean": {
"get": {
"summary": "Test inheritance / polymorphism",
"operationId": "getBean",
"parameters": [
{
"name": "number",
"in": "query",
"description": "Test inheritance / polymorphism",
"required": true,
"schema": {
"type": "integer",
"format": "int32"
},
"example": 1
}
],
"responses": {
"200": {
"description": "bean answer",
"content": {
"application/json": {
"schema": {
"type": "string"
}
}
}
}
}
}
}
},
"components": {
"schemas": {
"MultipleSub2Bean": {
"type": "object",
"properties": {
"d": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub2Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
},
"MultipleBaseBean": {
"type": "object",
"properties": {
"beanType": {
"type": "string"
},
"a": {
"type": "integer",
"format": "int32"
},
"b": {
"type": "string"
}
},
"description": "MultipleBaseBean",
"example": {
"a": 0,
"b": "test"
}
},
"MultipleSub1Bean": {
"type": "object",
"properties": {
"c": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub1Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
}
}
}
}
I played around with the swagger test classes code, but I was not able to produce the requested result. I also tried to use the "implementation" property, but by doing so my "oneOf" property definition gets completely ignored.
Am I doing something wrong, or is this a missing feature?
It was indeed a missing feature / bug; correct support has been added in #2585, and is part of latest 2.0.0-SNAPSHOT version available on sonatype, and will be part of next release; closing ticket, please reopen if your issue isn't fixed.
Appreciate your work! Thank you very much!!!
|
2025-04-01T06:40:32.113701
| 2024-10-21T16:20:58
|
2603066083
|
{
"authors": [
"Kajalmehta29",
"swaraj-das"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11052",
"repo": "swaraj-das/Collect-your-GamingTools",
"url": "https://github.com/swaraj-das/Collect-your-GamingTools/pull/633"
}
|
gharchive/pull-request
|
Improved footer
Issue resolved: #630
Pull Request Format
PR Title
Issue #0 : ISSUE NAME solved
Type of PR
Add X in the box to specify the improvement type.
[X] Bug fix
[ ] Feature enhancement
[ ] Documentation update
[ ] Other (specify): ___________
Description
This pull request aims to [briefly describe what the pull request does, e.g., "add a new feature," "fix a bug," "improve documentation," etc.].
Screenshots / Videos (if applicable)
Before:
[Describe the state before the changes, e.g., "The section headings in the footer were misaligned, leading to an inconsistent and cluttered appearance."]
After:
[Describe the state after the changes, e.g., "The section headings in the footer are now properly aligned, providing a clean and professional look."]
Checklist
Add X in the box to specify.
[X] I have performed a self-review of my code.
[ ] I have tested the changes thoroughly before submitting this pull request.
[ ] I have provided relevant issue numbers, screenshots, and videos after making the changes.
[ ] I have commented my code, particularly in hard-to-understand areas.
Additional Context
[Provide any additional context about the changes, such as specific challenges you faced or decisions you made. This can help reviewers understand the rationale behind your changes.]
Thank you for reviewing my pull request!
@Kajalmehta29 Too poor footer. If you can make it better. Not need separate page for feedback.
|
2025-04-01T06:40:32.129330
| 2024-10-17T20:45:37
|
2595715716
|
{
"authors": [
"bethatasitmay",
"swaschkut"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11053",
"repo": "swaschkut/pan-os-php",
"url": "https://github.com/swaschkut/pan-os-php/issues/29"
}
|
gharchive/issue
|
exportToExcel getting Call to undefined method EDL::isGroup() on v2.1.27
Before we get into the details, a quick question: does v2.1.27 require a newer PHP version?
I'm still running PHP v8.3.11.
Also, my installed older version was showing as 2.1.25 but I think it was a bugfix dev release that rolled into v2.1.26 if I remember correctly (could be wrong).
Describe the bug
I ran a command that worked fine the last time I used it (then-current-release or possibly dev release v2.1.25 downloaded 2 Sep 2024), but after upgrading to v2.1.27 latest release today I get the error in the title.
Expected behavior
Normal output per previous versions.
Current behavior
> php -r "require_once 'C:/tools/pan/pan-os-php/utils/pan-os-php.php';" type=rule in=api://panorama.domain.com location=any ruletype=security 'actions=exportToExcel:DevQA-Rules.xls' 'filter=(src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)'
***********************************************
*********** pan-os-php.php type=rule UTILITY **************
- PAN-OS-PHP version: 2.1.27 [WIN] [8.3.11]
- Downloading config from API...
- Detected platform type is 'panorama'
- No 'template' provided so using default ='any'
- filter after sanitization : (src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)
- Loading configuration through PAN-OS-PHP library...
(0.97 seconds, 52.43 mb memory)
- PAN-OS-PHP APP-ID version: 8902-9003
- PAN-OS APP-ID version: 8904-9011
- PAN-OS AV version: 4974-5492
- PAN-OS WF version:<PHONE_NUMBER>90
- PAN-OS THREAT version: 8904-9011
- PAN-OS version: 110
- PAN-OS Device timezone: US/Pacific is used. actual time: 2024/10/17 13:09:44
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
* processing ruleset 'PanoramaConf: / RuleStore:Security' that holds 454 rules
PHP Fatal error: Uncaught Error: Call to undefined method EDL::isGroup() in C:\tools\pan\pan-os-php\lib\container-classes\AddressRuleContainer.php:555
Stack trace:
#0 C:\tools\pan\pan-os-php\lib\misc-classes\RQuery.php(199) : eval()'d code(1): AddressRuleContainer->hasObjectRecursive()
#1 C:\tools\pan\pan-os-php\lib\misc-classes\RQuery.php(199): eval()
#2 C:\tools\pan\pan-os-php\lib\misc-classes\RQuery.php(331): RQuery->matchSingleObject()
#3 C:\tools\pan\pan-os-php\utils\lib\RULEUTIL.php(371): RQuery->matchSingleObject()
#4 C:\tools\pan\pan-os-php\utils\lib\RULEUTIL.php(37): RULEUTIL->time_to_process_objects()
#5 C:\tools\pan\pan-os-php\utils\lib\UTIL.php(215): RULEUTIL->utilStart()
#6 C:\tools\pan\pan-os-php\lib\misc-classes\PH.php(1060): UTIL->__construct()
#7 C:\tools\pan\pan-os-php\utils\pan-os-php.php(118): PH::callPANOSPHP()
#8 Command line code(1): require_once('...')
#9 {main}
thrown in C:\tools\pan\pan-os-php\lib\container-classes\AddressRuleContainer.php on line 555
Possible solution
On v2.1.25 (again, not sure if this was the stable or dev release from my last issue report), I get this:
> php -r "require_once 'C:/tools/pan/pan-os-php/utils/pan-os-php.php';" type=rule in=api://panorama.domain.com location=any ruletype=security 'actions=exportToExcel:DevQA-Rules.xls' 'filter=(src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)'
***********************************************
*********** pan-os-php.php type=rule UTILITY **************
- PAN-OS-PHP version: 2.1.25 [WIN] [8.3.11]
- Downloading config from API...
- Detected platform type is 'panorama'
- No 'template' provided so using default ='any'
- filter after sanitization : (src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)
- Loading configuration through PAN-OS-PHP library...
(1.20 seconds, 52.43 mb memory)
- PAN-OS-PHP APP-ID version: 8867-8824
- PAN-OS APP-ID version: 8904-9011
- PAN-OS AV version: 4974-5492
- PAN-OS WF version:<PHONE_NUMBER>90
- PAN-OS THREAT version: 8904-9011
- PAN-OS version: 110
- PAN-OS Device timezone: US/Pacific is used. actual time: 2024/10/17 13:14:09
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
* processing ruleset 'PanoramaConf: / RuleStore:Security' that holds 454 rules
- object 'rule001' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'File-Sharing-01' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'File-Sharing-02' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule002' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule003' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule004' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule005' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule006' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule007' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule008' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule009' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
* objects processed in DG/Vsys '' : 11 filtered over 454 available
* processing ruleset 'PanoramaConf: / DeviceGroup:FW1 / RuleStore:Security' that holds 950 rules
- object 'rule010' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule011' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule012' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014b' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule015' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule016' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule017' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule020' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule022' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule023' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule024' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule026' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule027' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule028' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule029' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule030' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
* objects processed in DG/Vsys 'FW1' : 28 filtered over 950 available
* processing ruleset 'PanoramaConf: / DeviceGroup:FW2 / RuleStore:Security' that holds 1469 rules
- object 'rule011' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule010' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule015' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule016' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule012' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule029' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule030' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule017' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014b' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule020' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule023' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule024' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule026' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule031' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule028' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013-1' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule012b' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013-2' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'File-Sharing-01-T2' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
* objects processed in DG/Vsys 'FW2' : 30 filtered over 1469 available
- action 'exportToExcel' has tasks to process before shutdown.
**** PROCESSED 69 objects over 2873 available ****
************* END OF SCRIPT pan-os-php.php type=rule ************
Your Environment
Environment name and version: PowerShell v5.1
Operating System and version (desktop or mobile): Windows 10
Thanks for sharing this.
Sorry I did not saw your ticket.
This bug is now fixed in latest develop container, and it will be published in the next few week to latest container.
in your case please use:
https://github.com/swaschkut/pan-os-php/tree/develop_2-1-X
No worries - I just assumed you were busy or on holiday. Thanks for getting to it.
I'll check out the dev build in a few hours.
Sorry, I was doing some punishing PANOS upgrades that did not go well (nice 16 hour day). I'll check it out later today (Wednesday).
My apologies for the delay - I finally was able to test the dev build and it worked fine.
Thanks!
version 2.1.28 is now published and include this fix
#30
|
2025-04-01T06:40:32.140162
| 2024-05-31T12:04:01
|
2327689154
|
{
"authors": [
"CLAassistant",
"silverwind",
"yeliex"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11054",
"repo": "swc-project/swc-node",
"url": "https://github.com/swc-project/swc-node/pull/782"
}
|
gharchive/pull-request
|
doc: add @swc/core dependency to usage example
On a fresh repo, running the previous example would result in a Error: Cannot find module '@swc/core' otherwise.
I think it'd also be good to add @swc/core as a peer dependency to @swc-node/register, so that @swc/core will be installed automatically when missing.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
@swc/core in peerDependencies works as expected
I see it: https://github.com/swc-project/swc-node/blob/a1098be19a37077c419e42a60cf81889c5c94704/packages/register/package.json#L51
I guess my npm is not set up to automatically install peer dependencies.
Found the reason why npm did not install @swc/core: I had legacy-peer-deps=true in my .npmrc. Removing it made it install the peer dep.
|
2025-04-01T06:40:32.146447
| 2022-03-05T02:12:17
|
1160189902
|
{
"authors": [
"alexander-akait",
"clhiker",
"kdy1",
"kwonoj"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11055",
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/issues/3862"
}
|
gharchive/issue
|
Unicode character U+309B not accepted in identifiers
Describe the bug
Some unicode character not accepted in identifiers.
If using unicode character, it will get the error “”Invalid character in identifier"
If using encoded character, it will get the error: Unexpected character '゛'
Input code
var ゛; // U+309B
var \u309B;
var \u309C;
var a\u309B;
var a\u309C;
Config
{
"jsc": {
"parser": {
"syntax": "ecmascript"
},
"target": "es5"
},
"module": {
"type": "commonjs"
},
"isModule": false
}
Playground link
https://play.swc.rs/?version=1.2.148&code=H4sIAAAAAAAAAytLLFJ43DTbWkFfXyFU29jA0gkAY3O%2FZhIAAAA%3D&config=H4sIAAAAAAAAAy2LQQqAMAwE%2F5KzVy%2F9g48INUpL05YkgkX6dw142t0Z9oGsEcIDHUVJvOmohjcEoMioUVI3mAsYyknmWFff3ParkB9s9C8hNuZWs7pMuv36wKI0X3tLUmhnAAAA
https://play.swc.rs/?version=1.2.148&code=H4sIAAAAAAAAAytLLFKIKTU2sHSy5iqDsZ0h7ERkiUSoDAASLisXMQAAAA%3D%3D&config=H4sIAAAAAAAAAy2LQQqAMAwE%2F5KzVy%2F9g48INUpL05YkgkX6dw142t0Z9oGsEcIDHUVJvOmohjcEoMioUVI3mAsYyknmWFff3ParkB9s9C8hNuZWs7pMuv36wKI0X3tLUmhnAAAA
Expected behavior
SWC should properly parse these unicode as these code are valid identifiers.
Actual behavior
error: Unexpected character '゛'
--> input.js:1:5
|
1 | var ゛; // U+309B
| ^
Caused by:
0: failed to process js file
1: failed to process js file
2: Syntax Error
error: Invalid character in identifier
--> input.js:1:5
|
1 | var \u309B;
| ^^^^^^
Version
1.2.148
Additional context
I run swc use cmd as follows
npx swc --config-file=.swcrc poc.js -o poc.js
The environment is ubuntu 20.04 amd64.
Isn't this same as https://github.com/swc-project/swc/issues/3341 except charpoint?
@kwonoj Sounds like yes, maybe we can union them
It's an identical issue iff U+309B unicode id and not unicode xid.
I didn't close this as a duplicate as I need to check it.
I found https://github.com/Boshen/unicode-id
I'll patch it.
|
2025-04-01T06:40:32.155908
| 2021-11-04T09:53:28
|
1044536758
|
{
"authors": [
"CLAassistant",
"kdy1",
"magic-akari"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11056",
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/pull/2648"
}
|
gharchive/pull-request
|
fix(node-swc/types): add missing es6 config type(#2633)
Closes #2633
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Should I rebase it?
No, I rebased it.
|
2025-04-01T06:40:32.157830
| 2023-09-10T22:24:09
|
1889360731
|
{
"authors": [
"Austaras"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11057",
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/pull/7933"
}
|
gharchive/pull-request
|
feat(es/compat): impl pure_getters for optional_chaining
Description:
BREAKING CHANGE:
Related issue (if exists):
Closes #7921
I don't believe it would be easy. Both are referenced many times.
|
2025-04-01T06:40:32.197859
| 2023-01-18T01:52:48
|
1537277079
|
{
"authors": [
"adrian-kong",
"swiftnav-travis"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11058",
"repo": "swift-nav/swift-toolbox",
"url": "https://github.com/swift-nav/swift-toolbox/pull/914"
}
|
gharchive/pull-request
|
fixes branch ref
currently refs
ref: 912/merge
which is incorrect branch name.
Frontend and Release Workflow Started here
Frontend and Release Workflow Started here
|
2025-04-01T06:40:32.200017
| 2017-11-10T22:16:56
|
273076454
|
{
"authors": [
"hamin",
"swizzlr"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11059",
"repo": "swiftdocker/docker-swift",
"url": "https://github.com/swiftdocker/docker-swift/pull/95"
}
|
gharchive/pull-request
|
Use Dockerfile.template as root source of truth.
This commit takes the lead from
docker-node, separating
supported versions into their own folders, using Dockerfiles generated
from template. This commit closes #92 by backporting Clang 3.8 into 3.1.
I have also removed two soon to expire PGP keys from the file and added the new automation key.
@swizzlr i wanted to do this forever!!!! redis, ruby, and a bunch of other docker images are doing the same thing! I think i mentioned this in some discussion earlier Thank you!
I've smoke tested both 3 and 4 with two separate Swift 3 and 4 codebases respectively.
|
2025-04-01T06:40:32.205433
| 2024-10-18T21:19:58
|
2598404536
|
{
"authors": [
"ahoppen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11060",
"repo": "swiftlang/sourcekit-lsp",
"url": "https://github.com/swiftlang/sourcekit-lsp/issues/1771"
}
|
gharchive/issue
|
Tests can’t jump to generated interface of stdlib
I needed to disable a few tests in SwiftInterfaceTests because they can’t jump to generated interfaces of the stdlib. Investigate why.
Synced to Apple’s issue tracker as rdar://138210215
|
2025-04-01T06:40:32.207829
| 2024-11-18T20:46:42
|
2669873402
|
{
"authors": [
"Ernest0-Production",
"dschaefer2"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11061",
"repo": "swiftlang/swift-package-manager",
"url": "https://github.com/swiftlang/swift-package-manager/issues/8123"
}
|
gharchive/issue
|
Long cache retrieval when resolving dependencies
Description
SPM dependency resolution takes a very long time to extract. Each packet takes on average ~1 second to extract.
As I understand it, the problem is that the cache is stored in a compressed form and takes time to extract. But it seems to me that this is of less practical use than a faster resolution.
I would like the user cache to store the original artifacts (the same as in the local .build directory)
I assume you mean "package" not "packet". But yes, it's actually doing a git clone and checkout from the cache which is a very expensive operation.
I think we're probably at the point where we're doing the best we can. SwiftPM does a complete git clone into the user cache under .swiftpm/cache/repositories. We need that so we have all the versions available during dependency resolution and allows you to have multiple projects using different versions. We then do a shallow clone of the selected revision over to the package scratch directory. One second to do is about as fast as git can go.
The solution we're hoping for is the previous work with package registries where the registry would have the metadata we need and a source archive for each available version. Then we can cache those which would be much faster. But I'm not sure when a community wide solution for that will happen.
|
2025-04-01T06:40:32.211447
| 2024-09-24T11:54:30
|
2545186033
|
{
"authors": [
"grynspan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11062",
"repo": "swiftlang/swift-testing",
"url": "https://github.com/swiftlang/swift-testing/pull/727"
}
|
gharchive/pull-request
|
[DO NOT MERGE] Check the codepage used on Windows in Jenkins
[One line description of your change]
Motivation:
[Explain here the context, and why you're making that change. What is the problem you're trying to solve.]
Modifications:
[Describe the modifications you've done.]
Result:
[After your change, what will change.]
Checklist:
[ ] Code and documentation should follow the style of the Style Guide.
[ ] If public symbols are renamed or modified, DocC references should be updated.
@swift-ci test Windows
@swift-ci test
*** CODEPAGE IN CI: 1252 -- CONSOLE: 65001
|
2025-04-01T06:40:32.212323
| 2024-07-20T14:15:56
|
2420916565
|
{
"authors": [
"hamishknight"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11063",
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/75384"
}
|
gharchive/pull-request
|
[Macros] Remove SwiftOperators dependency
Looks like this got missed by #74502
@swift-ci please test
|
2025-04-01T06:40:32.213791
| 2024-08-26T03:24:46
|
2485777492
|
{
"authors": [
"hborla"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11064",
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/76079"
}
|
gharchive/pull-request
|
[Concurrency] Add missing Sendable conformances to nested value types in Async{Throwing}Stream.
Termination, BufferingPolicy, and YieldResult all have trivial conformances to Sendable. The conformance on YieldResult is conditional on Element: Sendable because one of the enum cases stores an element in its associated value.
@swift-ci please smoke test
|
2025-04-01T06:40:32.214996
| 2024-09-05T16:52:35
|
2508296069
|
{
"authors": [
"cachemeifyoucan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11065",
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/76284"
}
|
gharchive/pull-request
|
scan-deps-deterministic-check-windows
Test commit to see the test failure on windows.
@swift-ci please smoke test windows platform
Closing as I have an idea what is going on.
|
2025-04-01T06:40:32.221440
| 2024-09-10T02:29:50
|
2515264450
|
{
"authors": [
"DmT021",
"DougGregor"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11066",
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/76363"
}
|
gharchive/pull-request
|
[Diagnostics] Add -print-diagnostic-groups flag
This change adds the -print-diagnostic-groups flag as described by SE-0443.
@DougGregor please take a look
I've also added the flag to many relevant tests. This way we test that the messages we're getting are expected and belong to the expected group. IMO it's nice to have some extra coverage so we can be sure we don't change the group of a warning accidentally. But let me know if you think otherwise.
Also, there's a use of DiagnosticOptions and DiagnosticEngine in LLDB so I've opened a PR there as well https://github.com/swiftlang/llvm-project/pull/9240
Only a small comment and a question, this looks good!
@swift-ci please smoke test
@DougGregor Since there are changes in the public members of DiagnosticOptions and DiagnosticEngine LLDB won't build without this patch https://github.com/swiftlang/llvm-project/pull/9240
I don't know how to deal with it, both of these PRs depend on each other and should be tested and merged together.
Can you rerun the tests here with that PR?
https://github.com/swiftlang/llvm-project/pull/9240
@swift-ci please smoke test
I don't know how to deal with it, both of these PRs depend on each other and should be tested and merged together.
Can you rerun the tests here with that PR?
Yes, of course. I've kicked off the multi-PR test, sorry I missed that earlier.
@DougGregor The windows tests failed because of the missing quote marks in ucrt.swift. https://github.com/swiftlang/swift/pull/76363/files#diff-b68568b1ae0d461890b6f117d611184202d92a42af3a9caad91e666d6e542415
https://github.com/swiftlang/llvm-project/pull/9240
@swift-ci please smoke test
|
2025-04-01T06:40:32.223341
| 2024-06-12T13:58:33
|
2348852925
|
{
"authors": [
"award999"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11067",
"repo": "swiftlang/vscode-swift",
"url": "https://github.com/swiftlang/vscode-swift/pull/886"
}
|
gharchive/pull-request
|
Always register the swift-lldb debug adapter provider
Issue: #848
We can always register this and that way the user does not need to reload the window when use change to using lldb-dap. Now they can use the updated debug configurations right away
@adam-fowler the debug session does nothing if the path does not exist or cannot find in toolchain, so added a check to verify that the the debug adapter does exist so the user gets an error message instead of silently failing. I see that we are already listening to changes to swift.debugger.useDebugAdapterFromToolchain and swift.debugger.path settings, else where so seemed a little more complex than necessary to listen to this again to register the debug provider
|
2025-04-01T06:40:32.231787
| 2017-12-23T10:36:44
|
284299670
|
{
"authors": [
"Rotzbua",
"cech12"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11068",
"repo": "swiftmailer/swiftmailer",
"url": "https://github.com/swiftmailer/swiftmailer/issues/1033"
}
|
gharchive/issue
|
Brackets in SetFrom are missing
Q
A
Bug report?
yes
Feature request?
no
RFC?
no
How used?
in Yii 2
Swiftmailer version
5.4.3
PHP version
7.1.1
Observed behaviour
<EMAIL_ADDRESS>=> "Name (Something in Brackets) xxx"])
generates this header information:
From: Name (Something in Brackets) xxx<EMAIL_ADDRESS>But the received mail has this header information:
From: Name
xxx<EMAIL_ADDRESS>
The brackets are replaced with a line break.
Expected behaviour
The received message should have this header information:
From: Name (Something in Brackets) xxx<EMAIL_ADDRESS>or
From: "Name (Something in Brackets) xxx"<EMAIL_ADDRESS>Example
Yii::$app->mailer->setTextBody("This is a test mail.")
<EMAIL_ADDRESS>=> "Name (Something in Brackets) xxx"])
<EMAIL_ADDRESS> ->setSubject('test mail')
->setHtmlBody('<b>This is a test mail.</b>')
->send();
@cech12 Please add which transport you used. (Maybe the brackets filtered for security reasons.)
@Rotzbua Sorry, but I don't know, what you mean with "transport". :/
I tested the normal php mail function on 2 different web servers
<EMAIL_ADDRESS>"test mail", "This is a test mail.", "From: Name (Something in Brackets) xxx <name@mail.com>");
and the result is the same.
The received message had this header information:
From: Name
xxx<EMAIL_ADDRESS>
Maybe there are some security mechanisms. How can I test it?
I asked for the sending method: mail(), sendmail or smtp. Swiftm. calls them Swift_SmtpTransport, Swift_SendmailTransport, Swift_MailTransport.
I suggest your swiftmailer use phps mail(), because it shows the same behaviour. mail() is just a proxy function in php. It is processed very different on different computer systems. That's the reason why Swift_MailTransport is removed in v6.
Try smtp?
By the way: I tested your code on my webspace (php7.0) and the brackets are not removed. Seems that my provider has a different setup than your provider.
Ah okay! Yes I use the mail transport Swift_MailTransport.
If this function is removed in the new swiftmailer version, in the future nobody will have this problem. I reached that I wanted to reach. :D
I will try the SMTP mail transport and if there is a Problem I will open a new Issue.
So this Issue can be closed. :)
Thanks for your help! :)
|
2025-04-01T06:40:32.276875
| 2023-05-26T12:44:33
|
1727559068
|
{
"authors": [
"chainofhonor",
"cwilvx",
"mungai-njoroge",
"tokisak1kurum1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11069",
"repo": "swing-opensource/swingmusic",
"url": "https://github.com/swing-opensource/swingmusic/issues/126"
}
|
gharchive/issue
|
hope to support lyrics
my music folder like this:
/music/oldsong/song1-singer.mp3
/music/lightsong/songb-singer.mp3
/lyrics/oldsong/song1-singer.lrc
/lyrics/lightsong/songb-singer.lrc
I hope to be able to display the lyrics on the web interface
Hello @chainofhonor
The development of this project is currently on hold. (see pinned issue). However, lyric support is a planned feature.
Hello @chainofhonor
Lyrics supported has been added with v1.4.0 release.
Go To Release Page
I would really appreciate it if support for lyrics embedded in the music files could be provided.
@tokisak1kurum1
That does not work? I thought we had that. Maybe it's broken or something. I'll investigate that.
|
2025-04-01T06:40:32.279481
| 2015-07-30T17:48:53
|
98225005
|
{
"authors": [
"tlunter",
"y3ddet"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11070",
"repo": "swipely/docker-api",
"url": "https://github.com/swipely/docker-api/pull/299"
}
|
gharchive/pull-request
|
add user option to list passed along to exec call
Parameter is "user" with the value equal to the in-container username.
Example:
docker_obj = Docker::Container.get("mycontainer")
cmd = { "/bin/sh", "ls", "/" }
opts = { :user => "nobody" }
docker_obj.exec(cmd,opts)
The Docker engine will use the default value for user if the parameter provided is an empty string, so this should be safe. Tested against Docker 1.7.1 (boot2docker).
@y3ddet this has been released in v1.22.2
|
2025-04-01T06:40:32.282222
| 2024-10-16T00:24:26
|
2590239450
|
{
"authors": [
"SgiobairOg",
"sabberworm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11071",
"repo": "swisscom/JCR-Hopper",
"url": "https://github.com/swisscom/JCR-Hopper/pull/1"
}
|
gharchive/pull-request
|
Initial version
Includes editor and runner
Todo:
[ ] Documentation
[ ] Checks
Should we have ./gradlew check run on Github Actions or another runner?
|
2025-04-01T06:40:32.285147
| 2024-04-22T14:33:36
|
2256703291
|
{
"authors": [
"gfellerph",
"imagoiq"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11072",
"repo": "swisspost/design-system",
"url": "https://github.com/swisspost/design-system/issues/2990"
}
|
gharchive/issue
|
Spinner never stops in overlay of breadcrumb-buttons
When clicking on a breadcrumb-button ("Helps", "Contact"), a spinner is shown, but it has no condition to automatically be removed.
@imagoiq please re-check
|
2025-04-01T06:40:32.294680
| 2015-06-20T20:45:25
|
89816101
|
{
"authors": [
"blimmer",
"rwjblue",
"trabus"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11073",
"repo": "switchfly/ember-cli-mocha",
"url": "https://github.com/switchfly/ember-cli-mocha/issues/58"
}
|
gharchive/issue
|
Bring component integration test in line with ember-cli
This PR to ember-cli will likely be done soon and has a few differences from our implementation:
It generates a component integration test by default when generating a new component (and allows generating a unit test by specifying --unit)
It has a different argument (they use --test-type)
availableOptions: [
{
name: 'test-type',
type: ['integration', 'unit'],
default: 'integration',
aliases:[
{ 'i': 'integration'},
{ 'u': 'unit'},
{ 'integration': 'integration' },
{ 'unit': 'unit' }
]
}
]
Once the ember-cli PR is merged I propose we bring this project back in sync. I will gladly submit a PR for this change.
Thanks for keeping track and working on this @blimmer!
:+1:
@blimmer Thanks, and sorry for any crossed wires!
No problem. Happy to see this taking priority!
The dependency branch was merged this morning. This can be worked whenever - I might be able to get to it this week(end)
@blimmer - Thank you!
|
2025-04-01T06:40:32.327643
| 2017-11-16T10:36:40
|
274468253
|
{
"authors": [
"facontidavide",
"malban"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11074",
"repo": "swri-robotics/mapviz",
"url": "https://github.com/swri-robotics/mapviz/pull/543"
}
|
gharchive/pull-request
|
Added "keep image ratio" to Image plugin
Hi,
I believe this PR would help making the Image Plugin more usable.
Added "Keep original image ratio" to the Widget. When checked, the height is automatically adjust, both in "percent" and "pixel" mode. The user should only modify the Width whilst the Height is updated accordingly.
When changing the mode from "percent" to "pixel" and vice versa, the image size is preserved.
To be sure that no resize is introduced in step 2), I had to change the width and height type and QSpinBox to double and QDoubleSpinBox, and percend must have 1 decimal.
Cheers
https://vimeo.com/243268373
Does anyone at SWRI ever say "thanks for contributing"?
Some of us do at least: https://github.com/swri-robotics/mapviz/pull/539#issuecomment-345371749 https://github.com/swri-robotics/mapviz/pull/525#issuecomment-342520534
But regardless, thanks for contributing, we appreciate it and are glad you find the tool useful.
|
2025-04-01T06:40:32.332903
| 2020-04-06T07:18:23
|
594864229
|
{
"authors": [
"Kyungpyo-Kim",
"cbradyas",
"pjreed",
"zang09"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11075",
"repo": "swri-robotics/novatel_gps_driver",
"url": "https://github.com/swri-robotics/novatel_gps_driver/issues/84"
}
|
gharchive/issue
|
boost::thread sleep not working
I update my ros driver, after that this driver not publishing any ros topics!
Then, I figure out that while loop in Spin() function execute only first time.
So I change the code:
(+) ros::Rate loop_rate(1000);
while(gps_.IsConnected() && ros::ok()) {
(-) boost::this_thread::sleep(boost::posix_time::microseconds(1));
(+) loop_rate.sleep();
}
Then it finally publish the topics again!
Please fix this errors.
When you say it only executes the first time, what happens after that? Is something throwing an exception, or is it blocking somewhere? Has anybody else seen this happen?
It seems especially odd that using ros::Rate there would make a difference; one of the differences in using a ROS timer is that it follows ROS's clock when running in a simulation, and, if anything, I would expect that to cause problems since the timer will stop running if the simulation is paused. Normally that's not an issue for hardware drivers since you don't usually hardware drivers in simulation, though.
In my case after enter the command like this,
$ sudo apt-get update && sudo apt-get upgrade
I can see same problem in all my laptop and mini PC(NUC)
When I debugging the code, there was no execption but the code is block at below position.
-> boost::this_thread::sleep(boost::posix_time::microseconds(1));
Then, I launch the driver in real situation not with simulation.
My computer enviornment is: Ubuntu 18.04 && ROS Melodic
We also had this problem. I made that change and it fixed it.
I also got same problem.
My computer environment is: Ubuntu 18.04 && ROS Melodic
How strange, I had never seen this problem before, but I tested it on a different computer and now I see it there. Well, I merged in a change that should fix it everywhere. Thanks!
|
2025-04-01T06:40:32.353170
| 2024-01-16T11:16:10
|
2083685196
|
{
"authors": [
"Akmadan23",
"sxyazi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11076",
"repo": "sxyazi/yazi",
"url": "https://github.com/sxyazi/yazi/pull/523"
}
|
gharchive/pull-request
|
feat: add option to enter directory on open command
This is another nice feature to have in my opinion, and it's something that almost any TUI file manager has.
That's intentional. See https://yazi-rs.github.io/docs/faq#why-cant-open-and-enter-be-a-single-command
You can use this Smart enter tip to impl it.
hat's intentional. See https://yazi-rs.github.io/docs/faq#why-cant-open-and-enter-be-a-single-command
You can use this Smart enter tip to impl it.
Oh, I see, I did not know that. But why can't we just add this simple flag to achieve it?
I want to keep it concise - one command should only do one thing.
Opening and entering are not essentially the same, and if we add it, we must also consider whether "entering should support opening" or "opening should support entering". In the end, we might need to add them for both commands.
Opening and entering are not essentially the same, and if we add it, we must also consider whether "entering should support opening" (many people want l to support opening, but l is by default bound to the enter command) or "opening should support entering". In the end, we might need to add them for both commands.
Well, I actually tried to implement enter --or-open before, but since enter is a member of Tab it was really unpractical to access open, which is a Manager member, so I did it the other way around.
As for the l binding, we could add a note in enter's documentation to use open --or-enter-dir for that use case...
Anyways, I understand your point here. Feel free to close PR if you don't see any value in this addition.
Okay let me close it, sorry.
|
2025-04-01T06:40:32.358919
| 2024-10-29T19:16:37
|
2622120878
|
{
"authors": [
"Kongedam",
"Ltek",
"karlsnyder0",
"syd711"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11077",
"repo": "syd711/vpin-studio",
"url": "https://github.com/syd711/vpin-studio/issues/574"
}
|
gharchive/issue
|
Add WOL support
Add Wake-On-LAN support for VPin Studio hosts that go into sleep / stand by mode on a LAN.
Should work both from Windows and macOS Vpin Studio clients.
Should work for hosts configured in the connection launcher or with previously discovered clients via. broadcast.
nice idea!
I'd love to see Wake-on-LAN functionality added to VPin Studio in a future update.
My cabinet is located in my garage, and it would be incredibly convenient to be able to wake it up remotely using VPin Studio itself. Currently, I have to rely on a separate program to do this, which adds an extra step to my setup.
Integrating Wake-on-LAN would streamline the process and make VPin Studio an even more comprehensive solution for managing my virtual pinball setup.
Thanks for considering this request!
@karlsnyder0 I close this issue as fixed.
Let's create new tickets if there are follow-ups.
|
2025-04-01T06:40:32.795851
| 2020-10-19T02:42:49
|
724214768
|
{
"authors": [
"synfinatic",
"tiehfood"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11097",
"repo": "synfinatic/udp-proxy-2020",
"url": "https://github.com/synfinatic/udp-proxy-2020/issues/40"
}
|
gharchive/issue
|
Add --daemon flag / pid file
Would be nice if udp-proxy-2020 supported running as a daemon and writing a pid file to /var/run/udp-proxy-2020.pid or other user defined path.
I successfully build your site-to-site version for the Unifiy UDM and it works very well 👍 . A daemonized version as you suggested would be a really nice feature
FWIW, I have no idea how the UDM starts services on boot, but pretty much everything nowadays (upstart, systemd) don't require a "daemon" mode. If you do figure it out, i'd be curious to hear how you did it so I can document it for others.
Quick note: Apparently daemonizing go apps is non-traditional because of how go works. But this library seems the most popular way of doing so: https://github.com/sevlyar/go-daemon
As far as I know there's no easy way to run a persistent daemon on the UDM/UDMP besides a more or less hacky way abusing the way package installations are handled on these machines. But for my purpose it would be sufficient enough to start the daemon by hand every time the UDM/UDMP reboots.
So my idea: ssh into the UDM/UDMP and start your tool as a daemon :)
Unfortunately I have no experience in programming with go
Sadly the UDM series is really limited right now. That said, it is just linux and so you should be able to write a startup script to do it for you.
Yeah, got it running. Thanks again for the great tool :)
|
2025-04-01T06:40:32.803693
| 2021-11-16T22:44:06
|
1055502834
|
{
"authors": [
"Jasonvdb",
"john-zaprite",
"kiwiidb",
"rbndg"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11098",
"repo": "synonymdev/blocktank-client",
"url": "https://github.com/synonymdev/blocktank-client/issues/8"
}
|
gharchive/issue
|
"Local" / "Remote" is switched from the perspective of the buyer.
Hey there, just bought a channel on testnet using this API.
It seems like local_balance and remote_balance should be switched: right now it's from the perspective of the server, it makes more sense to have it from the POV of the client IMO.
Hey, yeah this makes sense. We'll probably update this server side and post back here once done. Thanks!
Did this change get implemented yet?
Seems like local_balance refers to Synonym's node. Is this expected to change, or should we assume it will stay as-is?
we will change it eventually, most likely we will use spending and receiving as the words
|
2025-04-01T06:40:32.813530
| 2021-11-22T23:29:44
|
1060689653
|
{
"authors": [
"ChristianMurphy",
"wooorm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11099",
"repo": "syntax-tree/mdast-util-to-markdown",
"url": "https://github.com/syntax-tree/mdast-util-to-markdown/issues/47"
}
|
gharchive/issue
|
roundtrippping content causes definition to disappear
Initial checklist
[X] I read the support docs
[X] I read the contributing guide
[X] I agree to follow the code of conduct
[X] I searched issues and couldn’t find anything (or linked relevant results below)
Affected packages and versions
1.2.4
Link to runnable example
https://stackblitz.com/edit/node-st1b3g?file=index.js
Steps to reproduce
Run
import { remark as remarFactory } from 'remark';
import { visit } from 'unist-util-visit';
const remark = remarFactory();
// clean extra attributes that make it hard to see issue
function scrubber(tree) {
visit(tree, function (node) {
node.value = undefined;
node.position = undefined;
node.spread = undefined;
node.lang = undefined;
node.identifier = undefined;
node.label = undefined;
node.title = undefined;
node.url = undefined;
});
}
const content = `[a]: `;
const originalAst = remark.parse(content);
const newContent = remark.stringify(originalAst);
const newAst = remark.parse(newContent);
scrubber(originalAst);
scrubber(newAst);
console.log(JSON.stringify(originalAst, null, 4));
console.log(JSON.stringify(newAst, null, 4));
result is:
{
"type": "root",
"children": [
{
"type": "definition"
}
]
}
then
{
"type": "root",
"children": [
{
"type": "paragraph",
"children": [
{
"type": "text"
}
]
}
]
}
and the output text is
[a]:
Expected behavior
definition content is preserved
Actual behavior
definition content disapepars
Runtime
Node v16
Package manager
npm v7
OS
Linux
Build and bundle tools
No response
also happens with
[a]
[a]: 
https://spec.commonmark.org/dingus/?text=[a]
[a]%3A %26%23xc%3B
released!
|
2025-04-01T06:40:32.817099
| 2023-02-13T11:13:37
|
1582152321
|
{
"authors": [
"dstallenberg"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11100",
"repo": "syntest-framework/syntest-core",
"url": "https://github.com/syntest-framework/syntest-core/pull/210"
}
|
gharchive/pull-request
|
Create 'create-plugin-template' command
Closes #166
We should create a github template which includes the ci scripts.
The create-template command then is used to create a specific plugin template.
Superseded by #245
|
2025-04-01T06:40:32.850373
| 2024-11-08T18:11:20
|
2644790625
|
{
"authors": [
"guettli",
"janiskemper"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11101",
"repo": "syself/cluster-api-provider-hetzner",
"url": "https://github.com/syself/cluster-api-provider-hetzner/issues/1508"
}
|
gharchive/issue
|
HetznerCluster does not react on changes in relevant secrets
/kind bug
What steps did you take and what happened:
The "HetznerCluster" object should listen to the changes in the secret that stores the Hetzner credentials. For example, because they have to be synced in the updated form to the workload clusters. This doesn't happen right now because the controller doesn't react on the relevant events.
In https://github.com/syself/cluster-api-provider-hetzner/blob/1904cabf66bdede229603836230edb7a5e29424b/controllers/hetznercluster_controller.go#L400 we "acquire" the secret but set "controlledByOwner" to false. This means that the hetznercluster-controller doesn't actually own the secret, so that the events are not shown, even though we set the event listener here: https://github.com/syself/cluster-api-provider-hetzner/blob/1904cabf66bdede229603836230edb7a5e29424b/controllers/hetznercluster_controller.go#L749
What did you expect to happen:
We should react on the events in the hetzner secret. Either by setting the hetznercluster controller as "controller" of the secret, or by changing the way we listen to events.
It would be obviously easier to set the hetznercluster as controller of the secret, that's just one value. I'm not aware if that has any drawbacks compared to the current state.
@janiskemper
if a controller owns a secret and you delete the controller, Kubernetes garbage collection (GC) will typically remove the secret.
Somehow I think an owneRef does not match. The user is responsible for that secret.
We can use an option for .Own():
From the Owns docstring:
// The default behavior reconciles only the first controller-type OwnerReference of the given type.
// Use Owns(object, builder.MatchEveryOwner) to reconcile all owners.
I suggest to use MatchEveryOwner
|
2025-04-01T06:40:32.894232
| 2021-11-04T16:16:34
|
1044941728
|
{
"authors": [
"DarkneSsDz",
"sysrom"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11102",
"repo": "sysrom/DcRatCHS",
"url": "https://github.com/sysrom/DcRatCHS/issues/1"
}
|
gharchive/issue
|
HRDP and Reverse Proxy
HRDP and Reverse Proxy not found in form
The original author did not set poxy。。。
working on to add them by myself he was rude in his answers last time so better continue his work and develop it by myself i already added reverse proxy and i am working on hrdp :)
OK
|
2025-04-01T06:40:33.246251
| 2015-09-02T20:03:58
|
104560732
|
{
"authors": [
"awalGarg",
"guybedford"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11103",
"repo": "systemjs/builder",
"url": "https://github.com/systemjs/builder/issues/299"
}
|
gharchive/issue
|
Support SystemJS builder dependency extraction without execution
We currently execute transpiled ES6 in the target environment to extract the dependencies, but we don't use the execute function at all, so this is unnecessary work.
In addition, not execution would allow transpiling to ES6 itself say just using the modules transformer when transpiling for browser environments that support more than Node does (previously discussed in https://github.com/jspm/jspm-cli/issues/884).
This can be done by having a truncated instantiate for the builder, and just applying a regex in the tracer to extract the System.register dependencies since we can rely on it being the first line of code after comments and meta strings.
This would be pretty awesome. Any updates? What parts of the code should potential contributors be looking at to help out?
This isn't a priority currently, and may be somewhat tricky for contributions. Such an intercept would be at https://github.com/systemjs/builder/blob/master/lib/trace.js#L284 skipping instantiate. It would be important to ensure any useful functions of instantiate are replicated in this case (such as concatting the dependencies with load.metadata.deps).
Released in 0.15.0.
|
2025-04-01T06:40:33.271810
| 2020-04-10T15:54:27
|
597960097
|
{
"authors": [
"szgabsz91"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11104",
"repo": "szgabsz91/iit-szg",
"url": "https://github.com/szgabsz91/iit-szg/issues/85"
}
|
gharchive/issue
|
Integrate Lighthouse analysis
Lighthouse should be integrated into CI.
First, let's save the initial analysis report, then create tickets for optimization tasks.
Ideas:
Lazy loading
Lazy loading images
RAIL
etc.
Lighthouse is integrated
|
2025-04-01T06:40:33.311893
| 2024-12-07T19:46:12
|
2724788330
|
{
"authors": [
"DMedina559",
"bassrock",
"t0bst4r"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11105",
"repo": "t0bst4r/home-assistant-matter-hub",
"url": "https://github.com/t0bst4r/home-assistant-matter-hub/issues/271"
}
|
gharchive/issue
|
[Bug]: Climate unnecessarily converting fahrenheit
Version
3.0.0-alpha.51
Matter Controller
Apple Home
Steps to reproduce
I have an Ecobee connected to HomeAssistant via the HomeKitBridge connection, I am then trying to export it via the Matter Hub add-on. When it shows up in the Home App, all the temperatures are around 162 degrees. My system currently uses fahrenheit for its units. It looks like the add on is converting the fahrenheit units into celcius
State and attributes
hvac_modes: off, heat, cool, heat_cool
min_temp: 45
max_temp: 92
fan_modes: on, auto
friendly_name: Bedrooms
supported_features: 395
current_temperature: 74
temperature: 74
target_temp_high: null
target_temp_low: null
current_humidity: 38
fan_mode: auto
hvac_action: idle
Relevant log output
No response
Documentation & Issues
[X] I have reviewed the documentation and the linked troubleshooting guide.
[X] I have searched the issues for a similar problem.
https://github.com/t0bst4r/home-assistant-matter-hub/discussions/261
This will be handled as part of a larger refactoring of climates (see #261)
|
2025-04-01T06:40:33.468992
| 2022-11-23T22:01:37
|
1462461589
|
{
"authors": [
"jbaird123",
"jeannekitchens",
"jgoodell2"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11106",
"repo": "t3-innovation-network/desm",
"url": "https://github.com/t3-innovation-network/desm/issues/352"
}
|
gharchive/issue
|
Spine terms added but then mapped terms disappear
After adding spine terms and then save. Returning to the mapping shows spine terms saved but the mapping is gone.
...this prevents clicking of "Done Alignment"
@jbaird123 @excelsior Jim is mapping CEDS.
@jgoodell2 - I'm having trouble duplicating this problem. Can you provide exact instructions including the file you're uploading to duplicate the issue? A screencast would be helpful if you can do it.
@excelsior - The issue here is that the elements in the screenshot above were added as synthetic elements and automatically mapped as "Identical". After saving, the mapping was lost, but the spine term remains. We went back in and tried to manually map the elements, and they do not get saved.
Resolved. Please reopen if the issue persists.
Yes. I confirm that this issue is resolved.
|
2025-04-01T06:40:33.475098
| 2023-01-19T02:56:51
|
1548439975
|
{
"authors": [
"leerob",
"nayaabkhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11107",
"repo": "t3-oss/create-t3-app",
"url": "https://github.com/t3-oss/create-t3-app/issues/1096"
}
|
gharchive/issue
|
bug: Starting with create-t3-app and trying app leads to 404
Provide environment information
N/A
Describe the bug
Hey, just a heads up – there is currently an open bug in Next.js for folks trying out the app directory (beta) as follows: if you are using i18n configuration inside of next.config.js, and the app directory, you will see a 404 when routing to your index route.
This is a bug. The i18n routing configuration inside next.config.js is not being ported to app. You can view the documentation for i18n routing here. To make sure you can incrementally adopt from pages -> app, or to have them coexist for awhile, we will be fixing this bug.
But at least for now, I just wanted to make y'all aware.
Reproduction repo
N/A
To reproduce
Stated above.
Additional information
No response
Related issue: https://github.com/vercel/next.js/issues/41980
To make sure you can incrementally adopt from pages -> app, or to have them coexist for awhile, we will be fixing this bug.
Thanks for the heads-up @leerob. Do you know what would be the behaviour after the fix?
Would i18n have no effect on the routes inside the app directory? or
Would locale be mapped to the app/[locale] folder? or
Something entirely different?
This may help us who are facing this issue proceed development with correct assumptions while we wait for the fix.
18n will exist, but only affect routes inside of the pages directory 👍
|
2025-04-01T06:40:33.481014
| 2017-01-19T23:03:50
|
202001608
|
{
"authors": [
"t9md"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11108",
"repo": "t9md/atom-narrow",
"url": "https://github.com/t9md/atom-narrow/issues/36"
}
|
gharchive/issue
|
Show rows at gutter instead of as-editor's normal-text.
Pros
Easier to correspond narrowEditor's text to underlying editor's text.
When introduce inline-edit feature(Update bounded-editor's text by editing narrowEditor.)
Cons
When copying, no longer copy rows(possibly, sometimes useful to share row info to other person).
Changed my mind.
|
2025-04-01T06:40:33.494471
| 2020-06-16T06:41:50
|
639412327
|
{
"authors": [
"ba32107"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11109",
"repo": "tabfloater/tabfloater",
"url": "https://github.com/tabfloater/tabfloater/issues/102"
}
|
gharchive/issue
|
Handle restoring last closed tab
When a floating tab is closed, and the user presses Ctrl+Shift+T (Ctrl+Shift+N on Firefox) to restore the last closed tab, the floating tab reappears (as a popup), but it's not always on top and TabFloater doesn't know it is restored.
Catch the tab restore event (if possible), and either:
Enable the floating state
or convert the floating tab straight into a normal tab
For this, we'd need to somehow identify if the tab that's being restored is the same floating tab that was closed last time. This is not straightforward, as the tab IDs are different. I suppose it would be possible to compare window types and URLs, but it quickly becomes too complex. It also needs a lot of extra events, adding performance overhead for every tab open operation.
Closing as won't fix.
|
2025-04-01T06:40:33.512995
| 2023-03-20T11:23:08
|
1631909806
|
{
"authors": [
"jacalata",
"peter-malcolm-bw"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11110",
"repo": "tableau/server-client-python",
"url": "https://github.com/tableau/server-client-python/issues/1210"
}
|
gharchive/issue
|
datasource item nullable project
Describe the bug
A clear and concise description of what the bug is.
I am unable to list all of the datasources on the server.
This looks like the same issue as https://github.com/tableau/server-client-python/pull/1028 except with DatasourceItem instead of WorkbookItem
Versions
Details of your environment, including:
Tableau Server version (or note if using Tableau Online) - Tableau Online
Python version: 3.9.12
TSC library version: 0.24
To Reproduce
Steps to reproduce the behaviour. Please include a code snippet where possible.
import tableauserverclient as TSC
# log in to tableau server
for datasource in TSC.Pager(server.datasources):
print(datasource)
The error can be fixed by removing one line from datasource_item.py:
class DatasourceItem(object):
...
@project_id.setter
#@property_not_nullable # TODO: remove this line
def project_id(self, value: str):
self._project_id = value
Results
What are the results or error messages received?
src/tableau_bot/refresh.py:142: in tableau_refresh
for datasource in Pager(server.datasources):
../../Library/Caches/pypoetry/virtualenvs/tableau-bot-sFRF_OxE-py3.9/lib/python3.9/site-packages/tableauserverclient/server/pager.py:40: in __iter__
current_item_list, last_pagination_item = self._endpoint(self._options)
../../Library/Caches/pypoetry/virtualenvs/tableau-bot-sFRF_OxE-py3.9/lib/python3.9/site-packages/tableauserverclient/server/endpoint/endpoint.py:205: in wrapper
return func(self, *args, **kwargs)
../../Library/Caches/pypoetry/virtualenvs/tableau-bot-sFRF_OxE-py3.9/lib/python3.9/site-packages/tableauserverclient/server/endpoint/datasources_endpoint.py:77: in get
all_datasource_items = DatasourceItem.from_response(server_response.content, self.parent_srv.namespace)
../../Library/Caches/pypoetry/virtualenvs/tableau-bot-sFRF_OxE-py3.9/lib/python3.9/site-packages/tableauserverclient/models/datasource_item.py:331: in from_response
datasource_item = cls(project_id)
../../Library/Caches/pypoetry/virtualenvs/tableau-bot-sFRF_OxE-py3.9/lib/python3.9/site-packages/tableauserverclient/models/datasource_item.py:58: in __init__
self.project_id = project_id
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <tableauserverclient.models.datasource_item.DatasourceItem object at 0x10995cdf0>, value = None
@wraps(func)
def wrapper(self, value):
if value is None:
error = "{0} must be defined.".format(func.__name__)
> raise ValueError(error)
E ValueError: project_id must be defined.
NOTE: Be careful not to post user names, passwords, auth tokens or any other private or sensitive information.
yep - this should be simple enough to get in a new release very soon.
|
2025-04-01T06:40:33.514379
| 2017-02-28T23:23:05
|
210933378
|
{
"authors": [
"RussTheAerialist",
"t8y8"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11111",
"repo": "tableau/server-client-python",
"url": "https://github.com/tableau/server-client-python/issues/152"
}
|
gharchive/issue
|
Update Contributing Guide
To include steps on how to branch off of development -- it'll avoid the merge conflicts from master we're seeing in PRs.
(Note: I'm not a git wizard, but my sequence of steps works every time :) )
@grbritz addressed this.
|
2025-04-01T06:40:33.518096
| 2024-07-09T14:24:22
|
2398370876
|
{
"authors": [
"JariKonstantin",
"ciminelli"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11112",
"repo": "tableflowhq/csv-import",
"url": "https://github.com/tableflowhq/csv-import/issues/233"
}
|
gharchive/issue
|
Localization
It would be helpful, if we could either manually localize the text or you offer localization.
This is now available with version 1.0.12!
The documentation can be found here: https://github.com/tableflowhq/csv-import?tab=readme-ov-file#internationalization
|
2025-04-01T06:40:33.583053
| 2016-12-18T01:57:04
|
196253365
|
{
"authors": [
"mrmrs",
"sebastienbarre"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11113",
"repo": "tachyons-css/tachyons-css.github.io",
"url": "https://github.com/tachyons-css/tachyons-css.github.io/pull/131"
}
|
gharchive/pull-request
|
Recent Components page, RSS feed
Hi folks.
This PR adds a new "Recent Components" page, and "Recent Components" RSS feed.
A couple more, small, build scripts were added -- check build.js
co(function* generator() {
yield componentsBuildList(options); // <- builds temporary components list (JSON)
yield componentsBuildIndex(options); // <- builds index pages (by category & most recent)
yield componentsBuildRSS(options); // <- builds RSS feed
yield componentsBuildPages(options); // <- comment to skip building pages
yield componentsBuildScreenshots(options); // <- comment to skip building screenshots
}).then(() => {
The temporary file that lists all components is now built by componentsBuildList,
The index pages are built componentsBuildIndex,
The RSS feed is built by componentsBuildRSS,
The rest is unchanged.
Note that componentsBuildList now uses the Git api (via nodegit) to find when a component was first committed to the repo. Took me some gitter/slack back and forth with the nodegit people, but I think this is robust to renaming and moving the component around. This introduces a new creationTime key in the component list (the temporary JSON file), which is leveraged later on. I also store a new author key, which is used in the RSS feed. Note that all this querying of the repo slows down this step, but it shouldn't be too bad (about 5 seconds on my laptop for all of the components).
The index pages still use the same template file, but new options allow for more flexibility. The list of index pages to build is in options.components.index:
components: {
index: { // list index pages to build
byCategory: { // components by category
title: 'Components',
path: 'components/index.html', // target location of index by category
sortAllBy: [['src'], ['asc']], // sort by file location will do
limitAll: false, // use all components
createSectionsBy: 'category', // create a section for each category
showSectionsTOC: true, // show Table of Contents (e.g. categories)
},
The byCategory key above describes the previously existing index page, the one listing all components by category. These options can be passed to a small function in components-build-sections that return a new list of components organized by sections. In the above example,
the list of components is first sorted by the src attribute (the path on disk),
all of the components are taken into account (limitAll is false),
sections are created by looking at the category key on each component,
there is a Table of Contents.
Now here is another index, and that's all it takes to describe the new "Recent Components" page:
mostRecent: { // most recent components
title: 'Recent Components',
path: 'components/recent.html', // target location of recent index
sortAllBy: [['creationTime'], ['desc']], // sort by most recent component first
limitAll: 50, // use the 50 most recent ones
createSectionsBy: creationTimeToYMD, // group by day
prettifySection: v => moment(v).format('LL'), // display as day
showSectionsTOC: false, // no need for Table of Contents
},
the list of components is first sorted by the creationTime attribute, most recent first,
only 50 components are on that page -- set to false if you prefer to list all of them,
sections could have been created using the creationTime key, but I thought it looked better when grouping all components created the same day together. The creationTimeToYMD callback takes a component and return whatever value you want to group components by (e.g. the section) -- in this case it converts the component creation time (milliseconds) to a YYYY-MM-DD format. Change it back to 'creationTime' if you'd prefer a section for each component.
prettifySection is a callback that will format the section for display on the page. Here, it converts the YYYY-MM-DD value to something more human-readable. Remove that property altogether if you end up switching createSectionsBy back to 'creationTime', since formatting the creationTime property is already handled by the options.components.prettify.creationTime callback.
I'm hiding the Table of Contents, which would only be a list of dates, didn't find it very useful.
The options.components.rss object should be self-explanatory:
rss: { // RSS feed
title: 'Tachyons Recent Components',
categories: ['CSS', 'Functional CSS'], // Categories this feed belongs to
ttl: 60, // Number of mins feed can be cached
path: 'components/rss.xml', // target location of feed (sync head.html)
count: 20, // how many in feed
},
I decided to only include 20 of the most recent components in that feed, that's a common value for feeds; feel free to customize.
Building a RSS feed that properly features the component screenshot was... tricky. I tried RSS enclosures, RSS custom elements -- no dice. I ended up looking at how http://unsplash.com does it. It seems to work fine. For RSS auto-discovery I added this to the templates/head.html:
<link rel="alternate" type="application/rss+xml" title="RSS Feed for Tachyons Recent Components" href="/components/rss.xml" />
This should hopefully let you put any pages from http://tachyons.io inside a RSS feed reader, and get the RSS feed directly.
What I leave up to you:
the new "Recent Components" page (aka components/recent.html) is not referenced from any other page right now. Feel free to add a link to it from the header? Making it look good is not my area of expertise :)
same for http://tachyons.io/components/rss.xml -- RSS auto-discovery should work, but feel free to explicitly mention the feed on the home page.
once this is up online, let's test the feed together -- I'll subscribe to it and hopefully you can add a component later on. If/when I see it in my reader I'll let you know that it all worked nicely.
Let me know if you have any questions, feedback, etc.
This is really amazing work.
Thanks guys. I'll keep an eye open, and once the RSS feed is up I'll let you know if it works as expected.
|
2025-04-01T06:40:33.586363
| 2022-06-23T13:18:29
|
1282395552
|
{
"authors": [
"mattwr18",
"tillprochaska"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11114",
"repo": "tactilenews/100eyes",
"url": "https://github.com/tactilenews/100eyes/issues/1388"
}
|
gharchive/issue
|
Set up Dependabot for signal-cli-rest-api
Dependabot currently doesn't support Docker Compose, a possible fork around is to create a Dockerfile that contains only the following line
FROM user/repo:version
… and reference that Dockerfile in the compose config.
@roschaefer do you think you might get to this this quarter or should we move it to a future one?
|
2025-04-01T06:40:33.594104
| 2023-02-22T20:05:16
|
1595760285
|
{
"authors": [
"HyperUpscale",
"tae-yeop"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11115",
"repo": "tae-yeop/insetgan",
"url": "https://github.com/tae-yeop/insetgan/issues/1"
}
|
gharchive/issue
|
extension for stable-diffusion-webui?
Hello,
Not sure how this github works... sorry about that in advance!
I am just a user ...
Very popular is using https://github.com/AUTOMATIC1111/stable-diffusion-webui
I was just wondering if you can make body generation from a single portrait as extension to be added to the stable-diffusion-webui?
Thank you in advance!
Hi, sorry for the late reply.
I'm not the original authors of this works. I was just trying to test out how this model works.
This model uses two StyleGAN to generate separate images of a face and a body. These generated images then go through an optimization process to make them look more natural, which means it refines the finer details. So this model don't generate the whole body in one-shot fashion.
the interface of stable-diffusion-webui is basically gradio library which is not that hard to learn. But i don't have to much time to make this model to webui.
Thank you so much for the information!
|
2025-04-01T06:40:33.640327
| 2024-05-20T16:02:34
|
2306301094
|
{
"authors": [
"davaymne",
"mratsim",
"smtmfft"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11116",
"repo": "taikoxyz/raiko",
"url": "https://github.com/taikoxyz/raiko/issues/227"
}
|
gharchive/issue
|
SGX prove fails with: Can not serialize input for SGX io error
Describe the bug
BACKGROUND:
Node successfully generated and submitted proof for block 179947
But when it tried to prove follow yo blocks, e.g. 180,466 it gets error:
ERROR[05-20|14:49:17.718] Failed to request proof height=180,466 error="Can not serialize input for SGX io error: Broken pipe (os error 32), output is Ok(Output { status: ExitStatus(unix_wait_status(256)), stdout: \"Starting one shot mode\\nGlobal options: GlobalOpts { secrets_dir: \\\"/root/.config/raiko/secrets\\\", config_dir: \\\"/root/.config/raiko/config\\\" }, OneShot options: OneShotArgs { sgx_instance_id: 5466 }\\nmemory allocation of 184 bytes failed\\n\", stderr: \"Gramine is starting. Parsing TOML manifest file, this may take some time...\\n-----------------------------------------------------------------------------------------------------------------------\\nGramine detected the following insecure configurations:\\n\\n - loader.insecure__use_cmdline_argv = true (forwarding command-line args from untrusted host to the app)\\n - sys.insecure__allow_eventfd = true (host-based eventfd is enabled)\\n - sgx.allowed_files = [ ... ] (some files are passed through from untrusted host without verification)\\n\\nGramine will continue application execution, but this configuration must not be used in production!\\n-----------------------------------------------------------------------------------------------------------------------\\n\\n[P1:T4:sgx-guest] error: Out-of-memory in library OS\\n\" })" endpoint=http://<IP_ADDRESS>:8080
INFO [05-20|14:49:55.410] Proof generated height=180,466 time=1m4.972853124s producer=SGXProofProducer
SGX server is configured with Processor memory reserved: 512 MB:
Steps to reproduce
Steps to reproduce here.
Spam policy
[X] I verify that this issue is NOT SPAM and understand SPAM issues will be closed and reported to GitHub, resulting in ACCOUNT TERMINATION.
UPDATE:
SGX prover keeps throwing error messages, but at the same time keep generating proof:
INFO [05-20|22:54:26.674] Request proof from raiko-host service blockID=182,180 coinbase=0x44c7dB3ac68d92398f88Cb2BD98C118925080e11 height=182,180 hash=4d1881..b6ea37
ERROR[05-20|22:54:28.793] Failed to request proof height=182,180 error="Can not serialize input for SGX io error: Broken pipe (os error 32), output is Ok(Output { status: ExitStatus(unix_wait_status(256)), stdout: \"Starting one shot mode\\nGlobal options: GlobalOpts { secrets_dir: \\\"/root/.config/raiko/secrets\\\", config_dir: \\\"/root/.config/raiko/config\\\" }, OneShot options: OneShotArgs { sgx_instance_id: 5466 }\\nmemory allocation of 184 bytes failed\\n\", stderr: \"Gramine is starting. Parsing TOML manifest file, this may take some time...\\n-----------------------------------------------------------------------------------------------------------------------\\nGramine detected the following insecure configurations:\\n\\n - loader.insecure__use_cmdline_argv = true (forwarding command-line args from untrusted host to the app)\\n - sys.insecure__allow_eventfd = true (host-based eventfd is enabled)\\n - sgx.allowed_files = [ ... ] (some files are passed through from untrusted host without verification)\\n\\nGramine will continue application execution, but this configuration must not be used in production!\\n-----------------------------------------------------------------------------------------------------------------------\\n\\n[P1:T15:sgx-guest] error: Out-of-memory in library OS\\n\" })" endpoint=http://<IP_ADDRESS>:8080
INFO [05-20|22:54:38.324] Check synced L1 snippet from anchor blockID=182,180 l1Height=1,581,891
ERROR[05-20|22:54:40.855] Failed to request proof height=182,180 error="Can not serialize input for SGX io error: Broken pipe (os error 32), output is Ok(Output { status: ExitStatus(unix_wait_status(256)), stdout: \"Starting one shot mode\\nGlobal options: GlobalOpts { secrets_dir: \\\"/root/.config/raiko/secrets\\\", config_dir: \\\"/root/.config/raiko/config\\\" }, OneShot options: OneShotArgs { sgx_instance_id: 5466 }\\nmemory allocation of 184 bytes failed\\n\", stderr: \"Gramine is starting. Parsing TOML manifest file, this may take some time...\\n-----------------------------------------------------------------------------------------------------------------------\\nGramine detected the following insecure configurations:\\n\\n - loader.insecure__use_cmdline_argv = true (forwarding command-line args from untrusted host to the app)\\n - sys.insecure__allow_eventfd = true (host-based eventfd is enabled)\\n - sgx.allowed_files = [ ... ] (some files are passed through from untrusted host without verification)\\n\\nGramine will continue application execution, but this configuration must not be used in production!\\n-----------------------------------------------------------------------------------------------------------------------\\n\\n[P1:T13:sgx-guest] error: Out-of-memory in library OS\\n\" })" endpoint=http://<IP_ADDRESS>:8080
INFO [05-20|22:54:41.456] Check synced L1 snippet from anchor blockID=182,180 l1Height=1,581,891
INFO [05-20|22:54:53.298] Proof generated height=182,180 time=26.623564692s producer=SGXProofProducer
INFO [05-20|22:54:53.298] NewProofSubmitter block proof blockID=182,180 coinbase=0x44c7dB3ac68d92398f88Cb2BD98C118925080e11 parentHash=3ce44b..de1151 hash=4d1881..b6ea37 stateRoot=77c37b..e9d758 proof=0000155a717639a029c7e5db6eddd6561b101269a4a17ecabcd937853b3a5be3e8c0e0357f2f450fdf26da8fbf2505131cefea562f6e72f868081312ece07165295e92d4194f76b353e02c50f54c17e49db04877963ab4df1c tier=200
INFO [05-20|22:54:53.308] Build proof submission transaction blockID=182,180 gasLimit=0 guardian=false
INFO [05-20|22:54:53.328] Publishing transaction service=prover tx=a7edc7..fde19a nonce=72 gasTipCap=1,000,000,000 gasFeeCap=12,598,439,738 gasLimit=298,584
INFO [05-20|22:54:53.331] Transaction successfully published service=prover tx=a7edc7..fde19a nonce=72 gasTipCap=1,000,000,000 gasFeeCap=12,598,439,738 gasLimit=298,584
INFO [05-20|22:54:56.456] Check synced L1 snippet from anchor blockID=182,180 l1Height=1,581,891
INFO [05-20|22:55:04.988] Proof assignment request body feeToken=0x0000000000000000000000000000000000000000 expiry=1,716,247,504 tierFees="[{Tier:100 Fee:+1000000000} {Tier:200 Fee:+1000000000} {Tier:1000 Fee:+0}]" blobHash=012680..ba5bfc currentUsedCapacity=0
INFO [05-20|22:55:04.988] Prover's ETH balance balance=36.56251266 address=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC
INFO [05-20|22:55:04.989] Prover's Taiko token balance balance=0 address=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC
WARN [05-20|22:55:04.989] Prover does not have required on-chain Taiko token balance providedProver=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC taikoTokenBalance=0 minTaikoTokenBalance=0
INFO [05-20|22:55:04.990] Proof assignment request body feeToken=0x0000000000000000000000000000000000000000 expiry=1,716,247,504 tierFees="[{Tier:100 Fee:+1100000000} {Tier:200 Fee:+1100000000} {Tier:1000 Fee:+0}]" blobHash=012680..ba5bfc currentUsedCapacity=0
INFO [05-20|22:55:04.991] Prover's ETH balance balance=36.56251266 address=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC
INFO [05-20|22:55:04.992] Prover's Taiko token balance balance=0 address=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC
WARN [05-20|22:55:04.992] Prover does not have required on-chain Taiko token balance providedProver=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC taikoTokenBalance=0 minTaikoTokenBalance=0
INFO [05-20|22:55:04.993] Proof assignment request body feeToken=0x0000000000000000000000000000000000000000 expiry=1,716,247,504 tierFees="[{Tier:100 Fee:+1320000000} {Tier:200 Fee:+1320000000} {Tier:1000 Fee:+0}]" blobHash=012680..ba5bfc currentUsedCapacity=0
INFO [05-20|22:55:04.993] Prover's ETH balance balance=36.56251266 address=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC
INFO [05-20|22:55:04.994] Prover's Taiko token balance balance=0 address=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC
WARN [05-20|22:55:04.994] Prover does not have required on-chain Taiko token balance providedProver=0xfbfd4F6993BC0D3481B9bf61AD0892f817a2e7aC taikoTokenBalance=0 minTaikoTokenBalance=0
INFO [05-20|22:55:05.340] Transaction confirmed service=prover tx=a7edc7..fde19a block=07b0e5..730636:1581901 effectiveGasPrice=7,418,930,979
INFO [05-20|22:55:05.341] "💰 Your block proof was accepted" blockID=182,180 parentHash=3ce44b..de1151 hash=4d1881..b6ea37 stateRoot=77c37b..e9d758 txHash=a7edc7..fde19a tier=200 isContest=false
It's OOM, try increasing gramine memory related setting maybe helpful.
Reference: https://gramine.readthedocs.io/en/latest/manifest-syntax.html#enclave-size
With mainnet and SGX running there for a whole month, I believe this issue is outdated, so closing it for now. Feel free to comment otherwise and we'll reopen for investigation.
|
2025-04-01T06:40:33.721769
| 2023-04-01T05:55:32
|
1650314259
|
{
"authors": [
"AjayDevInfy",
"Doesntmeananything",
"aaronmw",
"ddahan",
"gurgeous",
"jd1378",
"kmcgurty",
"kotaksempit",
"krisz094",
"mingtheanlay",
"or2e",
"thecrypticace",
"tjkohli",
"tkat",
"valgeirb"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11117",
"repo": "tailwindlabs/prettier-plugin-tailwindcss",
"url": "https://github.com/tailwindlabs/prettier-plugin-tailwindcss/issues/144"
}
|
gharchive/issue
|
vscode hanging with some vue files
What version of prettier-plugin-tailwindcss are you using?
0.2.5
What version of Tailwind CSS are you using?
3.2.7
What version of Node.js are you using?
18.12.1
What package manager are you using?
yarn
What operating system are you using?
macOS
Describe your issue
We added prettier-plugin-tailwindcss a couple of days ago. It worked initially, but I noticed that vscode+prettier was hanging trying to save certain .vue files. The hangs were obvious because files would fail to save and vscode would complain that prettier was still running. This occurred even after restarting vscode, etc. I removed the package and vscode was able to save those files again.
Sorry about the vague bug report. We are not using any other prettier plugins. We love the idea and really want it for our project! Thanks for all your hard work.
Does prettier hang when you run npx prettier -w . in your project?
I cannot get it to happen with prettier on the CLI. It's intermittent in vscode, but definitely present and unrecoverable. Here's the shortest file I could make that triggers it occasionally:
Something.vue:
<template>
<div v-if="error" class="bg-error text-xs mt-1 bg-error-content py-1 px-2 inline-block rounded">
{{ error }}
</div>
</template>
<script setup lang="ts">
import { ref } from "vue";
const error = ref<string>();
</script>
$ code --list-extensions | rg "vue|prett"
esbenp.prettier-vscode
Vue.volar
Vue.vscode-typescript-vue-plugin
settings.json:
"editor.formatOnSave": true,
"[vue]": {
"editor.codeActionsOnSave": { "source.organizeImports": true },
"editor.defaultFormatter": "esbenp.prettier-vscode"
},
.prettierrc:
{
"printWidth": 100
}
Are you by chance using the Tailwind CSS intellisense extension as well?
I am not using that. Anything else I should grep for? Thanks!!
Nope, I don't think so. I'll see what I can figure out.
@gurgeous So far I can't get the hang to happen. Any chance you could provide a project that you're able to reproduce it on? Also the full list of install extensions would be super helpful.
I can't share the project unfortunately. Maybe I can come up with a better repro. I'm also happy to pop open dev tools or mess around with vscode if you have any pointers. I've written some extensions too...
Here's the full list of installed extensions, apologies:
aki77.haml-lint alexcvzz.vscode-sqlite andersliu.insert-line-number astro-build.astro-vscode bierner.markdown-preview-github-styles BriteSnow.vscode-toggle-quotes bungcip.better-toml coolbear.systemd-unit-file cpylua.language-postcss csstools.postcss dbaeumer.vscode-eslint donjayamanne.githistory doublefint.pgsql earshinov.permute-lines EditorConfig.EditorConfig enkia.tokyo-night esbenp.prettier-vscode flesler.url-encode formulahendry.code-runner GitHub.copilot GitHub.vscode-pull-request-github golang.go Grumpydev.pico8vscodeeditor gurgeous.ruby-open-gem hogashi.crontab-syntax-highlight idleberg.applescript JuanBlanco.solidity kamikillerto.vscode-colorize karunamurti.haml Koihik.vscode-lua-format LoranKloeze.ruby-rubocop-revived mechatroner.rainbow-csv misogi.ruby-rubocop mrmlnc.vscode-duplicate ms-azuretools.vscode-docker ms-python.isort ms-python.python ms-python.vscode-pylance ms-vscode-remote.remote-ssh ms-vscode-remote.remote-ssh-edit ms-vscode.remote-explorer oderwat.indent-rainbow Orta.vscode-jest Prisma.prisma rebornix.ruby sgryjp.vscode-stable-sort Shopify.ruby-lsp skellock.just sldobri.gruvbox-5-stars sryze.uridecode stkb.rewrap stylelint.vscode-stylelint svelte.svelte-vscode syler.sass-indented tiehuis.zig Vue.volar Vue.vscode-typescript-vue-plugin william-voyek.vscode-nginx wingrunr21.vscode-ruby zkirkland.vscode-firstupper
Me too having the same issue on vscode. Had to removed this plugin to make saving vue file works again.
We've adopted a workaround for anyone interested:
Set pluginSearchDirs: false in .prettierrc. This fixes vscode.
Run prettier --plugin node_modules/prettier-plugin-tailwindcss from the CLI when you want to sort things.
We use just as our command runner and added a just sort to run the above. You could also use a pre-commit hook, etc.
Again, thanks for the great plugin! We love it (clearly) and I totally get that vscode/prettier are tough to debug. No worries and thanks for all your hard work.
For what it's worth, we're having the same issue with Next.js + TypeScript + TailwindCSS files. After spending over 4 hours trying to isolate the root cause, it was this plugin. Removing it solved the issue!
After days on a similar issue where one my CPU was stuck at 100% usage, I can confirm this comes from this plugin!
I'm using Nuxt 3.41 and I'm formatting Vue files.
On a 100-line Vue file:
Saving a file with Prettier only: 60ms and no CPU issue
Saving a file with Prettier + prettier-plugin-tailwind: 400ms, and minutes later, CPU usage raises, then my VS Code is completely stuck, and I have to reboot it.
Uninstalled :(
Too bad, it was useful to me.
In vue-cli based projects, downgrading Tailwind to 3.2.x resolves the issue.
Uninstall solve this :(
Any update on this? it seems pretty bad right now
You can run "Restart Extension Host" in VSCode to avoid rebooting the whole app and project — it seems to halt the runaway process and gets things working again. Sometimes I can just re-save the file afterwards without crash, but then a few more saves of the same file and it's crashing again.
Maybe that's a clue to what's happening with the plugin? I added pluginSearchDirs: false to prettier.config.js in my NextJS + Typescript project (as per @gurgeous' comment above) and I haven't had the issue so far. Though it's only been minutes, the one file that was crashing consistently seems not to be anymore ¯_(ツ)_/¯
👋 We're working towards a solution for this.
I've merged some (sometimes significant) performance and memory improvements in #153 that could potentially fix this problem. The fix is available to test via our insiders build: npm install prettier-plugin-tailwindcss@insiders
Could some of you give it a test and report back? It would be super helpful! You'll need to close/reopen VS Code after installing it so the prettier extension doesn't have the old version in memory.
@thecrypticace installing the plugin broke my Prettier, I have no formatting anymore when saving a file. Prettier logs when saving a file:
["INFO" - 3:35:55 PM] Formatting file:///workspaces/ppw/neuxt/pages/toolbox.vue
["INFO" - 3:35:55 PM] Using ignore file (if present) at /workspaces/ppw/.prettierignore
Notes:
I use Prettier for VS code extension
I have no .prettierignore file
I have no specific settings that could explain this.
I uninstalled the plugin and get a normal behaviour back, here are logs for a single save action:
["INFO" - 3:41:16 PM] Formatting file:///workspaces/ppw/neuxt/pages/toolbox.vue
["INFO" - 3:41:16 PM] Using ignore file (if present) at /workspaces/ppw/.prettierignore
["INFO" - 3:41:16 PM] File Info:
{
"ignored": false,
"inferredParser": "vue"
}
["INFO" - 3:41:16 PM] No local configuration (i.e. .prettierrc or .editorconfig) detected, falling back to VS Code configuration
["INFO" - 3:41:16 PM] Prettier Options:
{
"arrowParens": "always",
"bracketSpacing": true,
"endOfLine": "lf",
"htmlWhitespaceSensitivity": "css",
"insertPragma": false,
"singleAttributePerLine": true,
"bracketSameLine": false,
"jsxBracketSameLine": false,
"jsxSingleQuote": false,
"printWidth": 90,
"proseWrap": "preserve",
"quoteProps": "as-needed",
"requirePragma": false,
"semi": true,
"singleQuote": true,
"tabWidth": 2,
"trailingComma": "es5",
"useTabs": false,
"vueIndentScriptAndStyle": false,
"filepath": "/workspaces/ppw/neuxt/pages/toolbox.vue",
"parser": "vue"
}
["INFO" - 3:41:16 PM] Formatting completed in 69ms.
@ddahan There was a small hiccup that I just pushed a fix for a few minutes ago. Can you re-install the insiders build and give it one more test please?
The insiders version you'll need to test is 0.0.0-insiders.78bd35b (you can check your lock file to see if its the right one)
@thecrypticace it is definitely better than before!
Before, with the plugin activated: 400ms instead of 50-100ms to save a Vue file of 100 lines (and 100% CPU usage).
Now it seems to be almost the same.
However, it feels like if I wait for a few seconds to save the file, it takes longer (around 250ms). Then it's quick again. Is there anything that could explain this difference?
Yep so we removed object hashing to speed up config loading as for some config files it becomes very expensive but added an expiration timer so we can still detect changes to the config file (otherwise you'd have to reload VS Code every time the config changed). I'll see if there are other options for detecting config file changes to reduce the need for shorter expiration times.
Ok thanks! Not sure to understand deeply what's happening behind the scene, but the DX is a little weird because of that, because for the same file you have an almost instantaneous formatting, and then a laggy one. Anything that could improve this would be welcome imo. Thanks!
Still getting the same issue even with the insiders build 0.0.0-insiders.78bd35b. Disabling formatting on save with Prettier completely eliminates the issue, so it seems that the underlying issue is still there.
Not sure how to troubleshoot this further, but would be happy to provide any information that could be helpful here.
Still getting the same issue even with the insiders build 0.0.0-insiders.78bd35b. Disabling formatting on save with Prettier completely eliminates the problem, so it seems that the underlying cause is still there.
Not sure how to troubleshoot this further, but would be happy to provide any information that could be helpful here.
Same thing for me with 0.0.0-insiders.78bd35b.
@Doesntmeananything @valgeirb Could either one or both of you provide reproduction projects , details on VSCode plugins and VSCode version, and info on how you reproduce it? A video / screen recording might be helpful too so we can see what you're doing to repro it.
Additionally, specs on your computer could be useful (like CPU, memory, etc…)
I'm having the same issue and I'm rather curious what the common denominator is between people who have this.
I think this is related to a combination of installed NPM packages, because I've tried different VS Code versions, different Prettier extension versions, used different computers (one was a brand new Win11 installation), etc.
But there is just one project that is giving me issues while others work flawlessly.
Are any of you using DaisyUI for example?
Meanwhile I'm trying to create a minimal repro example.
@krisz094 Even if the repro isn't minimal it might still be useful as long as the issue is reproducible. Is that repo public by chance?
FWIW, I'm using Sveltekit 1.15.8 + Tailwindcss 3.2.4 + Typescript 5.0.4 + this plugin. The insiders version 0.0.0-insiders.d3f787d made a significant improvement. When VSCode upgraded to Typescript 5 last month, that's when we noticed the performance hit when autosaving. The workaround at the time was to downgrade VSCode to a patch before 1.77.
We are reproducing
https://github.com/verdie-g/crpg/tree/master/src/WebUI
can't we watch config file and update instead of expiring it after x seconds ?
@jd1378
can't we watch config file and update instead of expiring it after x seconds ?
We chose this router because, as far as I know, Prettier does not offer an API to do this. This means that it would need to be done via Node APIs — possibly using chokidar or some other file watcher. This also means that if we did that then we'd likely need to detect that we're running in VSCode because there's overhead to doing that and you wouldn't want to do it when running prettier from the command line.
that makes sense. but can't we detect and only do it when running in vscode ? so we don't get the overhead in cli, but get the benefits in vscode ? also it should be only one file I guess, so the overhead should be small. I mean if it means we don't reload the config for the rest of the life time of the extension, that would be a lot less work
I'm on a nuxt 3 project, it it not big yet, but still has ~ 50+ components already
here's without prettier-plugin-tailwindcss:
Rule | Time (ms) | Relative
:---------------------------------|----------:|--------:
prettier/prettier | 1735.189 | 50.0%
import/namespace | 1262.053 | 36.4%
@typescript-eslint/no-unused-vars | 70.073 | 2.0%
vue/attributes-order | 63.656 | 1.8%
n/no-deprecated-api | 25.548 | 0.7%
no-redeclare | 14.051 | 0.4%
import/order | 13.544 | 0.4%
unicorn/escape-case | 12.815 | 0.4%
vue/component-tags-order | 9.894 | 0.3%
spaced-comment | 9.673 | 0.3%
also as a thought: after squeezing as much as performance we can, can't we then create some web workers and offload the processing to web workers to process expressions in parallel ? assuming each expression must be done on it's own, wouldn't that help ?
@jd1378 While it would be amazing if that were possible — processing things in parallel in web workers would unfortunately require prettier integration. JS/TS expressions in non-JS/JSX/TS/TSX files are usually handled as embedded documents which means ultimately Prettier itself is responsible for calling parse in the appropriate plugins.
@or2e Since you're using WSL2 — are your projects files under /mnt/{drive-letter-here} (e.g. /mnt/c)? WSL2 has serious performance issues accessing files shared with Windows via /mnt/* and any access of the filesystem slows things down significantly. I took your project, updated it to the latest version of the prettier plugin (v0.3.0) and ran the CLI against one of the files. It takes roughly 5.4s to process ./src/pages/clans/\[id\]/index.vue under WebUI. About 4.7s of which is everything required to load your local copy of Tailwind, load and compile the config, and set up everything for processing. And the majority of this time is Node processing any require() calls.
If I move the repo so it's under /home/wsl the time drops significantly to 600ms total — with about 270ms of that being all the Tailwind, config loading, and setup. require() itself is slowed down a ton.
This might be something you can look at if you're still having performance problems.
@krisz094 Any chance you're using WSL as well?
@thecrypticace
1/ The repository has been cloned into the distro (not mounted), /root/WORK/crpg
2/
v0.2.8 - ~1500ms
v0.3.0 - 50-200ms
ty!
Fantastic!
Hey all — since we've released v0.3.0, with the discovery regarding WSL and windows file system sharing, and that there's been little movement on this issue we're going to close it as the issues should be mostly solved.
If you're still experiencing problems and can provide a reproduction please feel free to open a new issue that we can take a look at.
Thanks everyone for their input! ✨
Hi there. I'm curious to have extensive feedbacks from people who had the issue, an tested newer versions (0.3.0+) then. Thanks!
@gurgeous @kotaksempit @tjkohli @richardtallent-erm @Mattinton @Doesntmeananything @valgeirb @krisz094 @jd1378
|
2025-04-01T06:40:33.760991
| 2024-11-22T09:58:02
|
2682673704
|
{
"authors": [
"sukhbinder",
"taketwo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11118",
"repo": "taketwo/llm-ollama",
"url": "https://github.com/taketwo/llm-ollama/pull/20"
}
|
gharchive/pull-request
|
Fixes a bug when using llm chat is used with Ollama Model.
The error:
After the fix, this works.
Do you have an up-to-date version of ollama Python package installed?
Do you have an up-to-date version of ollama Python package installed?
Let me check.
No this is an issue with the llm-plugin.
with this commit. 678a50f11055e1ae8007134ef0c038df476f553e I get this error
Versions:
With the fix 0.8.0
Can you please check from your end.
This happens when we use
llm chat -m <some-ollama-model>
and do a muti turn.
Hm, so everything works fine with ollama==0.3.3. However, after I updated to the newest version (0.4.0)[https://github.com/ollama/ollama-python/releases/tag/v0.4.0] that was released today, I started to get errors:
no "name" attribute in model objects
no "model_info" attribute in in show response objects
I've made corresponding changes to use new names of these renamed attributes and then I was able to chat with Ollama models. This is strange:
Why do I not get the "attachment" error?
Why don't you get same attribute errors as I do?
Note that response objects should always have "attachments" attribute: https://github.com/simonw/llm/blob/335b3e635aa1439edafb13b0c2a225ce5840cc98/llm/models.py#L214
Hm, so everything works fine with ollama==0.3.3. However, after I updated to the newest version (0.4.0)[https://github.com/ollama/ollama-python/releases/tag/v0.4.0] that was released today, I started to get errors:
no "name" attribute in model objects
no "model_info" attribute in in show response objects
I've made corresponding changes to use new names of these renamed attributes and then I was able to chat with Ollama models. This is strange:
Why do I not get the "attachment" error?
Why don't you get same attribute errors as I do?
Strange.
I did get the attribute errors, in the previous update but then I started with a clean state on new venv with Python 3.9.12. What version of python are you testing this on?
And ollama version is 0.4.3 while Ollama python version is ollama 0.4.0
I'm testing on Python v3.12.3. I've pushed a bunch of updates that fix everything for me locally. Would you mind pulling latest master and creating a new venv from scratch?
Yes I will check in a bit. I think it will work. With an isolated env using uv I am getting the behaviour you are getting. This pr might not be needed. With ollama==0.3.3 and llm-ollama=0.7.0 it’s working like yours.
Only ollama==0.4.0 fixes are needed
I can confirm it's working.
versions
Thanks for confirming; I've tagged a new release: https://github.com/taketwo/llm-ollama/releases/tag/0.7.1.
|
2025-04-01T06:40:33.764544
| 2015-05-01T14:10:52
|
72439855
|
{
"authors": [
"McFoggy",
"takezoe"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11119",
"repo": "takezoe/gitbucket",
"url": "https://github.com/takezoe/gitbucket/issues/737"
}
|
gharchive/issue
|
Suggestion for good IDE setup
Hi,
can someone give advices on how to correctly setup an IDE to work on the project?
I tried Eclipse (& scala-ide), Netbeans & IDEA without lot of success.
In all 3 I can import the project, but in none of them I have a correct setup in which I have completion & navigation between Java, Scala & templates.
Of course using sbt on command line of course I build/run correctly the project
Thanks for any help.
PS: as you might have understand I am new to scala world
I recommend IntelliJ + Scala Plugin. You can simply open the /gitbucket directory as a SBT project.
Hi, thanks for the hint.
But as I told it is exactly what I did.
For example, opening the html templates does not lead in having Intellij interpreting them as "scala templates". Thus completion on contollers, routes, ... does not occur. Is there anything special to configure? I think IDEA see them as standard html files.
I come from pure java/maven world and was used to Netbeans & Eclipse. I perhaps missed something in IDEA setup but I doubt. I am using latest IDEA community edition (14.1.2) in which I activated scala support.
|
2025-04-01T06:40:33.817560
| 2022-07-20T11:39:42
|
1310911857
|
{
"authors": [
"meryacine",
"sr-gi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11120",
"repo": "talaia-labs/rust-teos",
"url": "https://github.com/talaia-labs/rust-teos/pull/79"
}
|
gharchive/pull-request
|
Pump rust-lightning to v0.0.108
This is needed because recent versions of rust-lightning expose the BigSize struct (used in TLV encoding) which will be used in some messages in the tower's Lightning interface (check bolt13).
Related to #31
I guess we can bump this as part of your SoB main PR, or do you need it early for any other purpose?
@sr-gi
Not for a specific thing, just separation of concerns.
These diffs look quite unrelated between the others.
Right. I'll merge it after the plugin release then.
|
2025-04-01T06:40:33.866812
| 2023-07-14T20:07:22
|
1805495509
|
{
"authors": [
"matthewhausman",
"natew"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11121",
"repo": "tamagui/tamagui",
"url": "https://github.com/tamagui/tamagui/issues/1423"
}
|
gharchive/issue
|
npm create tamagui failing
Current Behavior
run npm create tamagui
error:
@tamagui/toast@npm:1.39.9: The remote server failed to provide the requested resource
➤ YN0035: │ Response Code: 404 (Not Found)
➤ YN0035: │ Request Method: GET
➤ YN0035: │ Request URL<EMAIL_ADDRESS>➤ YN0035: │ @tamagui/web@npm:1.39.9: The remote server failed to provide the requested resource
➤ YN0035: │ Response Code: 404 (Not Found)
➤ YN0035: │ Request Method: GET
➤ YN0035: │ Request URL<EMAIL_ADDRESS>Expected Behavior
npm create tamagui is successful
Tamagui Version
not applicable
Reproduction
not applicable
System Info
System:
OS: macOS 13.0
CPU: (12) arm64 Apple M2 Pro
Memory: 76.44 MB / 16.00 GB
Shell: 5.8.1 - /bin/zsh
Binaries:
Node: 16.18.1 - ~/.nvm/versions/node/v16.18.1/bin/node
Yarn: 1.22.19 - ~/.nvm/versions/node/v16.18.1/bin/yarn
npm: 8.19.2 - ~/.nvm/versions/node/v16.18.1/bin/npm
Browsers:
Chrome: 114.0.5735.198
Safari: 16.1
npm was down
|
2025-04-01T06:40:33.872176
| 2015-09-21T14:51:33
|
107529810
|
{
"authors": [
"jochenvanwylick",
"tamasvajk"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11122",
"repo": "tamasvajk/SkyScanner",
"url": "https://github.com/tamasvajk/SkyScanner/issues/18"
}
|
gharchive/issue
|
Add search overloads that take in CancellationToken
Once a search is fired off - there is no way to cancel the search process. I would suggest we add an overload to the search methods that takes in a cancellation token - which will stop the polling process in the background.
In order to save resources - I would like to be able to cancel the process.
Something along the lines of
Task<List<Itinerary>> QueryFlight(FlightQuerySettings flightQuerySettings, CancellationToken token);
Good idea @jochenvanwylick.
Off:
Sorry I was not too responsive lately on the other open issues. I owe you at least one fix for two weeks now. I'm on my way of implementing these.
@tamasvajk No worries !
|
2025-04-01T06:40:33.904188
| 2015-12-25T17:05:15
|
123890723
|
{
"authors": [
"DenRozhko",
"IhorDavydenko",
"tananaev"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11123",
"repo": "tananaev/traccar",
"url": "https://github.com/tananaev/traccar/issues/1621"
}
|
gharchive/issue
|
table name "USER" can not use in oracle database
in the database Oracle can not be created with a table named "user".
http://docs.oracle.com/cd/B19306_01/server.102/b14200/ap_keywd.htm
and also
VARCHAR2 Maximum size: 4000 bytes
https://docs.oracle.com/cd/B28359_01/server.111/b28320/limits001.htm
The fact that "user" is a keyword doesn't mean that you can't use it. It's a keyword in MySQL as well and it works fine.
As for VARCHAR size:
http://docs.oracle.com/database/121/SQLRF/sql_elements001.htm#SQLRF55623
Beginning with Oracle Database 12c, you can specify a maximum size of 32767 bytes for the VARCHAR2, NVARCHAR2, and RAW data types.
12c is not in the free version and in Oracle12c 32767bytes mode for varchar2 must be enabled additionally
in Oracle 11xe (free oracle) varchar2 can only be 4000
into Oracle i can`t create table with name "user" or "USER" (these are different names in the oracle) and it is a fact
scripts to create the database traccar can not work with Oracle and I had to work hard to start the traccar server
It's really hard to cater for all available database engines. If you have any ideas how to fix the problem, please let me know.
You can do some scripts for various databases
I just don't have time to write scripts for all available databases. If you provide scripts, I can include them into the project.
Seems you can use the workaround. I was really looking very quickly. Need add into \database\DataManager.java\getObjectsTableName special case for Oracle DB - Wrap the name in double quotes. Also, do the same when DB is deployed. The main idea is [select * from user] cause error in Oracle, but [select * from "user"] will work fine if you create table in this manner: [create table "user" ( id number );]
(NOTE: in this case table name is case sensitive)
Sure you also need to adjust field types.
There is no "user" table in Traccar anymore, so it shouldn't be a problem.
|
2025-04-01T06:40:33.918000
| 2017-06-28T03:44:37
|
239044808
|
{
"authors": [
"Turbovix",
"neroshin",
"tananaev"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11124",
"repo": "tananaev/traccar",
"url": "https://github.com/tananaev/traccar/issues/3301"
}
|
gharchive/issue
|
Web not found
Hello folks I have an issues within web UI , the web view is alway blank from localhost http://prntscr.com/fp0fn8
Have you downloaded the web app?
yes folk, I download the traccar web UI and follow the instruction for traccar web in netbean but it said
"Attaching to localhost:8000
Connection refused.
"
I need to see your the source code folder structure that you have.
Also, if something doesn't work, please provide full information. What steps you followed, operating system, on what step it failed, ideally with screenshots.
I successfull to run it, but there something error popup after I add a new user device http://prntscr.com/fp2swu
That's not Traccar.
here screenshot from netbean http://prntscr.com/fp3c0f , really not sure if need to run this both traccar and traccar web.
I think you are confused. There is official traccar-web here:
https://github.com/tananaev/traccar-web
There is also unofficial traccar-web. That's the one you are using. If that's what you want, you asking in the wrong place.
where can I able to get the latest open source traccar web?
https://github.com/tananaev/traccar-web/releases
Em qui, 29 de jun de 2017 às 03:23, romano Mojica ednalan <
<EMAIL_ADDRESS>escreveu:
where can I able to get the latest open source traccar web?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/tananaev/traccar/issues/3301#issuecomment-311873731,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ADu5jlU6GEA5Vjn7SrjofsImMR10TiYgks5sI0LSgaJpZM4OHdM0
.
--
Att.
Marcio Torres®
Thank you for response @Turbovix and @tananaev
|
2025-04-01T06:40:33.941011
| 2015-09-23T22:58:35
|
108024193
|
{
"authors": [
"nvkelso",
"sensescape"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11125",
"repo": "tangrams/eraser-map",
"url": "https://github.com/tangrams/eraser-map/issues/33"
}
|
gharchive/issue
|
Mix of landuse kind wood and park in Prospect Park Brooklyn
Not sure what to do here:
Add wood to the park layer’s filter.
On Wed, Sep 23, 2015 at 3:58 PM, Geraldine Sarmiento <
<EMAIL_ADDRESS>wrote:
Not sure what to do here:
[image: screen shot 2015-09-23 at 3 56 36 pm]
https://cloud.githubusercontent.com/assets/466585/10061239/ef4f3faa-620b-11e5-85c5-e3a541a7491b.png
[image: screen shot 2015-09-23 at 3 58 15 pm]
https://cloud.githubusercontent.com/assets/466585/10061240/ef67bc38-620b-11e5-98e3-092bc6295331.png
—
Reply to this email directly or view it on GitHub
https://github.com/tangrams/eraser-map/issues/33.
ok, added
https://github.com/tangrams/eraser-map/blob/gh-pages/eraser-map.yaml#L1705
|
2025-04-01T06:40:33.941930
| 2016-01-26T21:41:52
|
128951932
|
{
"authors": [
"blair1618",
"tallytalwar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11126",
"repo": "tangrams/tangram-es",
"url": "https://github.com/tangrams/tangram-es/pull/504"
}
|
gharchive/pull-request
|
Use rounding when encoding position attributes as integers
Fixes some small (but visible) discrepancies in vertex positions at the edges of tiles.
LGTM.
|
2025-04-01T06:40:33.948230
| 2020-12-26T23:34:20
|
774948528
|
{
"authors": [
"bcamper",
"d3d9"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11127",
"repo": "tangrams/tangram",
"url": "https://github.com/tangrams/tangram/issues/772"
}
|
gharchive/issue
|
Firefox assumes that created data blobs (using data Object instead of an URL) are XML, outputs errors
TANGRAM VERSION: 0.21.1
ENVIRONMENT: win7 64bit Firefox 84.0.1, win10 64bit Firefox 85.0, (no errors:) win10 64bit Edge 87
TO REPRODUCE THE ISSUE, FOLLOW THESE STEPS:
use scene.setDataSource with type "GeoJSON" and data property instead of an URL
look at the console output (at layer rebuild)
RESULT:
There are errors at the blob locations: "XML Parsing Error: not well-formed".
Apparently the browser assumes blobs to be XML by default, if no type is given.
Outside of the console, I could not notice negative effects.
EXPECTED RESULT:
No such errors appearing in the console.
It seems like this is fixed, at least for (Geo)JSON data, by providing an object like { type: 'application/geo+json' } or at least { type: 'application/json' } to the Blob constructor here: https://github.com/tangrams/tangram/blob/990d2608c7dce2c3801c2cfd676e5c2e5b74c743/src/scene/scene.js#L1037
Thanks for the report and easy fix @d3d9! I wonder if this behavior has changed at some point in Firefox versions, but no matter now... this will be released in v0.21.2 and the issue will be closed then.
Thanks for the report and easy fix @d3d9! I wonder if this behavior has changed at some point in Firefox versions, but no matter now... this will be released in v0.21.2 and the issue will be closed then.
Fixed in v0.22.0
|
2025-04-01T06:40:33.969308
| 2021-03-08T08:09:35
|
824299227
|
{
"authors": [
"scala-steward"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11128",
"repo": "tanishiking/scalaunfmt",
"url": "https://github.com/tanishiking/scalaunfmt/pull/109"
}
|
gharchive/pull-request
|
Update scalatest to 3.2.6
Updates org.scalatest:scalatest from 3.1.4 to 3.2.6.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scalatest", artifactId = "scalatest" } ]
labels: test-library-update, semver-minor
Superseded by #113.
|
2025-04-01T06:40:33.986154
| 2019-08-14T14:05:37
|
480699170
|
{
"authors": [
"mbrandau",
"tannerlinsley"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11129",
"repo": "tannerlinsley/react-table",
"url": "https://github.com/tannerlinsley/react-table/issues/1446"
}
|
gharchive/issue
|
Header groups colSpan doesn't adjust with hidden columns
Using v7?
Thanks for using the alpha version of React Table v7! We're very excited about it.
Yes, and I'm excited too! 🔥
Describe the bug
When hiding columns, the colSpan of their respective headers doesn't change, resulting in columns being shifted to wrong headers.
To Reproduce
Steps to reproduce the behavior:
Go to https://codesandbox.io/s/tannerlinsleyreact-table-basic-b33h9
Play around with the checkboxes to hide the columns
Notice how the colSpan of the header groups doesn't change and columns get shifted to wrong headers
Expected behavior
The colSpan property of the header groups should be calculated with hidden columns in mind.
Screenshots
All columns show in this one.
Last name is hidden, age gets shifted to the Name header group.
This bug seems to be easy to fix. I'm happy to follow up with a PR!
See my commit for a possible fix.
I don't know why prettier changed that many lines though. 😕
I would love a PR!
On Aug 14, 2019, 8:56 AM -0600, Maximilian Brandau<EMAIL_ADDRESS>wrote:
See my commit for a possible fix.
I don't know why prettier changed that many lines though. 😕
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub, or mute the thread.
The issue is occurring again. https://codesandbox.io/s/tannerlinsleyreact-table-basic-b33h9
I‘m gonna have a look at this again. A test for this case might be a good idea.
Please reopen this issue @tannerlinsley
|
2025-04-01T06:40:33.989459
| 2019-09-08T10:23:43
|
490737084
|
{
"authors": [
"Codar97",
"elivoa",
"tannerlinsley"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11130",
"repo": "tannerlinsley/react-table",
"url": "https://github.com/tannerlinsley/react-table/issues/1505"
}
|
gharchive/issue
|
JSX cell is unexpected render to '[Object Object]' in v7.
JSX cell is render to '[Object Object]'
Affected version: 7.0.0-alpha.32 - 7.0.0-alpha.34
You can achieve the same with
columns: [ { Header: 'First Name', Cell: ({cell}) => <div>cell.value.firstName</div> accessor: 'firstName' }]
This may be its intentional functionality to combat cross site scripting.
This is fixed in the latest release. All renderers now support Function components, Class components, non-component functions, JSX elements, and primitives.
Also,@elivoa, Its important to note that accessor is meant to resolve a primitive data type (number, string, boolean, etc), so that it can be used to sort and filter the table. If you want to customize it's display, please do the following:
{
accessor: row => row.firstName,
Cell: ({ cell: { value }}) => <div>{value}<div>`
}
|
2025-04-01T06:40:33.994596
| 2018-02-23T12:29:11
|
299696114
|
{
"authors": [
"ishan123456789",
"remyblancke",
"tanoy009",
"wilgert"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11131",
"repo": "tanoy009/ng4-geoautocomplete",
"url": "https://github.com/tanoy009/ng4-geoautocomplete/issues/24"
}
|
gharchive/issue
|
How to remove the script from the index.html to use useGoogleGeoApi ?
I tried to removed the script of index.html
<script type="text/javascript" src="https://maps.google.com/maps/api/js?key=myKey&libraries=places"></script>
to
public userSettings5: any = {
geoCountryRestriction: ['fr'],
geoTypes: ['address'],
showSearchButton: false,
showRecentSearch: false,
useGoogleGeoApi: true
};
But not work
hi,
without the google map script mentioned in index.html, the directive won't work with Google API.
If u want to remove the script then u have to implement your own API which u can give the link to the directive to use. please refer demo and read the document.
Exactly, it's because I use agm-core at the same time, so with the src script I have a conflict with 2 google API called at the same time.
I had the same issue because two scripts were being loaded at the same time which was causing the issue.
I was able to make it work by not loading the Google Maps API in AgmCoreModule.
If you put this in your AppModule, AgmCoreModule does not load Google Maps API:
providers: [ ...BROWSER_GLOBALS_PROVIDERS, {provide: MapsAPILoader, useClass: NoOpMapsAPILoader} ],
Of course you have to leave the <script> in index.html.
This is a suboptimal but easy solution to make the two libraries cooperate.
Unfortunately I need the agm core to display a Google map (with markers, ...) on another page of my website, so ..
Same here, but for me that is still working with this setup.
AgmCoreModule.forRoot({
apiKey: 'AIzaSyDH1n-WWp1WgfRbK17-J0-BTlkF7i_czMg'
}),
//ng4-geoautocomplete
Ng4GeoautocompleteModule.forRoot(),
//infinite scroll
InfiniteScrollModule
],
This code got me api conflict when i call ng4-geoautcomplete and agm-core same time
Console screen error
|
2025-04-01T06:40:34.067528
| 2020-07-18T23:05:03
|
660436568
|
{
"authors": [
"colinrgodsey",
"jackjameshoward",
"jaredwolff"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11132",
"repo": "tarm/serial",
"url": "https://github.com/tarm/serial/pull/113"
}
|
gharchive/pull-request
|
custom baudrate for darwin and linux
Uses the IOCTL for darwin, and termios2 for linux custom baudrate. Retains non-darwin (termios 1) behavior for other posix targets. Change also modified the DTR "hangup" behavior for linux to more accurately mimic a terminal.
I was able to test this on OSX, Windows and Linux running 1M BAUD on an external device. Working a-ok here.
I've been using this for Raspbian and it works great.
|
2025-04-01T06:40:34.070266
| 2022-08-01T16:18:25
|
1324663373
|
{
"authors": [
"StanSvec"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11133",
"repo": "taro-suite/taro",
"url": "https://github.com/taro-suite/taro/issues/60"
}
|
gharchive/issue
|
Change configuration strategy
Auto-init strategy didn't work well during testing. We need a new strategy for handling of the configuration file:
The configuration file must be loaded when taro cli is executed. (Config dir resolution has been fixed to behave exactly according to the XDG spec)
~When the config file is not found then the executed command will print a message, suspend the execution, run config create and continues in case the config was created.~
~An exception to this is the taro exec command which will print a warning and load the default config file to be able to continue the execution. This is important because this command is expected to be executed in a non-interactive shell.~
We need also smarter taro config create command. When this command is executed we need a prompt similar to this one:
Select where to create the config file:
User specific config directory:
[1] $XDG_CONFIG_HOME/taro or ~/.config/taro
System directories (You must have the write permissions!)
[2] {The first directory from $XDG_CONFIG_DIRS}/taro or /etc/xdg/taro
[3] /etc/taro
Choose any of [1,2,3]: _
Bonus: taro config create --jobs for creating the jobs.yaml file template
The work will be done in #65
|
2025-04-01T06:40:34.090598
| 2021-06-15T02:15:22
|
920921487
|
{
"authors": [
"dbr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11134",
"repo": "taskchampion/taskchampion",
"url": "https://github.com/taskchampion/taskchampion/pull/276"
}
|
gharchive/pull-request
|
CI tests on Windows+macOS
..hopefully. As per https://github.com/taskchampion/taskchampion/issues/274
These failures now seem like legitimate failures which should be fixed by #275
🚀
|
2025-04-01T06:40:34.106434
| 2016-02-23T20:49:00
|
135859874
|
{
"authors": [
"bretd25",
"evantahler"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11135",
"repo": "taskrabbit/elasticsearch-dump",
"url": "https://github.com/taskrabbit/elasticsearch-dump/issues/170"
}
|
gharchive/issue
|
Will not overwrite file if it exists
My understanding is if a file already exists it should just overwrite. I am receiving this error message when trying to overwrite a file.
Tue, 23 Feb 2016 20:46:57 GMT | starting dump
Tue, 23 Feb 2016 20:46:57 GMT | got 1 objects from source elasticsearch (offset: 0)
Tue, 23 Feb 2016 20:46:57 GMT | Error Emitted => File /usr/data/car.json already exists, quitting
Tue, 23 Feb 2016 20:46:57 GMT | Total Writes: 0
Tue, 23 Feb 2016 20:46:57 GMT | dump ended with error (set phase) => Error: File /usr/data/car.json already exists, quitting
Thanks,
Bret
We changed this behavior (safety++) in v1 https://github.com/taskrabbit/elasticsearch-dump/releases/tag/v1.0.0
It looks like we forgot to update the README, which I've done here https://github.com/taskrabbit/elasticsearch-dump/commit/8aa9b7c37f579fc31c5997ff676873e17587f960
Thank you
|
2025-04-01T06:40:34.117049
| 2020-05-26T22:38:40
|
625244443
|
{
"authors": [
"monicakarnes",
"sevincent"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11136",
"repo": "tastefully-simple/cornerstone",
"url": "https://github.com/tastefully-simple/cornerstone/issues/123"
}
|
gharchive/issue
|
Credit- Multiple Line Credit- Created multiple Credit orders one for each line that was credited back
BC order 724
Myhq Order # 26929028
Issued Credit back... seen via myhq I see 2 credit orders created
26935340
26935341
It should be like this to match BC
Looks like it only gave me one credit order number in the table.
and note- I messed up and forgot to delete the tax part out... oops
TaxTotal and ShippingShippingCreditAmount are "order level" values, so they should be the same on all the items credited for the order. Being different caused the uspI_CreateCreditOrderForExternalProcess to process the items as separate orders. I did add code so that when this happens it will take them max of each of the values just in case this happens again.
Looks like this is fixed
Closing bug
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.