id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
58793007
Use prober results to return working WS server. This implements a simple strategy of returning whichever collider instance has been up the longest. For example, assume we have servers A and B which are both up, and GAE is returning A If A is reported as down by prober, GAE will start returning B. GAE will continue serving B even if A comes back up. B will returned until B if reported as down by problem. This will help avoid the situation where clients within a room are in different collider instances. It does not get us any load balancing benefits. If we want to do something more advanced here to ensure clients get into the right collider instance together, the next steps would be: When the first client joins a room, store the collider instance assigned to the client along with the room data. When the second client joins a room, ensure we return that same collider instance. Client would need to be updated to reconnect to the socket if the wss address returned by /join is different than the one it started with. The situation this would prevent would be: GAE is currently returning A, Client 2 loads /, gets A in the initial params A goes down, B is now returned Client 1 loads 1/, gets B, joins room, gets assigned B, connects to B Client 2 now joins a room, but connects to A because we connect before /join. When join returns B, client 2 needs to disconnect from A (possibly failed anyway) and connect to B. I suspect the effort for the more advanced tracking/reconnecting is not worth it at this time. @jiayliu @tkchin @milkypostman Switched to what we discussed, also added a couple test cases to cover the invalid data scenario. @milkypostman @tkchin I simplified what was being stored. Now memcache only stores the active host, since that is all we were using anyway. The update step now can just check if the active host is still active, or if not, pick one that is up and store that. Thanks for the simplification :) lgtm Could you also update cron.yaml to run the probers every 5min? lgtm %nits
gharchive/pull-request
2015-02-24T19:42:37
2025-04-01T06:46:13.845246
{ "authors": [ "chuckhays", "jiayliu", "tkchin" ], "repo": "webrtc/apprtc", "url": "https://github.com/webrtc/apprtc/pull/92", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1692275109
Revisit mandatory fields Description Remove required criteria. Justification: https://teamds.atlassian.net/browse/WOS-175 Motivation and Context Screenshots (if appropriate) Upgrade notes (if appropriate) Types of changes [ ] Bug fix (non-breaking change which fixes an issue) labeled with bug [x] New feature (non-breaking change which adds functionality) labeled with enhancement [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) Checklist: [x] My code follows the code standards of this project. [ ] My change requires updating the documentation. I have updated the documentation accordingly. @paweljozwicki Closed without merge as things got messed up because of major refactoring in the main branch. I opened https://github.com/websight-io/bulma/pull/192 instead.
gharchive/pull-request
2023-05-02T11:47:26
2025-04-01T06:46:13.899685
{ "authors": [ "gfazek" ], "repo": "websight-io/bulma", "url": "https://github.com/websight-io/bulma/pull/179", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1851882217
🛑 The Good Cyclist is down In 3ff9353, The Good Cyclist (https://thegoodcyclist.iwview.com) was down: HTTP code: 521 Response time: 107 ms Resolved: The Good Cyclist is back up in b67838c.
gharchive/issue
2023-08-15T17:58:26
2025-04-01T06:46:13.912994
{ "authors": [ "webworldview" ], "repo": "webworldview/uptime", "url": "https://github.com/webworldview/uptime/issues/1219", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1858377392
🛑 Jeff with Susan is down In 2b22f1b, Jeff with Susan (https://www.jeffwithsusan.com) was down: HTTP code: 521 Response time: 146 ms Resolved: Jeff with Susan is back up in 071a1bd after 417 days, 1 hours, 34 minutes.
gharchive/issue
2023-08-21T01:08:42
2025-04-01T06:46:13.915421
{ "authors": [ "webworldview" ], "repo": "webworldview/uptime", "url": "https://github.com/webworldview/uptime/issues/1736", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1598787708
🛑 TestPlayDump is down In 7897d36, TestPlayDump (https://www.testplaydump.com) was down: HTTP code: 400 Response time: 236 ms Resolved: TestPlayDump is back up in 8e0933d.
gharchive/issue
2023-02-24T14:57:32
2025-04-01T06:46:13.917813
{ "authors": [ "webworldview" ], "repo": "webworldview/uptime", "url": "https://github.com/webworldview/uptime/issues/703", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2041108628
created by d2-ai This PR implements #64, created by d2-AI. do something else [d2-AI] filter me image from rancher 在详情区域,增加对 ingress 和 egress 的展示。 在 ingress 详情中,每个 ingress 用一个 tab 展示。渲染每个 ingress 时,根据 ingress.from 中各条规则的类型分别渲染: IP block 类型,展示 CIDR 和 exceptions 字段; Namespace Selector 类型,展示 key、operator、value,并统计选中了多少个 namespace; Pod selector 类型,展示 key、operator、value,并统计选中了多少个 pod。 最后再将 ingress.ports 展示出来,包含 port 和 protocol 信息。 详情中 ingress.from 的数据示例如下: from: - ipBlock: cidr: 172.17.0.0/16 except: - 172.17.1.0/24 - namespaceSelector: matchLabels: project: myproject - podSelector: matchLabels: role: frontend 在 egress 详情中,每个 egress 用一个 tab 展示。渲染每个 egress 时,根据 egress.to 中各条规则的类型分别渲染: IP block 类型,展示 CIDR 和 exceptions 字段; 将 egress.ports 展示出来,包含 port 和 protocol 信息。 详情中 egress.to 的数据示例如下: to: - ipBlock: cidr: 10.0.0.0/24 这个是给 AI 的指令,还是关于PR的描述?
gharchive/pull-request
2023-12-14T07:31:03
2025-04-01T06:46:13.925309
{ "authors": [ "Yuyz0112", "tanbowensg" ], "repo": "webzard-io/dovetail-v2", "url": "https://github.com/webzard-io/dovetail-v2/pull/65", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
181920569
Remove redondant python file Salut, Un fichier redondant que tu as laissé passé, à supprimer ;) Hi, I understand that you thought 04_birthday_manager/birthday_countdown.py (1) was redundant with 03_birthday_countdown/program.py (2). But look, I don't think so. In the current state, (1) is indeed a copy-paste of (2), but you have to remark light differences: (1) is documented (because docstrings are only introduced in lesson 04, they are for example heavily used in 04_birthday_manager/container.py) (1) has two commented lines (30, 31), that are used during lesson 04 to show the interest of __name__ and __main__; besides in (1) main() is called within a if __name__ == '__main__': (otherwise, when you run 04_birthday_manager/program.py, the banner BIRTHDAY COUNTDOWN is displayed) More importantly, (1) is used to be imported in 04_birthday_manager/program.py via import birthday_countdown in line 2. Do you think of a better way to do things there? Yes, forget what I said yesterday ! I ran the wrong command, and I thought it was the same file. (git diff program.py birthday_countdown.py instead of diff program.py birthday_countdown.py). However, I should send you another PR soon. Happy to hear that :) can't wait!
gharchive/pull-request
2016-10-09T23:54:34
2025-04-01T06:46:13.966171
{ "authors": [ "CharbelGui", "Nairwolf" ], "repo": "wecodetogethere/appetizers_python", "url": "https://github.com/wecodetogethere/appetizers_python/pull/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
442941436
result of paper I am a beginner. I want to ask you how you got the result of your paper.The results of each validation are the max, and then the max in 10? I'm also very curious. I try to run the code and to adjust the hyper-parameters, but I cannot got expected results.
gharchive/issue
2019-05-11T01:42:38
2025-04-01T06:46:14.011715
{ "authors": [ "LYF14020510036", "veophi" ], "repo": "weihua916/powerful-gnns", "url": "https://github.com/weihua916/powerful-gnns/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
230661326
How to bind plotOptions event with Vue component Function I want to bind a vue component function with highchart plot, but It can't bind with function with option attribute. Can refs work? function createOptions(vm) { return { plotOptions: { series: { events: { click() { // ... vm.yourComponentFunction(); }, }, }, }, }, } { data() { return { options: createOptions(this), }; }, } Creating options when you can access vm Thank you. It works. I'm stupid.
gharchive/issue
2017-05-23T10:31:52
2025-04-01T06:46:14.021683
{ "authors": [ "moJiXiang", "weizhenye" ], "repo": "weizhenye/vue-highcharts", "url": "https://github.com/weizhenye/vue-highcharts/issues/19", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
376100240
Feature Request: Use Snap channels to separate stable/beta builds Would it be possible to begin using channels for snap releases as documented here: https://forum.snapcraft.io/t/channels/551 Currently Wekan channels are all the same release, and often there are multiple simultaneous releases to the point where I don't have the "latest" version despite using snap: What is actually more concerning to me is introducing new features to my users that might require a learning curve. It would be great to see most of the daily updates occurring on a candidate or beta channel and then added to stable releases on a schedule that I can anticipate (e.g. ~1/month). I love all of the work that you're doing, I'm just worried that introduction/removal of new features so rapidly is going to confuse my users and potentially break things (e.g. I've instructed them not to use custom fields because I don't want to repair JSON files every time they decide to delete the custom fields; I've also told them to avoid using subtasks due to the way the feature behaves). The latest version reintroduced rules, which canbe a great feature, but might be one that requires a learning curve. Maybe an alternative to this is admin ability to disable certain features. I'm using beta/edge etc channels when there is something that could be broken. @xet7 are the stable/edge channels supposed to be separate now? Per https://wekan.github.io/ the stable release is 2.35, but I'm seeing 2.42 across all channels when I run snap info wekan on my installation: refresh-date: today at 03:35 UTC channels: stable: 2.42 2019-03-07 (467) 133MB - candidate: 2.42 2019-03-07 (467) 133MB - beta: 2.42 2019-03-07 (467) 133MB - edge: 2.42 2019-03-07 (468) 133MB - installed: 2.42 (467) 133MB - @nmd3 I updated stable version to be 2.42 at website. The reason why 2.35 was marked stable was that after it there was major changes that did bring some new bugs, but many of those are already fixed, and there is no plans to remove any features and fixes that are at 2.42. Some months ago there really was separate stable and edge, because new features at edge were broken. When they did get more usable, they did get merged to stable. But now I'm thinking, would there still be a need to be separate stable and edge. I have used separate channels when needed, as I did for some days recently.
gharchive/issue
2018-10-24T13:31:33
2025-04-01T06:46:14.029101
{ "authors": [ "nmd3", "xet7" ], "repo": "wekan/wekan-snap", "url": "https://github.com/wekan/wekan-snap/issues/67", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2335932977
Distinguishing born-digital item from CALM data in pipeline In order to support the experience team to better display born digital items, we should investigate identifying these as a material type from CALM. See this description from @gestchild: We currently display the structure of digitised items in the viewer, e.g. the table of contents for books. It will do the same for born digital things, e.g. these images (not a great example as it only has one level, but the only example I have to hand). Ashley feels it doesn't work in this latter scenario and that we need to show all the items under each folder, which I think makes sense. However, if I were to do that it would also show all the pages in the table of contents for books. So we need someway to distinguish between born digital things, which should show the item level in the structure and non born digital things that shouldn't. We can't use the IIIF Manifest for this, so would need to implement your suggestion of surfacing that information in the catalogue API More context: https://wellcome.slack.com/archives/CGXDT2GSH/p1716394693535169 On investigation it appears that we can identify these works by the Material type specified in CALM. Adding a new Format type specific to material labelled "Archives - Digital" should suffice. Some more context in Slack: https://wellcome.slack.com/archives/CGXDT2GSH/p1718028991823389 Implemented a temporary fix on the FE to identify BD works, but that method won't be valid forever as more works come in. When this ticket here is done, this TODO should then be addressed ASAP, so ticket creation for it should be part of the DoD of this piece. @kenoir can this be closed now? Yup 👍
gharchive/issue
2024-06-05T13:31:30
2025-04-01T06:46:14.084279
{ "authors": [ "kenoir", "pollecuttn", "rcantin-w" ], "repo": "wellcomecollection/catalogue-pipeline", "url": "https://github.com/wellcomecollection/catalogue-pipeline/issues/2659", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1108965061
Bagit: stream zip directly to S3 Right now, we are creating a bagit zip file on EFS. The zip file is then uploaded to S3 with a TransferManager, as it might exceed the size limit of 5GB for standard object uploads. We now want to stream the zip directly to S3, utilizing S3 multipart uploads. For this, we will gather 50 MB chunks of data in a buffer on the heap and compute the md5 of that buffer. Then we will upload that chunk as a part of the multipart upload, setting the md5 sum as the ETag. This will continue until we reach the end of our zip stream. The last chunk may be smaller. As the multipart upload allows 10 000 chunks, we can upload bags of up to 500 000 MB (~500GB) with this method. We are good to go ahead with this 👍
gharchive/issue
2022-01-20T07:43:28
2025-04-01T06:46:14.086059
{ "authors": [ "aray-wellcome", "opaetzel" ], "repo": "wellcomecollection/goobi-infrastructure", "url": "https://github.com/wellcomecollection/goobi-infrastructure/issues/422", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
595276120
Storage tiering for video master files Objective Sustainable digitised and born-digital preservation Measures Are all existing MXF files only stored in cold storage tiers? Are new MXF files only stored in cold storage tiers during ingest? Initiatives [ ] Support new AV workflow in Goobi [ ] Support new AV workflow in DDS/DLCS [ ] Store new MXF files in cold storage tiers [ ] Migrate existing MXF files to cold storage tiers Related: https://github.com/wellcomecollection/platform/issues/4403
gharchive/issue
2020-04-06T17:02:22
2025-04-01T06:46:14.088779
{ "authors": [ "alexwlchan", "jtweed" ], "repo": "wellcomecollection/platform", "url": "https://github.com/wellcomecollection/platform/issues/4413", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
662800287
Spinup/down Staging environments as required Investigate any potential methods for spinning down/up staging and staging-prod environments when not in use. Initial look at using event-based auto-scaling to scale up on first request (and out after a period of time) doesn't look feasible due to the available metrics. Could use scheduled auto-scaling to turn off services on a schedule. Staging and Staging-Prod have different ECS services but share all other resources. Updated staging env to stop tasks at 1900 and start at 0700 Applied to staging env only for now, will test it tonight and apply to all envs if it is successful. PR: https://github.com/wellcomecollection/iiif-builder-infrastructure/pull/17 Verified overnight that this is working as expected, closing.
gharchive/issue
2020-07-21T09:30:35
2025-04-01T06:46:14.091079
{ "authors": [ "donaldgray" ], "repo": "wellcomecollection/platform", "url": "https://github.com/wellcomecollection/platform/issues/4679", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
617375773
Add load function to semantic similarity and kwargs to fit Description as above ^ If you fancy you can implement this #49 and release 2020.05.1
gharchive/pull-request
2020-05-13T11:38:01
2025-04-01T06:46:14.092019
{ "authors": [ "aCampello", "nsorros" ], "repo": "wellcometrust/WellcomeML", "url": "https://github.com/wellcometrust/WellcomeML/pull/47", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
237439464
add nocache to preview and fix promise function names Type 🔧 Fix Value Makes the redirect service work. Fixed mixed promise function names (resolve and reject). Also does't use the cache for /preview and /preview/*. All [x] Demoed to the relevant people [x] PR labelled and assigned [x] Any introduced code complexity has been flagged Merging as it's broken and the cache behaviour has already been deployed.
gharchive/pull-request
2017-06-21T07:29:01
2025-04-01T06:46:14.094963
{ "authors": [ "jamesgorrie" ], "repo": "wellcometrust/wellcomecollection.org", "url": "https://github.com/wellcometrust/wellcomecollection.org/pull/1031", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
356016651
add nicer logging to remapper Just to give you info and to stop me counting documents in folders. Doesn't affect prod systems, merging.
gharchive/pull-request
2018-08-31T15:02:51
2025-04-01T06:46:14.095812
{ "authors": [ "jamesgorrie" ], "repo": "wellcometrust/wellcomecollection.org", "url": "https://github.com/wellcometrust/wellcomecollection.org/pull/3262", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
121919884
WebView in content leaves patch behind Very simple to replicate. When there is a webview in content, it leaves a trace behind. Specifically, webview won't rotate Did anyone check this? I can not reproduce it, seems to be ok @wenchaojiang Can you put the code to load a webView? https://github.com/wenchaojiang/AndroidSwipeableCardStack/issues/96
gharchive/issue
2015-12-13T13:38:03
2025-04-01T06:46:14.114516
{ "authors": [ "SUBINPTPM", "christian160984", "neeradkumar", "wenchaojiang" ], "repo": "wenchaojiang/AndroidSwipeableCardStack", "url": "https://github.com/wenchaojiang/AndroidSwipeableCardStack/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
139689261
Get current top view I have an "like" textview to show when card is swipped to right, but I can´t get the view to apply alpha when needed. How can I do it? Not supported, you have to track the view collection status by your self Ok, thanks.
gharchive/issue
2016-03-09T20:16:09
2025-04-01T06:46:14.115901
{ "authors": [ "ihariza", "wenchaojiang" ], "repo": "wenchaojiang/AndroidSwipeableCardStack", "url": "https://github.com/wenchaojiang/AndroidSwipeableCardStack/issues/34", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1402320086
[Suggestion]: Use ForgeGui#rightHeight for overlays to allow for mod compat Hey there, nice mod. It would be really cool if you could use and update the ForgeGui#rightHeight field when drawing your overlays (specifically the hydration and water overlay) instead of only taking vanilla icon offsets into account. This will allow your overlays to be compatible with other mods that add new/change icon overlays above the hotbar. There's currently a visual conflict due to this with my Leave My Bars Alone mod: https://github.com/Fuzss/leavemybarsalone/issues/1 Using the Forge field would easily fix that. Ok, been messing around with this. While, yes, this should return a correct value, it doesn't update until after the custom overlay is rendered :( It looks like I need to do separate event handler outside of the Forge built-in to do this ... maybe in the POST render ... I have no idea how Forge does this to be honest. I see that theoretically when renderFood or renderAir should update this value, but it doesn't. I really don't know about rewriting a significant portion of the mod to support this. I know the Forge stuff is supposed to be there to help, but it looks like it actually causes more issues than it solves. I'll keep playing with this, let me know if you have any suggestions. Ok, I need to be using registerAbove to get this to work. Looks good now! I'll get this pushed out, thanks for the suggestion. Also, great mod, btw. I'm adding it to all my modpacks from here on out. Awesome, thank you! @Fuzss FYI, registerAbove and registerBelow happen before rightHeight is rendered. I think these are really reserved for manipulating Vanilla GUI elements. registerAboveAll and registerBelowAll are what I wanted to target for Mod elements, as they can take advantage of rightHeight. For example, registerAbove (or after) VanillaGuiOverlay.FOOD_LEVEL doesn't provide the updated rightHeight to my overlay element. I didn't bother digging in further to see how the ordering code actually works. Also, side note: This is so confusing, given that the methods and variables are so poorly named. Above == After and Below == Before in ordering and right/left height deal with the offset for the gui elements. I literally had to go to the Forge source to understand that it is: order and offset, not layer and (whatever rightHeight is supposed to tell me)???!!! Yeah well documentation (or more the existence thereof) really is not a strength of Forge haha This is where the Fabric ecosystem really shines, the documentation over there is amazing rightHeight and leftHeight are the names of those fields though as they control the height of icon bars on the right and left side of above the hotbar respectively. So leftHeight is for hearts/armor, rightHeight is for food/mount health/air meter And yeah, the above and below naming confused me too, but I think what they are referring to is where you want your icon bar to display relative to the one you're referencing when registering. So if I want my thirst bar to appear above the hunger bar, I need to use registerAbove. This is only makes sense for icon rows above the hotbar though, so for any other kind of overlay somewhere on the screen the aboveAll/belowAll methods are probably what I would go for.
gharchive/issue
2022-10-09T16:08:13
2025-04-01T06:46:14.122461
{ "authors": [ "Fuzss", "wendall911" ], "repo": "wendall911/Homeostatic", "url": "https://github.com/wendall911/Homeostatic/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
488577897
View [grid.fixed-table] not found. ` /** * Find the given view in the list of paths. * * @param string $name * @param array $paths * @return string * * @throws \InvalidArgumentException */ protected function findInPaths($name, $paths) { foreach ((array) $paths as $path) { foreach ($this->getPossibleViewFiles($name) as $file) { if ($this->files->exists($viewPath = $path.'/'.$file)) { return $viewPath; } } } throw new InvalidArgumentException("View [$name] not found."); } /** * Get an array of possible view files. * * @param string $name * @return array */ protected function getPossibleViewFiles($name) { return array_map(function ($extension) use ($name) { return str_replace('.', '/', $name).'.'.$extension; }, $this->extensions); } /** * Add a location to the finder. * * @param string $location * @return void Arguments "View [grid.fixed-table] not found." ` blade 视图文件覆盖不全 垃圾扩展没法用
gharchive/issue
2019-09-03T12:33:26
2025-04-01T06:46:14.129469
{ "authors": [ "sdjtu502", "wonbin2011" ], "repo": "wenxuanjun/laravel-adminlte3", "url": "https://github.com/wenxuanjun/laravel-adminlte3/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
269224606
Fix bug where Debezium Logical Converters were not used correctly. Also fix some found issues with the debezium date type converter. Related tests. LGTM!
gharchive/pull-request
2017-10-27T20:22:09
2025-04-01T06:46:14.149737
{ "authors": [ "criccomini", "mtagle" ], "repo": "wepay/kafka-connect-bigquery", "url": "https://github.com/wepay/kafka-connect-bigquery/pull/99", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
196195646
regex exclusion on --containers not working we have 2 containers in every pod, one always named consul which i want to filter. Would assume the following to work but no dice. Am I missing something? stern --container "[^cC].*" "*" Hey @dwelch2344 Sorry for the late response here. I think your problem is that the .* after the negated c/C still matches so basically it'll match any string. You could limit it so the text can't start with a c and things should work fine. Also the pod query needs .* (not just *), but that could be a copy-paste error. Can you see if this works as expected: stern --container "^[^cC].*" ".*" Assuming you figured this out. Feel free to open a new issue if not
gharchive/issue
2016-12-17T04:26:31
2025-04-01T06:46:14.152298
{ "authors": [ "akupila", "dwelch2344" ], "repo": "wercker/stern", "url": "https://github.com/wercker/stern/issues/21", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1041482340
chore: release 0.2.1 :robot: I have created a release *beep* *boop* 0.2.1 (2021-11-01) Bug Fixes server: fix error in the periodic task when pgp key not generated (7a29a8e) This PR was generated with Release Please. See documentation. :robot: Release is at https://github.com/werf/trdl/releases/tag/v0.2.1 :sunflower:
gharchive/pull-request
2021-11-01T18:49:03
2025-04-01T06:46:14.158333
{ "authors": [ "flant-team-sysdev" ], "repo": "werf/trdl", "url": "https://github.com/werf/trdl/pull/101", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
332427068
String after < is completely removed, if it is not followed by a space SNIPPET TO REPRODUCE const htmlToText = require('html-to-text'); let textResponse = htmlToText.fromString('<p>there are definitely <10,000 terrestrial planets in the universe. Only few of them would be habitable for future human.</p>', {     wordwrap: false }); console.log(textResponse); EXPECTED there are definitely <10,000 terrestrial planets in the universe. Only few of them would be habitable for future human. ACTUAL OUTPUT there are definitely The problem is that < is interpreted as an opening tag you need to replace it with &lt;. This is not a problem of this module the problems seems to be the used html parser. Cannot replace all < with &lt; as the input is dynamic and I need to preserve both, the HTML text as well as plain text. It wouldn't be a problem with correct use of punctuation by the user 😄 Anyway, I will try to do a workaround on this and will post once I am done. This particular example doesn't reproduce in version 7. htmlparser2 got smarter recently and doesn't consider <10,000 ... as a tag anymore. Still, it's not perfect and can be confused in other situations, such as <ten thousand .... You know what? Even Blink (Chrome's engine) is confused by <ten thousand .... I suppose it might be a performance optimization - being ready to unroll the parser state when something doesn't make sense might be costly and not worth it on a scale. HTML spec also doesn't seem to be helpful - it is really permissive about tag attributes and doesn't even ban < character from occurrence inside a tag. It requires some effort to collect the behavior across numerous JS HTML parsers. So far I know that Angular has a particularly smart parser, but that's probably not a great dependency for a project like html-to-text. The majority seems to allow out-of-spec stuff such as non-alphanumeric tag names, much like Blink. Ok, now I'm pretty confident there is no parser to switch to in order to address this issue. https://astexplorer.net/ contains most of the ones worth looking, and I made I PR there for the only one missing. There are more projects but those are either unhealthy or reusing one of the parsers such as parse5. @angular/compiler contains a nice parser but in itself it doesn't look like a good dependency. Forking it might be a way to go but I'm not convinced it is the right way to go. I would prefer not to maintain a parser too... If there is a nice example on how a certain html fragment should be interpreted according to the spec and how it is different in AST explorer - that better be filed upstream (in the parser repo, htmlparser2). I'll keep this issue open as a reference but I don't have any more to do about it, for now at least. I am facing the same issue even if my html being passed has $lt; instead of <. My html: ` Point 1 - this is point 1 Point 2 - <this is point 2> ` Output completely skips this is point 2 @sairupesh I can't reproduce this. Sounds like you're unescaping html somewhere in your pipeline before calling html-to-text. const text = htmlToText( `<div> <ul> <li><i>Point 1 - this is point 1</i></li> <li><span style="font-weight: 700;">Point 2 - &lt;this is point 2&gt;</span></li> </ul> </div>` ); console.log(text); * Point 1 - this is point 1 * Point 2 - <this is point 2>
gharchive/issue
2018-06-14T14:26:32
2025-04-01T06:46:14.168112
{ "authors": [ "KillyMXI", "ajayRaghav37", "mlegenhausen", "sairupesh" ], "repo": "werk85/node-html-to-text", "url": "https://github.com/werk85/node-html-to-text/issues/160", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
916887357
Update munit to 0.7.26 Updates org.scalameta:munit from 0.7.22 to 0.7.26. GitHub Release Notes - Version Diff I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scalameta", artifactId = "munit" } ] labels: test-library-update, semver-patch Codecov Report Merging #126 (c1f081d) into master (c68d563) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #126 +/- ## ======================================= Coverage 35.24% 35.24% ======================================= Files 21 21 Lines 1132 1132 Branches 18 18 ======================================= Hits 399 399 Misses 733 733 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update c68d563...c1f081d. Read the comment docs. Superseded by #136.
gharchive/pull-request
2021-06-10T04:40:09
2025-04-01T06:46:14.244772
{ "authors": [ "codecov-commenter", "scala-steward" ], "repo": "weso/shacl-s", "url": "https://github.com/weso/shacl-s/pull/126", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2044893069
832203201-work3 实现了基本的商品和订单的增删改查,功能还不够完善,正在改进中,主要用到mysql语言和jdbc连接,其他还在学,bonus努力中 提交要求看群公告
gharchive/pull-request
2023-12-16T18:20:31
2025-04-01T06:46:14.245828
{ "authors": [ "Poldroc", "wyxbnbl555" ], "repo": "west2-online-reserve/collection-java", "url": "https://github.com/west2-online-reserve/collection-java/pull/235", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2110807836
[COMP]: InView: off-screen exit animations should be instant We should set the animation durations to zero for exit transitions, only when matchRootMarginToThreshold={true}. Currently, if you have even a slight duration on your InView animations, you can quickly scroll back up and see them in a half-transitioned state. I'm going to treat this one as blocked until this issue is resolved on the wethegit/react-hooks package: https://github.com/wethegit/react-hooks/issues/16
gharchive/issue
2024-01-31T19:25:50
2025-04-01T06:46:14.325117
{ "authors": [ "andrewrubin" ], "repo": "wethegit/component-library", "url": "https://github.com/wethegit/component-library/issues/102", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2160343913
Cannot add chains with empty blockExplorers to MetaMask Describe the bug When attempting to add the Foundry chain to MetaMask, I encounter the following error: MetaMask - RPC Error: Expected null or array with at least one valid string HTTPS URL 'blockExplorerUrl'. I discovered that adding the blockExplorers parameter resolves the issue. Wagmi configures wallet_addEthereumChain's blockExplorerUrls based on the chain's blockExplorers. The current Wagmi implementation leads to an empty array for blockExplorerUrls when chains lack blockExplorers. https://github.com/wevm/wagmi/blob/4ba761df16ceaa4746ccb0705ea8e514630b1c1b/packages/core/src/connectors/injected.ts#L373-L393 According to MetaMask documentation, they require "One or more URLs", indicating that the empty array triggers the error. I propose excluding blockExplorerUrls for chains that do not have blockExplorers. Link to Minimal Reproducible Example No response Steps To Reproduce Create a project: pnpm create wagmi@2.0.10 --template vite-vanilla Apply this diff: diff --git a/src/main.ts b/src/main.ts index 257488e..47c0748 100644 --- a/src/main.ts +++ b/src/main.ts @@ -1,11 +1,17 @@ -import { connect, disconnect, reconnect, watchAccount } from '@wagmi/core' +import { connect, disconnect, reconnect, watchAccount, switchChain } from '@wagmi/core' import { Buffer } from 'buffer' +import { foundry } from '@wagmi/core/chains' import './style.css' import { config } from './wagmi' globalThis.Buffer = Buffer +const button = document.createElement('button') +button.textContent = 'Switch to foundry' +button.onclick = () => switchChain(config, { chainId: foundry.id }) +document.body.appendChild(button) + document.querySelector<HTMLDivElement>('#app')!.innerHTML = ` <div> <div id="account"> diff --git a/src/wagmi.ts b/src/wagmi.ts index 3bce606..ebadbf5 100644 --- a/src/wagmi.ts +++ b/src/wagmi.ts @@ -1,16 +1,15 @@ -import { coinbaseWallet, injected, walletConnect } from '@wagmi/connectors' +import { injected } from '@wagmi/connectors' import { http, createConfig } from '@wagmi/core' -import { mainnet, sepolia } from '@wagmi/core/chains' +import { mainnet, sepolia, foundry } from '@wagmi/core/chains' export const config = createConfig({ - chains: [mainnet, sepolia], + chains: [mainnet, sepolia, foundry], connectors: [ injected(), - coinbaseWallet({ appName: 'Create Wagmi' }), - walletConnect({ projectId: import.meta.env.VITE_WC_PROJECT_ID }), ], transports: { [mainnet.id]: http(), [sepolia.id]: http(), + [foundry.id]: http(), }, }) Run the project, connect MetaMask, and switch to the Foundry chain; then, the following error occurs: MetaMask - RPC Error: Expected null or array with at least one valid string HTTPS URL 'blockExplorerUrl'. By modifying as follows: diff --git a/src/wagmi.ts b/src/wagmi.ts index ebadbf5..1a91b83 100644 --- a/src/wagmi.ts +++ b/src/wagmi.ts @@ -2,8 +2,10 @@ import { injected } from '@wagmi/connectors' import { http, createConfig } from '@wagmi/core' import { mainnet, sepolia, foundry } from '@wagmi/core/chains' +const f = { ...foundry, blockExplorers: { default: { name: "", url: "https://example.com" } } } + export const config = createConfig({ - chains: [mainnet, sepolia, foundry], + chains: [mainnet, sepolia, f], connectors: [ injected(), ], The error is resolved, and everything works as expected: Wagmi Version 2.6.5 Viem Version 2.7.16 TypeScript Version 5.3.3 Check existing issues [X] I checked there isn't already an issue for the bug I encountered. Anything else? No response Thanks for the detailed issue @tash-2s! You are welcome to submit a PR (excluding blockExplorerUrls for chains that do not have blockExplorers) if you want credit. Otherwise, will get to this soon. @tmm Thanks for taking a look. Let me try quickly!
gharchive/issue
2024-02-29T04:36:27
2025-04-01T06:46:14.335057
{ "authors": [ "tash-2s", "tmm" ], "repo": "wevm/wagmi", "url": "https://github.com/wevm/wagmi/issues/3652", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
313518338
Add validator tests Fixes #51 Coverage decreased (-18.6%) to 45.712% when pulling 886cfaf1af63059f339a1363be0f7e65ae7a22f8 on bug/51 into a7c9f27d0d718f0105287b5e468f06d9ee8e53c8 on master. Coverage decreased (-18.6%) to 45.712% when pulling 886cfaf1af63059f339a1363be0f7e65ae7a22f8 on bug/51 into a7c9f27d0d718f0105287b5e468f06d9ee8e53c8 on master. Coverage decreased (-18.6%) to 45.712% when pulling 886cfaf1af63059f339a1363be0f7e65ae7a22f8 on bug/51 into a7c9f27d0d718f0105287b5e468f06d9ee8e53c8 on master. Coverage decreased (-18.6%) to 45.712% when pulling 886cfaf1af63059f339a1363be0f7e65ae7a22f8 on bug/51 into a7c9f27d0d718f0105287b5e468f06d9ee8e53c8 on master. Coverage decreased (-18.6%) to 45.712% when pulling 886cfaf1af63059f339a1363be0f7e65ae7a22f8 on bug/51 into a7c9f27d0d718f0105287b5e468f06d9ee8e53c8 on master. Coverage decreased (-18.6%) to 45.712% when pulling 886cfaf1af63059f339a1363be0f7e65ae7a22f8 on bug/51 into a7c9f27d0d718f0105287b5e468f06d9ee8e53c8 on master.
gharchive/pull-request
2018-04-11T23:09:41
2025-04-01T06:46:14.342936
{ "authors": [ "coveralls", "philsturgeon" ], "repo": "wework/speccy", "url": "https://github.com/wework/speccy/pull/52", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1990369477
🛑 Mumble (HK) is down In 5e26564, Mumble (HK) (hk.keepstar.net) was down: HTTP code: 0 Response time: 0 ms Resolved: Mumble (HK) is back up in 8789b85 after 11 minutes.
gharchive/issue
2023-11-13T10:34:43
2025-04-01T06:46:14.349638
{ "authors": [ "wfjsw" ], "repo": "wfjsw/status-winterco-org", "url": "https://github.com/wfjsw/status-winterco-org/issues/546", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2070537269
从别的apps过来 Content 我之前用的是nclientv2,最近想换成pica,如果我把在nclientv2下载过的漫画搬到pica可行吗? 不可行
gharchive/issue
2024-01-08T14:08:30
2025-04-01T06:46:14.354433
{ "authors": [ "HayashiSora", "wgh136" ], "repo": "wgh136/PicaComic", "url": "https://github.com/wgh136/PicaComic/issues/352", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
397017157
Add support for Lustre 2.12.1 This patch adds support for Lustre 2.12.0 Fixes #743 Signed-off-by: Joe Grund jgrund@whamcloud.io This is superseded by #825 Pieces that was unrelated was split out to #829
gharchive/pull-request
2019-01-08T17:51:36
2025-04-01T06:46:14.359134
{ "authors": [ "jgrund", "utopiabound" ], "repo": "whamcloud/integrated-manager-for-lustre", "url": "https://github.com/whamcloud/integrated-manager-for-lustre/pull/773", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1650499274
How to without serializer So for reasons I have the endpoint and I do some various manipulation aggregating various stuff (without a model) and generating a dict with the data that I want in the exported file. An example of my code: class Export(XLSXFileMixin, Viewset): queryset = "" renderer_classes = (XLSXRenderer,) filename = "export.xlsx" http_method_names = [ "get", ] def list(self, request): new_output = {} if self.request.query_params.get("export") == "data": self.column_header = { "titles": [ [...] ] } [my code here] return self.response(new_output) I get AssertionError at /api/export/ 'Export' should either include a `serializer_class` attribute, or override the `get_serializer_class()` method. So What is the right serializer to use to get just this dict in the generated file? The rendering code relies on customizing the Serializer. Can you try passing the DRF generic one? from rest_framework.serializers import Serializer class Export(XLSXFileMixin, Viewset): queryset = "" serializer_class = Serializer renderer_classes = (XLSXRenderer,) filename = "export.xlsx" http_method_names = [ "get", ] def list(self, request): new_output = {} if self.request.query_params.get("export") == "data": self.column_header = { "titles": [ [...] ] } [my code here] return self.response(new_output) I tried with that one but didn't worked. Seems that required the various fields (with a field that doesn't exist in the serializer itself) define otherwise wasn't generating anything. Investigating the code seems that for any row and any value check if exist in the fields, so share like a json with the data is not working for this package (also the column headers were defined). At the end I did it manually generating a csv for the moment.
gharchive/issue
2023-04-01T14:31:43
2025-04-01T06:46:14.362328
{ "authors": [ "FlipperPA", "Mte90" ], "repo": "wharton/drf-excel", "url": "https://github.com/wharton/drf-excel/issues/75", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
922672186
Update but still shows 1.0.2 (possible min ver issue) Describe the bug After updating, version still shows 1.0.2. Even after uninstalling/reinstalling the mod. Screenshots Additional context After looking at the module.json file, I see the minimum Foundry version is listed as 0.85. I am still running 0.79 as there are mods I use that have not been updated yet. Does current ver 1.1.4 of your mod work with 0.79? If it does, updating the "minimumCoreVersion" to drop it back to .079 should resolve the issue. If it doesn't work with 0.79 you can close this out. Sucks though as I was the user that gave you feedback on the new Merchant Description feature and I can't use it! lol Did a little testing. Downloaded the zip file for 1.1.4 and just changed the minimumversion to 0.7.9. All seems to be working OK except the description. I add a description and after hitting save, it disappears. I would just write it off as the reason being incompatibility except for 1 oddity - it shows up for a second after saving and then disappears. See below for a better explanation. So I'm not sure if it actually is just a compatibility issue with 0.7.9 or if something else is going on here. Nevermind @whelan, I'm closing the issue. Seems there are also issues with player permissions when trying to open the shop that don't exist for me in 1.0.2. Guess I'll have to wait until I upgrade Foundry to use the new features. :-( @Shuggaloaf I'm sorry for you. But I'm also struggling a bit with some of the 0.8.x where they have changed some of the locations for actor data.
gharchive/issue
2021-06-16T14:15:44
2025-04-01T06:46:14.466609
{ "authors": [ "Shuggaloaf", "whelan" ], "repo": "whelan/fvtt-merchant-sheet-npc", "url": "https://github.com/whelan/fvtt-merchant-sheet-npc/issues/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1892281839
Picture-in-picture toggle event Hi everyone, Is there any event for a Picture-in-picture toggle? I want the main video element to hide while triggering Picture-in-picture to access my web page below. Thanks in advance. Hi @sivajiDsmart, we've added a PiP toggle event, you can listen for it like so: const elm = document.querySelector("whereby-embed"); This works perfectly. Thank you so much @kevinwhereby
gharchive/issue
2023-09-12T11:10:00
2025-04-01T06:46:14.468938
{ "authors": [ "kevinwhereby", "sivajiDsmart" ], "repo": "whereby/browser-sdk", "url": "https://github.com/whereby/browser-sdk/issues/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
62178978
Unable to set 'response.body' in 'after_request' callback In order to conform to the REST conventions of a Rails app - specifically a 'GET index' results contain a 'root' node corresponding to the 'underscorerized / pluralized' name of the resource - I need to add an 'after_request' callback. While experimenting with implementing this callback - I thought I'd try the 'base' case listed in the documentation ... class Person < ActiveRestClient::Base after_request :fix_empty_content private def fix_empty_content(name, response) if response.status == 204 && response.body.blank? response.body = '{"empty": true}' end end end This doesn't work and returns this error ... NoMethodError: undefined method body=' for #<Faraday::Response:0x007f9361ac9058> from /Users/dekhaus/projects/rails4/groot/app/models/wip_model.rb:26:in adjust_body' Presumably - because no setter exists for the 'body' attribute. Regards, Dave This is how I change the body of the request before it is being serialized into ARC objects (using proxy): module Proxies class ProxyBase < ActiveRestClient::ProxyBase def self.underscore_key(string) string.gsub(/::/, '/') .gsub(/([A-Z]+)([A-Z][a-z])/, '\1_\2') .gsub(/([a-z\d])([A-Z])/, '\1_\2') .tr('-', '_') .downcase end def self.convert_hash_keys(value) case value when Array value.map { |v| convert_hash_keys(v) } when Hash Hash[value.map { |k, v| [underscore_key(k), convert_hash_keys(v)] }] else value end end # converts camelCase keys to snake_case symbols get(/.*/) do response = passthrough translate(response) do |body| convert_hash_keys(body) end end end end
gharchive/issue
2015-03-16T19:02:28
2025-04-01T06:46:14.483113
{ "authors": [ "dekhaus", "dkarter" ], "repo": "whichdigital/active-rest-client", "url": "https://github.com/whichdigital/active-rest-client/issues/85", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1102003057
arm更新之后进不去面板 提示ql ckeck 执行之后还是一样 拉新镜像也是这样 arm设备(树莓派)更新之后进不去面板 提示 ql ckeck 执行之后还是一样 拉新镜像也是这样没有用一直进不去 我也是这样,回退就好了 把ql check贴出来,看下pm2 logs panel的日志 开始执行... 2022-01-14 10:25:54 =====> 开始检测 ---> 1. 复制通知文件 ---> 复制一份 /ql/sample/notify.py 为 /ql/scripts/notify.py '/ql/sample/notify.py' -> '/ql/scripts/notify.py' ---> 复制一份 /ql/sample/notify.js 为 /ql/scripts/sendNotify.js '/ql/sample/notify.js' -> '/ql/scripts/sendNotify.js' ---> 通知文件复制完成 ---> 2. 复制nginx配置文件 '/ql/docker/nginx.conf' -> '/etc/nginx/nginx.conf' '/ql/docker/front.conf' -> '/etc/nginx/conf.d/front.conf' ---> 配置文件复制完成 =====> 检测面板 <script src="https://gw.alipayobjects.com/os/lib/react/16.13.1/umd/react.production.min.js"></script> <script src="https://gw.alipayobjects.com/os/lib/react-dom/16.13.1/umd/react-dom.production.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/darkreader@4.9.40/darkreader.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/lib/codemirror.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/shell/shell.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/python/python.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/javascript/javascript.js"></script> <script src="https://cdn.jsdelivr.net/npm/sockjs-client@1/dist/sockjs.min.js"></script> <script src="/umi.0b59d978.js"></script> =====> 面板服务启动正常 =====> 检测nginx服务 43 root 0:00 nginx: master process nginx -c /etc/nginx/nginx.conf 4116 root 0:00 nginx: worker process 4117 root 0:00 nginx: worker process 4118 root 0:00 nginx: worker process 4119 root 0:00 nginx: worker process =====> nginx服务正常 ---> pm2日志 2022-01-14T10:22:54: sql: 'INSERT INTO Envs_backup SELECT id, value, timestamp, status, position, name, remarks, createdAt, updatedAt FROM Envs;' 2022-01-14T10:22:54: }, 2022-01-14T10:22:54: original: [Error: SQLITE_CONSTRAINT: UNIQUE constraint failed: Envs_backup.id] { 2022-01-14T10:22:54: errno: 19, 2022-01-14T10:22:54: code: 'SQLITE_CONSTRAINT', 2022-01-14T10:22:54: sql: 'INSERT INTO Envs_backup SELECT id, value, timestamp, status, position, name, remarks, createdAt, updatedAt FROM Envs;' 2022-01-14T10:22:54: }, 2022-01-14T10:22:54: sql: 'INSERT INTO Envs_backup SELECT id, value, timestamp, status, position, name, remarks, createdAt, updatedAt FROM Envs;' 2022-01-14T10:22:54: } 2022-01-14T10:22:54: info: ✌️ DB loaded 2022-01-14T10:22:55: info: ✌️ Dependency Injector loaded 2022-01-14T10:22:55: info: ✌️ Express loaded 2022-01-14T10:23:00: info: ✌️ init data loaded 2022-01-14T10:23:00: info: ✌️ link deps 2022-01-14T10:23:00: info: ✌️ Sock loaded 2022-01-14T10:23:00: info: 2022-01-14T10:23:00: ################################################ 2022-01-14T10:23:00: 🛡️ Server listening on port: 5600 🛡️ 2022-01-14T10:23:00: ################################################ 2022-01-14T10:23:00: 2022-01-14T10:19:14: path: '/ql/db/crontab.db~', 2022-01-14T10:19:14: dest: '/ql/db/crontab.db' 2022-01-14T10:19:14: } 2022-01-14T10:19:14: Error: ENOENT: no such file or directory, rename '/ql/db/crontab.db~' -> '/ql/db/crontab.db' 2022-01-14T10:19:14: [Error: ENOENT: no such file or directory, rename '/ql/db/auth.db~' -> '/ql/db/auth.db'] { 2022-01-14T10:19:14: errno: -2, 2022-01-14T10:19:14: code: 'ENOENT', 2022-01-14T10:19:14: syscall: 'rename', 2022-01-14T10:19:14: path: '/ql/db/auth.db~', 2022-01-14T10:19:14: dest: '/ql/db/auth.db' 2022-01-14T10:19:14: } 2022-01-14T10:19:14: Error: ENOENT: no such file or directory, rename '/ql/db/auth.db~' -> '/ql/db/auth.db' 2022-01-14T10:19:14: [Error: ENOENT: no such file or directory, rename '/ql/db/app.db~' -> '/ql/db/app.db'] { 2022-01-14T10:19:14: errno: -2, 2022-01-14T10:19:14: code: 'ENOENT', 2022-01-14T10:19:14: syscall: 'rename', 2022-01-14T10:19:14: path: '/ql/db/app.db~', 2022-01-14T10:19:14: dest: '/ql/db/app.db' 2022-01-14T10:19:14: } 2022-01-14T10:19:14: Error: ENOENT: no such file or directory, rename '/ql/db/app.db~' -> '/ql/db/app.db' =====> 检测后台 {"code":200,"data":{"username":"admin"}} =====> 后台服务启动正常 =====> 检测结束 把ql check贴出来,看下pm2 logs panel的日志 日志如下 好像是依赖安装不成功 root@qinglong:/ql $ ql check 开始执行... 2022-01-14 10:29:50 =====> 开始检测 ---> 1. 复制通知文件 ---> 复制一份 /ql/sample/notify.py 为 /ql/scripts/notify.py '/ql/sample/notify.py' -> '/ql/scripts/notify.py' ---> 复制一份 /ql/sample/notify.js 为 /ql/scripts/sendNotify.js '/ql/sample/notify.js' -> '/ql/scripts/sendNotify.js' ---> 通知文件复制完成 ---> 2. 复制nginx配置文件 '/ql/docker/nginx.conf' -> '/etc/nginx/nginx.conf' '/ql/docker/front.conf' -> '/etc/nginx/conf.d/front.conf' ---> 配置文件复制完成 =====> 检测面板 <script src="https://gw.alipayobjects.com/os/lib/react/16.13.1/umd/react.production.min.js"></script> <script src="https://gw.alipayobjects.com/os/lib/react-dom/16.13.1/umd/react-dom.production.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/darkreader@4.9.40/darkreader.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/lib/codemirror.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/shell/shell.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/python/python.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/javascript/javascript.js"></script> <script src="https://cdn.jsdelivr.net/npm/sockjs-client@1/dist/sockjs.min.js"></script> <script src="/umi.0b59d978.js"></script> =====> 面板服务启动正常 =====> 检测nginx服务 42 root 0:00 nginx: master process nginx -c /etc/nginx/nginx.conf 474377 root 0:00 nginx: worker process 474378 root 0:00 nginx: worker process 474379 root 0:00 nginx: worker process 474380 root 0:00 nginx: worker process =====> nginx服务正常 ---> pm2日志 2022-01-12T04:33:22: ################################################ 2022-01-12T04:33:22: 2022-01-13T04:35:46: info: ✌️ Server need close 2022-01-13T04:35:46: info: ✌️ Server closed 2022-01-13T04:35:49: info: ✌️ Init file down 2022-01-13T04:35:49: info: ✌️ Sentry loaded 2022-01-13T11:47:18: info: ✌️ Init file down 2022-01-13T11:47:18: info: ✌️ Sentry loaded 2022-01-13T11:53:28: info: ✌️ Init file down 2022-01-13T11:53:29: info: ✌️ Sentry loaded 2022-01-13T11:57:58: info: ✌️ Init file down 2022-01-13T11:57:59: info: ✌️ Sentry loaded 2022-01-13T12:09:06: info: ✌️ Init file down 2022-01-13T12:09:06: info: ✌️ Sentry loaded 2022-01-13T12:10:26: info: ✌️ Init file down 2022-01-13T12:10:27: info: ✌️ Sentry loaded 2022-01-13T12:14:16: info: ✌️ Init file down 2022-01-13T12:14:16: info: ✌️ Sentry loaded 2022-01-14T04:33:16: info: ✌️ Init file down 2022-01-14T04:33:17: info: ✌️ Sentry loaded 2022-01-14T04:33:18: at new SqliteDialect (/ql/node_modules/.pnpm/sequelize@7.0.0-next.1/node_modules/sequelize/lib/dialects/sqlite/index.js:14:30) 2022-01-14T04:33:18: at new Sequelize (/ql/node_modules/.pnpm/sequelize@7.0.0-next.1/node_modules/sequelize/lib/sequelize.js:318:20) 2022-01-14T04:33:18: at Object. (/ql/back/data/index.ts:4:26) 2022-01-14T04:33:18: at Module._compile (internal/modules/cjs/loader.js:1072:14) 2022-01-14T04:33:18: at Object.Module._extensions..js (internal/modules/cjs/loader.js:1101:10) 2022-01-14T04:33:18: at Module.load (internal/modules/cjs/loader.js:937:32) 2022-01-14T04:33:18: at Function.Module._load (internal/modules/cjs/loader.js:778:12) 2022-01-14T04:33:18: at Module.require (internal/modules/cjs/loader.js:961:19) 2022-01-14T04:33:18: at Module.Hook._require.Module.require (/usr/local/pnpm-global/5/node_modules/.pnpm/require-in-the-middle@5.1.0/node_modules/require-in-the-middle/index.js:80:39) 2022-01-14T04:33:18: at require (internal/modules/cjs/helpers.js:92:18) 2022-01-14T04:33:18: at Object. (/ql/back/data/env.ts:1:1) 2022-01-14T04:33:18: at Module._compile (internal/modules/cjs/loader.js:1072:14) 2022-01-14T04:33:18: at Object.Module._extensions..js (internal/modules/cjs/loader.js:1101:10) 2022-01-14T04:33:18: at Module.load (internal/modules/cjs/loader.js:937:32) 2022-01-14T04:33:18: at Function.Module._load (internal/modules/cjs/loader.js:778:12) 2022-01-14T04:33:18: at Module.require (internal/modules/cjs/loader.js:961:19) 2022-01-14T04:33:18: at Module.Hook._require.Module.require (/usr/local/pnpm-global/5/node_modules/.pnpm/require-in-the-middle@5.1.0/node_modules/require-in-the-middle/index.js:80:39) 2022-01-14T04:33:18: at require (internal/modules/cjs/helpers.js:92:18) 2022-01-14T04:33:18: at Object. (/ql/back/loaders/db.ts:6:1) 2022-01-14T04:33:18: at Module._compile (internal/modules/cjs/loader.js:1072:14) =====> 检测后台 =====> 后台服务异常,重置基础环境并重启后台 ---> 1. 开始检测配置文件 ---> 配置文件检测完成 ---> 2. 开始安装青龙依赖 检测到 /ql 的依赖包有变化,运行 npm install... Progress: resolved 1, reused 0, downloaded 0, added 0 ╭──────────────────────────────────────────────────────────────────╮ │ │ │ Update available! 6.15.1 → 6.26.1. │ │ Changelog: https://github.com/pnpm/pnpm/releases/tag/v6.26.1 │ │ Run pnpm add -g pnpm to update. │ │ │ │ Follow @pnpmjs for updates: https://twitter.com/pnpmjs │ │ │ ╰──────────────────────────────────────────────────────────────────╯ Progress: resolved 31, reused 24, downloaded 0, added 0 Progress: resolved 78, reused 71, downloaded 0, added 0 Progress: resolved 79, reused 72, downloaded 0, added 0 Progress: resolved 104, reused 76, downloaded 0, added 0 Progress: resolved 155, reused 79, downloaded 0, added 0 Progress: resolved 223, reused 100, downloaded 0, added 0 Progress: resolved 250, reused 105, downloaded 0, added 0 Progress: resolved 266, reused 114, downloaded 0, added 0 Progress: resolved 288, reused 126, downloaded 0, added 0 Progress: resolved 304, reused 134, downloaded 0, added 0 Progress: resolved 324, reused 152, downloaded 0, added 0 Progress: resolved 342, reused 164, downloaded 0, added 0 Progress: resolved 370, reused 193, downloaded 0, added 0 Progress: resolved 393, reused 216, downloaded 0, added 0 Progress: resolved 396, reused 219, downloaded 0, added 0 Progress: resolved 410, reused 233, downloaded 0, added 0 Progress: resolved 464, reused 287, downloaded 0, added 0 Progress: resolved 530, reused 353, downloaded 0, added 0 Progress: resolved 531, reused 354, downloaded 0, added 0 Progress: resolved 534, reused 414, downloaded 0, added 0 Progress: resolved 595, reused 594, downloaded 0, added 0 Progress: resolved 664, reused 663, downloaded 0, added 0 Progress: resolved 720, reused 719, downloaded 0, added 0 Progress: resolved 790, reused 789, downloaded 0, added 0 Progress: resolved 819, reused 818, downloaded 0, added 0 Progress: resolved 873, reused 872, downloaded 0, added 0 Progress: resolved 997, reused 996, downloaded 0, added 0 Progress: resolved 1051, reused 1050, downloaded 0, added 0 Progress: resolved 1147, reused 1144, downloaded 0, added 0 Progress: resolved 1228, reused 1225, downloaded 0, added 0 Progress: resolved 1318, reused 1315, downloaded 0, added 0 Progress: resolved 1428, reused 1426, downloaded 0, added 0 Progress: resolved 1524, reused 1522, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1598, reused 1596, downloaded 0, added 0 Packages: +388 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1598, reused 1596, downloaded 0, added 33 Progress: resolved 1598, reused 1596, downloaded 0, added 388, done .../node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../node_modules/sqlite3 install: gyp info spawn args [ .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../node_modules/sqlite3 install: gyp info spawn args '-f', .../node_modules/sqlite3 install: gyp info spawn args 'make', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../node_modules/sqlite3 install: gyp info spawn args 'build', .../node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../node_modules/sqlite3 install: gyp info spawn args ] .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info spawn make .../node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../node_modules/sqlite3 install: AR(target) Release/obj.target/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../node_modules/sqlite3 install: COPY Release/nothing.a .../node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../node_modules/sqlite3 install: /bin/sh: python: not found .../node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: gyp ERR! build error .../node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../node_modules/sqlite3 install: gyp ERR! not ok .../node_modules/sqlite3 install: node-pre-gyp ERR! build error .../node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. Progress: resolved 1, reused 0, downloaded 0, added 0 Progress: resolved 49, reused 48, downloaded 0, added 0 Progress: resolved 76, reused 70, downloaded 0, added 0 Progress: resolved 79, reused 78, downloaded 0, added 0 Progress: resolved 121, reused 103, downloaded 0, added 0 Progress: resolved 199, reused 154, downloaded 0, added 0 Progress: resolved 241, reused 170, downloaded 0, added 0 Progress: resolved 315, reused 190, downloaded 0, added 0 Progress: resolved 369, reused 209, downloaded 0, added 0 Progress: resolved 429, reused 270, downloaded 0, added 0 Progress: resolved 510, reused 355, downloaded 0, added 0 Progress: resolved 512, reused 358, downloaded 0, added 0 Progress: resolved 573, reused 554, downloaded 0, added 0 Progress: resolved 662, reused 661, downloaded 0, added 0 Progress: resolved 729, reused 728, downloaded 0, added 0 Progress: resolved 770, reused 769, downloaded 0, added 0 Progress: resolved 795, reused 794, downloaded 0, added 0 Progress: resolved 873, reused 872, downloaded 0, added 0 Progress: resolved 979, reused 978, downloaded 0, added 0 Progress: resolved 1060, reused 1058, downloaded 0, added 0 Progress: resolved 1153, reused 1150, downloaded 0, added 0 Progress: resolved 1243, reused 1241, downloaded 0, added 0 Progress: resolved 1339, reused 1334, downloaded 0, added 0 Progress: resolved 1430, reused 1428, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1478, reused 1476, downloaded 0, added 0 Packages: +364 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1478, reused 1476, downloaded 0, added 1 Progress: resolved 1478, reused 1476, downloaded 0, added 127, done .../sqlite3@5.0.2/node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-f', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'make', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'build', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args ] .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn make .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../sqlite3@5.0.2/node_modules/sqlite3 install: AR(target) Release/obj.target/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: COPY Release/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../sqlite3@5.0.2/node_modules/sqlite3 install: /bin/sh: python: not found .../sqlite3@5.0.2/node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. 安装 /ql 的依赖包运行不成功,再次尝试一遍... 运行 npm install... Progress: resolved 1, reused 0, downloaded 0, added 0 Progress: resolved 42, reused 35, downloaded 0, added 0 Progress: resolved 78, reused 71, downloaded 0, added 0 Progress: resolved 79, reused 72, downloaded 0, added 0 Progress: resolved 121, reused 77, downloaded 0, added 0 Progress: resolved 183, reused 79, downloaded 0, added 0 Progress: resolved 264, reused 103, downloaded 0, added 0 Progress: resolved 361, reused 183, downloaded 0, added 0 Progress: resolved 434, reused 257, downloaded 0, added 0 Progress: resolved 506, reused 329, downloaded 0, added 0 Progress: resolved 533, reused 356, downloaded 0, added 0 Progress: resolved 545, reused 527, downloaded 0, added 0 Progress: resolved 639, reused 638, downloaded 0, added 0 Progress: resolved 714, reused 713, downloaded 0, added 0 Progress: resolved 780, reused 779, downloaded 0, added 0 Progress: resolved 838, reused 837, downloaded 0, added 0 Progress: resolved 925, reused 923, downloaded 0, added 0 Progress: resolved 1049, reused 1046, downloaded 0, added 0 Progress: resolved 1137, reused 1134, downloaded 0, added 0 Progress: resolved 1231, reused 1228, downloaded 0, added 0 Progress: resolved 1321, reused 1319, downloaded 0, added 0 Progress: resolved 1427, reused 1422, downloaded 0, added 0 Progress: resolved 1534, reused 1532, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1598, reused 1596, downloaded 0, added 0 Packages: +388 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1598, reused 1596, downloaded 0, added 9 Progress: resolved 1598, reused 1596, downloaded 0, added 151, done .../node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../node_modules/sqlite3 install: gyp info spawn args [ .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../node_modules/sqlite3 install: gyp info spawn args '-f', .../node_modules/sqlite3 install: gyp info spawn args 'make', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../node_modules/sqlite3 install: gyp info spawn args 'build', .../node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../node_modules/sqlite3 install: gyp info spawn args ] .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info spawn make .../node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../node_modules/sqlite3 install: AR(target) Release/obj.target/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../node_modules/sqlite3 install: COPY Release/nothing.a .../node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../node_modules/sqlite3 install: /bin/sh: python: not found .../node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: gyp ERR! build error .../node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../node_modules/sqlite3 install: gyp ERR! not ok .../node_modules/sqlite3 install: node-pre-gyp ERR! build error .../node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. Progress: resolved 1, reused 0, downloaded 0, added 0 Progress: resolved 28, reused 27, downloaded 0, added 0 Progress: resolved 75, reused 70, downloaded 0, added 0 Progress: resolved 79, reused 78, downloaded 0, added 0 Progress: resolved 115, reused 96, downloaded 0, added 0 Progress: resolved 201, reused 154, downloaded 0, added 0 Progress: resolved 287, reused 205, downloaded 0, added 0 Progress: resolved 365, reused 241, downloaded 0, added 0 Progress: resolved 413, reused 257, downloaded 0, added 0 Progress: resolved 474, reused 318, downloaded 0, added 0 Progress: resolved 510, reused 355, downloaded 0, added 0 Progress: resolved 546, reused 419, downloaded 0, added 0 Progress: resolved 620, reused 599, downloaded 0, added 0 Progress: resolved 686, reused 682, downloaded 0, added 0 Progress: resolved 745, reused 744, downloaded 0, added 0 Progress: resolved 794, reused 793, downloaded 0, added 0 Progress: resolved 889, reused 888, downloaded 0, added 0 Progress: resolved 1020, reused 1018, downloaded 0, added 0 Progress: resolved 1108, reused 1106, downloaded 0, added 0 Progress: resolved 1198, reused 1196, downloaded 0, added 0 Progress: resolved 1288, reused 1286, downloaded 0, added 0 Progress: resolved 1391, reused 1389, downloaded 0, added 0 Progress: resolved 1478, reused 1475, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1478, reused 1476, downloaded 0, added 0 Packages: +364 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1478, reused 1476, downloaded 0, added 1 Progress: resolved 1478, reused 1476, downloaded 0, added 47, done .../sqlite3@5.0.2/node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-f', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'make', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'build', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args ] .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn make .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../sqlite3@5.0.2/node_modules/sqlite3 install: AR(target) Release/obj.target/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: COPY Release/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../sqlite3@5.0.2/node_modules/sqlite3 install: /bin/sh: python: not found .../sqlite3@5.0.2/node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. npm install 运行不成功,请进入 /ql 目录后手动运行 npm install... ---> 青龙依赖安装完成 ---> 3. 开始安装脚本依赖 检测到 /ql/scripts 的依赖包有变化,运行 npm install... Lockfile is up-to-date, resolution step is skipped Progress: resolved 1, reused 0, downloaded 0, added 0 Packages: +212 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Progress: resolved 212, reused 44, downloaded 0, added 0 Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /ql/scripts/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 212, reused 179, downloaded 0, added 174 Progress: resolved 212, reused 212, downloaded 0, added 212, done dependencies: crypto-js 4.1.1 download 8.0.0 got 11.8.2 http-server 0.12.3 qrcode-terminal 0.12.0 request 2.88.2 tough-cookie 4.0.0 tunnel 0.0.6 ws 7.5.5 ---> 脚本依赖安装完成 =====> 检测结束 执行结束... 2022-01-14 10:32:50 耗时 180 秒 我也是,提示npm安装不了 我也是同样的问题,手动执行 npm install 提示下面的内容: npm ERR! code ERESOLVE npm ERR! ERESOLVE unable to resolve dependency tree npm ERR! npm ERR! While resolving: undefined@undefined npm ERR! Found: react@17.0.2 npm ERR! node_modules/react npm ERR! dev react@"17.x" from the root project npm ERR! npm ERR! Could not resolve dependency: npm ERR! peer react@">=15.5 <=16.x" from react-codemirror2@7.2.1 npm ERR! node_modules/react-codemirror2 npm ERR! dev react-codemirror2@"^7.2.1" from the root project npm ERR! npm ERR! Fix the upstream dependency conflict, or retry npm ERR! this command with --force, or --legacy-peer-deps npm ERR! to accept an incorrect (and potentially broken) dependency resolution. npm ERR! npm ERR! See /root/.npm/eresolve-report.txt for a full report. npm ERR! A complete log of this run can be found in: npm ERR! /root/.npm/_logs/2022-01-14T06_38_01_957Z-debug.log 你们还可以ql check,我的直接System load 800%,直接死机,nanopim4,rk3399设备。无法查看日志,,,, 同样的问题,我今天手动重启了宿主机,然后就发现我后台进不去了。重新创建容器也是一样的问题。 pm2 logs panel 0|panel | 2022-01-14T19:11:33: Error: More than 10% of the data file is corrupt, the wrong beforeDeserialization hook may be used. Cautiously refusing to start NeDB to prevent dataloss 0|panel | 2022-01-14T19:11:33: at Persistence.treatRawData (/ql/node_modules/.pnpm/nedb@1.8.0/node_modules/nedb/lib/persistence.js:242:11) 0|panel | 2022-01-14T19:11:33: at /ql/node_modules/.pnpm/nedb@1.8.0/node_modules/nedb/lib/persistence.js:281:38 0|panel | 2022-01-14T19:11:33: at FSReqCallback.readFileAfterClose [as oncomplete] (node:internal/fs/read_file_context:68:3) 0|panel | 2022-01-14T19:11:33: Error: More than 10% of the data file is corrupt, the wrong beforeDeserialization hook may be used. Cautiously refusing to start NeDB to prevent dataloss 0|panel | 2022-01-14T19:11:33: at Persistence.treatRawData (/ql/node_modules/.pnpm/nedb@1.8.0/node_modules/nedb/lib/persistence.js:242:11) 0|panel | 2022-01-14T19:11:33: at /ql/node_modules/.pnpm/nedb@1.8.0/node_modules/nedb/lib/persistence.js:281:38 0|panel | 2022-01-14T19:11:33: at FSReqCallback.readFileAfterClose [as oncomplete] (node:internal/fs/read_file_context:68:3) 0|panel | 2022-01-14T19:11:34: info: ✌️ Init file down 0|panel | 2022-01-14T19:11:35: info: ✌️ Sentry loaded 0|panel | 2022-01-14T19:11:35: info: ✌️ DB loaded 把ql check贴出来,看下pm2 logs panel的日志 日志如下 好像是依赖安装不成功 root@qinglong:/ql $ ql check 开始执行... 2022-01-14 10:29:50 =====> 开始检测 ---> 1. 复制通知文件 ---> 复制一份 /ql/sample/notify.py 为 /ql/scripts/notify.py '/ql/sample/notify.py' -> '/ql/scripts/notify.py' ---> 复制一份 /ql/sample/notify.js 为 /ql/scripts/sendNotify.js '/ql/sample/notify.js' -> '/ql/scripts/sendNotify.js' ---> 通知文件复制完成 ---> 2. 复制nginx配置文件 '/ql/docker/nginx.conf' -> '/etc/nginx/nginx.conf' '/ql/docker/front.conf' -> '/etc/nginx/conf.d/front.conf' ---> 配置文件复制完成 =====> 检测面板 <script src="https://gw.alipayobjects.com/os/lib/react/16.13.1/umd/react.production.min.js"></script> <script src="https://gw.alipayobjects.com/os/lib/react-dom/16.13.1/umd/react-dom.production.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/darkreader@4.9.40/darkreader.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/lib/codemirror.min.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/shell/shell.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/python/python.js"></script> <script src="https://cdn.jsdelivr.net/npm/codemirror@5/mode/javascript/javascript.js"></script> <script src="https://cdn.jsdelivr.net/npm/sockjs-client@1/dist/sockjs.min.js"></script> <script src="/umi.0b59d978.js"></script> =====> 面板服务启动正常 =====> 检测nginx服务 42 root 0:00 nginx: master process nginx -c /etc/nginx/nginx.conf 474377 root 0:00 nginx: worker process 474378 root 0:00 nginx: worker process 474379 root 0:00 nginx: worker process 474380 root 0:00 nginx: worker process =====> nginx服务正常 ---> pm2日志 2022-01-12T04:33:22: ################################################ 2022-01-12T04:33:22: 2022-01-13T04:35:46: info: ✌️ Server need close 2022-01-13T04:35:46: info: ✌️ Server closed 2022-01-13T04:35:49: info: ✌️ Init file down 2022-01-13T04:35:49: info: ✌️ Sentry loaded 2022-01-13T11:47:18: info: ✌️ Init file down 2022-01-13T11:47:18: info: ✌️ Sentry loaded 2022-01-13T11:53:28: info: ✌️ Init file down 2022-01-13T11:53:29: info: ✌️ Sentry loaded 2022-01-13T11:57:58: info: ✌️ Init file down 2022-01-13T11:57:59: info: ✌️ Sentry loaded 2022-01-13T12:09:06: info: ✌️ Init file down 2022-01-13T12:09:06: info: ✌️ Sentry loaded 2022-01-13T12:10:26: info: ✌️ Init file down 2022-01-13T12:10:27: info: ✌️ Sentry loaded 2022-01-13T12:14:16: info: ✌️ Init file down 2022-01-13T12:14:16: info: ✌️ Sentry loaded 2022-01-14T04:33:16: info: ✌️ Init file down 2022-01-14T04:33:17: info: ✌️ Sentry loaded 2022-01-14T04:33:18: at new SqliteDialect (/ql/node_modules/.pnpm/sequelize@7.0.0-next.1/node_modules/sequelize/lib/dialects/sqlite/index.js:14:30) 2022-01-14T04:33:18: at new Sequelize (/ql/node_modules/.pnpm/sequelize@7.0.0-next.1/node_modules/sequelize/lib/sequelize.js:318:20) 2022-01-14T04:33:18: at Object. (/ql/back/data/index.ts:4:26) 2022-01-14T04:33:18: at Module._compile (internal/modules/cjs/loader.js:1072:14) 2022-01-14T04:33:18: at Object.Module._extensions..js (internal/modules/cjs/loader.js:1101:10) 2022-01-14T04:33:18: at Module.load (internal/modules/cjs/loader.js:937:32) 2022-01-14T04:33:18: at Function.Module._load (internal/modules/cjs/loader.js:778:12) 2022-01-14T04:33:18: at Module.require (internal/modules/cjs/loader.js:961:19) 2022-01-14T04:33:18: at Module.Hook._require.Module.require (/usr/local/pnpm-global/5/node_modules/.pnpm/require-in-the-middle@5.1.0/node_modules/require-in-the-middle/index.js:80:39) 2022-01-14T04:33:18: at require (internal/modules/cjs/helpers.js:92:18) 2022-01-14T04:33:18: at Object. (/ql/back/data/env.ts:1:1) 2022-01-14T04:33:18: at Module._compile (internal/modules/cjs/loader.js:1072:14) 2022-01-14T04:33:18: at Object.Module._extensions..js (internal/modules/cjs/loader.js:1101:10) 2022-01-14T04:33:18: at Module.load (internal/modules/cjs/loader.js:937:32) 2022-01-14T04:33:18: at Function.Module._load (internal/modules/cjs/loader.js:778:12) 2022-01-14T04:33:18: at Module.require (internal/modules/cjs/loader.js:961:19) 2022-01-14T04:33:18: at Module.Hook._require.Module.require (/usr/local/pnpm-global/5/node_modules/.pnpm/require-in-the-middle@5.1.0/node_modules/require-in-the-middle/index.js:80:39) 2022-01-14T04:33:18: at require (internal/modules/cjs/helpers.js:92:18) 2022-01-14T04:33:18: at Object. (/ql/back/loaders/db.ts:6:1) 2022-01-14T04:33:18: at Module._compile (internal/modules/cjs/loader.js:1072:14) =====> 检测后台 =====> 后台服务异常,重置基础环境并重启后台 ---> 1. 开始检测配置文件 ---> 配置文件检测完成 ---> 2. 开始安装青龙依赖 检测到 /ql 的依赖包有变化,运行 npm install... Progress: resolved 1, reused 0, downloaded 0, added 0 ╭──────────────────────────────────────────────────────────────────╮ │ │ │ Update available! 6.15.1 → 6.26.1. │ │ Changelog: https://github.com/pnpm/pnpm/releases/tag/v6.26.1 │ │ Run pnpm add -g pnpm to update. │ │ │ │ Follow @pnpmjs for updates: https://twitter.com/pnpmjs │ │ │ ╰──────────────────────────────────────────────────────────────────╯ Progress: resolved 31, reused 24, downloaded 0, added 0 Progress: resolved 78, reused 71, downloaded 0, added 0 Progress: resolved 79, reused 72, downloaded 0, added 0 Progress: resolved 104, reused 76, downloaded 0, added 0 Progress: resolved 155, reused 79, downloaded 0, added 0 Progress: resolved 223, reused 100, downloaded 0, added 0 Progress: resolved 250, reused 105, downloaded 0, added 0 Progress: resolved 266, reused 114, downloaded 0, added 0 Progress: resolved 288, reused 126, downloaded 0, added 0 Progress: resolved 304, reused 134, downloaded 0, added 0 Progress: resolved 324, reused 152, downloaded 0, added 0 Progress: resolved 342, reused 164, downloaded 0, added 0 Progress: resolved 370, reused 193, downloaded 0, added 0 Progress: resolved 393, reused 216, downloaded 0, added 0 Progress: resolved 396, reused 219, downloaded 0, added 0 Progress: resolved 410, reused 233, downloaded 0, added 0 Progress: resolved 464, reused 287, downloaded 0, added 0 Progress: resolved 530, reused 353, downloaded 0, added 0 Progress: resolved 531, reused 354, downloaded 0, added 0 Progress: resolved 534, reused 414, downloaded 0, added 0 Progress: resolved 595, reused 594, downloaded 0, added 0 Progress: resolved 664, reused 663, downloaded 0, added 0 Progress: resolved 720, reused 719, downloaded 0, added 0 Progress: resolved 790, reused 789, downloaded 0, added 0 Progress: resolved 819, reused 818, downloaded 0, added 0 Progress: resolved 873, reused 872, downloaded 0, added 0 Progress: resolved 997, reused 996, downloaded 0, added 0 Progress: resolved 1051, reused 1050, downloaded 0, added 0 Progress: resolved 1147, reused 1144, downloaded 0, added 0 Progress: resolved 1228, reused 1225, downloaded 0, added 0 Progress: resolved 1318, reused 1315, downloaded 0, added 0 Progress: resolved 1428, reused 1426, downloaded 0, added 0 Progress: resolved 1524, reused 1522, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1598, reused 1596, downloaded 0, added 0 Packages: +388 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1598, reused 1596, downloaded 0, added 33 Progress: resolved 1598, reused 1596, downloaded 0, added 388, done .../node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../node_modules/sqlite3 install: gyp info spawn args [ .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../node_modules/sqlite3 install: gyp info spawn args '-f', .../node_modules/sqlite3 install: gyp info spawn args 'make', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../node_modules/sqlite3 install: gyp info spawn args 'build', .../node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../node_modules/sqlite3 install: gyp info spawn args ] .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info spawn make .../node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../node_modules/sqlite3 install: AR(target) Release/obj.target/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../node_modules/sqlite3 install: COPY Release/nothing.a .../node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../node_modules/sqlite3 install: /bin/sh: python: not found .../node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: gyp ERR! build error .../node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../node_modules/sqlite3 install: gyp ERR! not ok .../node_modules/sqlite3 install: node-pre-gyp ERR! build error .../node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. Progress: resolved 1, reused 0, downloaded 0, added 0 Progress: resolved 49, reused 48, downloaded 0, added 0 Progress: resolved 76, reused 70, downloaded 0, added 0 Progress: resolved 79, reused 78, downloaded 0, added 0 Progress: resolved 121, reused 103, downloaded 0, added 0 Progress: resolved 199, reused 154, downloaded 0, added 0 Progress: resolved 241, reused 170, downloaded 0, added 0 Progress: resolved 315, reused 190, downloaded 0, added 0 Progress: resolved 369, reused 209, downloaded 0, added 0 Progress: resolved 429, reused 270, downloaded 0, added 0 Progress: resolved 510, reused 355, downloaded 0, added 0 Progress: resolved 512, reused 358, downloaded 0, added 0 Progress: resolved 573, reused 554, downloaded 0, added 0 Progress: resolved 662, reused 661, downloaded 0, added 0 Progress: resolved 729, reused 728, downloaded 0, added 0 Progress: resolved 770, reused 769, downloaded 0, added 0 Progress: resolved 795, reused 794, downloaded 0, added 0 Progress: resolved 873, reused 872, downloaded 0, added 0 Progress: resolved 979, reused 978, downloaded 0, added 0 Progress: resolved 1060, reused 1058, downloaded 0, added 0 Progress: resolved 1153, reused 1150, downloaded 0, added 0 Progress: resolved 1243, reused 1241, downloaded 0, added 0 Progress: resolved 1339, reused 1334, downloaded 0, added 0 Progress: resolved 1430, reused 1428, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1478, reused 1476, downloaded 0, added 0 Packages: +364 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1478, reused 1476, downloaded 0, added 1 Progress: resolved 1478, reused 1476, downloaded 0, added 127, done .../sqlite3@5.0.2/node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-f', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'make', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'build', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args ] .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn make .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../sqlite3@5.0.2/node_modules/sqlite3 install: AR(target) Release/obj.target/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: COPY Release/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../sqlite3@5.0.2/node_modules/sqlite3 install: /bin/sh: python: not found .../sqlite3@5.0.2/node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. 安装 /ql 的依赖包运行不成功,再次尝试一遍... 运行 npm install... Progress: resolved 1, reused 0, downloaded 0, added 0 Progress: resolved 42, reused 35, downloaded 0, added 0 Progress: resolved 78, reused 71, downloaded 0, added 0 Progress: resolved 79, reused 72, downloaded 0, added 0 Progress: resolved 121, reused 77, downloaded 0, added 0 Progress: resolved 183, reused 79, downloaded 0, added 0 Progress: resolved 264, reused 103, downloaded 0, added 0 Progress: resolved 361, reused 183, downloaded 0, added 0 Progress: resolved 434, reused 257, downloaded 0, added 0 Progress: resolved 506, reused 329, downloaded 0, added 0 Progress: resolved 533, reused 356, downloaded 0, added 0 Progress: resolved 545, reused 527, downloaded 0, added 0 Progress: resolved 639, reused 638, downloaded 0, added 0 Progress: resolved 714, reused 713, downloaded 0, added 0 Progress: resolved 780, reused 779, downloaded 0, added 0 Progress: resolved 838, reused 837, downloaded 0, added 0 Progress: resolved 925, reused 923, downloaded 0, added 0 Progress: resolved 1049, reused 1046, downloaded 0, added 0 Progress: resolved 1137, reused 1134, downloaded 0, added 0 Progress: resolved 1231, reused 1228, downloaded 0, added 0 Progress: resolved 1321, reused 1319, downloaded 0, added 0 Progress: resolved 1427, reused 1422, downloaded 0, added 0 Progress: resolved 1534, reused 1532, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1598, reused 1596, downloaded 0, added 0 Packages: +388 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1598, reused 1596, downloaded 0, added 9 Progress: resolved 1598, reused 1596, downloaded 0, added 151, done .../node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../node_modules/sqlite3 install: gyp info spawn args [ .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../node_modules/sqlite3 install: gyp info spawn args '-f', .../node_modules/sqlite3 install: gyp info spawn args 'make', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-I', .../node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3', .../node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../node_modules/sqlite3 install: gyp info spawn args 'build', .../node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../node_modules/sqlite3 install: gyp info spawn args ] .../node_modules/sqlite3 install: gyp info ok .../node_modules/sqlite3 install: gyp info it worked if it ends with ok .../node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../node_modules/sqlite3 install: gyp info spawn make .../node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../node_modules/sqlite3 install: AR(target) Release/obj.target/../../../registry.nlark.com+node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../node_modules/sqlite3 install: COPY Release/nothing.a .../node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../node_modules/sqlite3 install: /bin/sh: python: not found .../node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/build' .../node_modules/sqlite3 install: gyp ERR! build error .../node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../node_modules/sqlite3 install: gyp ERR! not ok .../node_modules/sqlite3 install: node-pre-gyp ERR! build error .../node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/registry.npmmirror.com+node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3 .../node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/registry.npmmirror.com+sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. Progress: resolved 1, reused 0, downloaded 0, added 0 Progress: resolved 28, reused 27, downloaded 0, added 0 Progress: resolved 75, reused 70, downloaded 0, added 0 Progress: resolved 79, reused 78, downloaded 0, added 0 Progress: resolved 115, reused 96, downloaded 0, added 0 Progress: resolved 201, reused 154, downloaded 0, added 0 Progress: resolved 287, reused 205, downloaded 0, added 0 Progress: resolved 365, reused 241, downloaded 0, added 0 Progress: resolved 413, reused 257, downloaded 0, added 0 Progress: resolved 474, reused 318, downloaded 0, added 0 Progress: resolved 510, reused 355, downloaded 0, added 0 Progress: resolved 546, reused 419, downloaded 0, added 0 Progress: resolved 620, reused 599, downloaded 0, added 0 Progress: resolved 686, reused 682, downloaded 0, added 0 Progress: resolved 745, reused 744, downloaded 0, added 0 Progress: resolved 794, reused 793, downloaded 0, added 0 Progress: resolved 889, reused 888, downloaded 0, added 0 Progress: resolved 1020, reused 1018, downloaded 0, added 0 Progress: resolved 1108, reused 1106, downloaded 0, added 0 Progress: resolved 1198, reused 1196, downloaded 0, added 0 Progress: resolved 1288, reused 1286, downloaded 0, added 0 Progress: resolved 1391, reused 1389, downloaded 0, added 0 Progress: resolved 1478, reused 1475, downloaded 0, added 0  WARN  @ant-design/pro-layout: @ant-design/pro-provider@1.4.20 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout: @ant-design/pro-utils@1.24.9 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @ant-design/pro-layout@6.26.6 requires a peer of antd@4.x but version 4.17.0-alpha.7 was installed.  WARN  @monaco-editor/react: @monaco-editor/loader@1.2.0 requires a peer of monaco-editor@>= 0.21.0 < 1 but none was installed.  WARN  @monaco-editor/react@4.3.1 requires a peer of monaco-editor@>= 0.25.0 < 1 but none was installed.  WARN  react-codemirror2@7.2.1 requires a peer of react@>=15.5 <=16.x but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-diff-viewer@3.1.1 requires a peer of react-dom@^15.3.0 || ^16.0.0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react@^16.0.0-0 but version 17.0.2 was installed.  WARN  react-split-pane@0.1.92 requires a peer of react-dom@^16.0.0-0 but version 17.0.2 was installed.  WARN  umi > @umijs/bundler-webpack > @umijs/bundler-utils > @umijs/types > @umijs/renderer-react: react-router-config@5.1.1 requires a peer of react-router@>=5 but none was installed. Progress: resolved 1478, reused 1476, downloaded 0, added 0 Packages: +364 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /root/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 1478, reused 1476, downloaded 0, added 1 Progress: resolved 1478, reused 1476, downloaded 0, added 47, done .../sqlite3@5.0.2/node_modules/sqlite3 install$ node-pre-gyp install --fallback-to-build .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node-pre-gyp@0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Using request for node-pre-gyp https download .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp info check checked for "/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" (not found) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http GET https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Tried to download(403): https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp WARN Pre-built binaries not found for sqlite3@5.0.2 and node@14.17.6 (node-v83 ABI, musl) (falling back to source compile with node-gyp) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp http 403 status code downloading tarball https://mapbox-node-binary.s3.amazonaws.com/sqlite3/v5.0.2/napi-v3-linux-arm64.tar.gz .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info find Python using Python version 3.8.10 found at "/usr/bin/python3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn /usr/bin/python3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/gyp/gyp_main.py', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'binding.gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-f', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'make', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build/config.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/addon.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-I', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '/root/.cache/node-gyp/14.17.6/include/node/common.gypi', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dlibrary=shared_library', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dvisibility=default', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_root_dir=/root/.cache/node-gyp/14.17.6', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_gyp_dir=/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_lib_file=/root/.cache/node-gyp/14.17.6/<(target_arch)/node.lib', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dmodule_root_dir=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Dnode_engine=v8', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--depth=.', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--no-parallel', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '--generator-output', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args 'build', .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args '-Goutput_dir=.' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args ] .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info it worked if it ends with ok .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node-gyp@8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info using node@14.17.6 | linux | arm64 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn make .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp info spawn args [ 'BUILDTYPE=Release', '-C', 'build' ] .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Entering directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: CC(target) Release/obj.target/nothing/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.o .../sqlite3@5.0.2/node_modules/sqlite3 install: AR(target) Release/obj.target/../../../node-addon-api@3.2.1/node_modules/node-addon-api/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: COPY Release/nothing.a .../sqlite3@5.0.2/node_modules/sqlite3 install: ACTION deps_sqlite3_gyp_action_before_build_target_unpack_sqlite_dep Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c .../sqlite3@5.0.2/node_modules/sqlite3 install: /bin/sh: python: not found .../sqlite3@5.0.2/node_modules/sqlite3 install: make: *** [deps/action_before_build.target.mk:13: Release/obj/gen/sqlite-autoconf-3340000/sqlite3.c] Error 127 .../sqlite3@5.0.2/node_modules/sqlite3 install: make: Leaving directory '/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/build' .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack Error: make failed with exit code: 2 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.onExit (/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/lib/build.js:194:23) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! command "/usr/local/bin/node" "/usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js" "build" "--fallback-to-build" "--module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node" "--module_name=node_sqlite3" "--module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64" "--napi_version=8" "--node_abi_napi=napi" "--napi_build_version=3" "--node_napi_label=napi-v3" .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! node-gyp -v v8.2.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! build error .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack Error: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess. (/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/lib/util/compile.js:83:29) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at ChildProcess.emit (events.js:400:28) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:1055:16) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:288:5) .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! System Linux 5.4.0-1041-raspi .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! command "/usr/local/bin/node" "/ql/node_modules/.pnpm/node-pre-gyp@0.11.0/node_modules/node-pre-gyp/bin/node-pre-gyp" "install" "--fallback-to-build" .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! cwd /ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node -v v14.17.6 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! node-pre-gyp -v v0.11.0 .../sqlite3@5.0.2/node_modules/sqlite3 install: node-pre-gyp ERR! not ok .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed to execute '/usr/local/bin/node /usr/local/lib/node_modules/pnpm/dist/node_modules/node-gyp/bin/node-gyp.js build --fallback-to-build --module=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64/node_sqlite3.node --module_name=node_sqlite3 --module_path=/ql/node_modules/.pnpm/sqlite3@5.0.2/node_modules/sqlite3/lib/binding/napi-v3-linux-arm64 --napi_version=8 --node_abi_napi=napi --napi_build_version=3 --node_napi_label=napi-v3' (1) .../sqlite3@5.0.2/node_modules/sqlite3 install: Failed  ELIFECYCLE  Command failed with exit code 1. npm install 运行不成功,请进入 /ql 目录后手动运行 npm install... ---> 青龙依赖安装完成 ---> 3. 开始安装脚本依赖 检测到 /ql/scripts 的依赖包有变化,运行 npm install... Lockfile is up-to-date, resolution step is skipped Progress: resolved 1, reused 0, downloaded 0, added 0 Packages: +212 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Progress: resolved 212, reused 44, downloaded 0, added 0 Packages are hard linked from the content-addressable store to the virtual store. Content-addressable store is at: /ql/scripts/.pnpm-store/v3 Virtual store is at: node_modules/.pnpm Progress: resolved 212, reused 179, downloaded 0, added 174 Progress: resolved 212, reused 212, downloaded 0, added 212, done dependencies: crypto-js 4.1.1 download 8.0.0 got 11.8.2 http-server 0.12.3 qrcode-terminal 0.12.0 request 2.88.2 tough-cookie 4.0.0 tunnel 0.0.6 ws 7.5.5 ---> 脚本依赖安装完成 =====> 检测结束 执行结束... 2022-01-14 10:32:50 耗时 180 秒 删了容器,更新镜像,重新启动容器 同样的问题,我今天手动重启了宿主机,然后就发现我后台进不去了。重新创建容器也是一样的问题。 pm2 logs panel 上面的版本是2.10.13,之后我删除容器和数据文件重新使用 2.11.0 版本还是同样的错,最后退回去使用 2.10.12 启动成功了。
gharchive/issue
2022-01-13T16:24:20
2025-04-01T06:46:15.010084
{ "authors": [ "2563411574", "Kaol-Fu", "Tao173", "Wxh16144", "daidaojianke", "kissfast", "lukemin", "whyour" ], "repo": "whyour/qinglong", "url": "https://github.com/whyour/qinglong/issues/1082", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
797199584
Solver iterates very slowly and does not obtain a solution The current implementation of the model using the Radau method with default tolerances does not converge to a solution. The model will run for several minutes and eventually output the warnings shown below. I terminated the program after 26 minutes because the solver had not obtained a solution. I'm not sure what's going on but suggestions on how to use SciPy's solve_ivp function for this gasifier model would be helpful. solid_phase.py:139: RuntimeWarning: invalid value encountered in power * ((kp * rhop * cpp)**(-0.5) + (ks * rhos * cps)**(-0.5))**(-1) gas_phase.py:104: RuntimeWarning: overflow encountered in power cpgg[:, j] = Acp[j] + Bcp[j] * Tg + Ccp[j] * Tg**2 + Dcp[j] * Tg**3 + Ecp[j] * Tg**4 gas_phase.py:104: RuntimeWarning: invalid value encountered in add cpgg[:, j] = Acp[j] + Bcp[j] * Tg + Ccp[j] * Tg**2 + Dcp[j] * Tg**3 + Ecp[j] * Tg**4 /numpy/core/fromnumeric.py:87: RuntimeWarning: invalid value encountered in reduce return ufunc.reduce(obj, axis, dtype, out, **passkwargs) gas_phase.py:172: RuntimeWarning: overflow encountered in double_scalars Ar = dp**3 * rhogi * (rhop - rhogi) * g / muin**2 gas_phase.py:175: RuntimeWarning: divide by zero encountered in double_scalars Umsr = (np.exp(-0.5405 * Lsi / Db) * (4.294e3 / Ar + 1.1) + 3.676e2 * Ar**(-1.5) + 1) gas_phase.py:181: RuntimeWarning: invalid value encountered in double_scalars Rrb = (1 - 0.103 * (Umsr * umf - umf)**(-0.362) * Drbs)**(-1) solid_phase.py:80: RuntimeWarning: overflow encountered in double_scalars Cs[i] = rhob_s[i] * cps[i] solid_phase.py:161: RuntimeWarning: invalid value encountered in sqrt Nud = 2 + 0.6 * Re_dc**0.5 * Pr**0.33 kinetics.py:97: RuntimeWarning: invalid value encountered in power yv = m0v * Tsv**b0v gas_phase.py:492: RuntimeWarning: invalid value encountered in sqrt Nud = 2 + 0.6 * Re_dc**0.5 * Pr**0.33 gas_phase.py:497: RuntimeWarning: invalid value encountered in power (7 - 10 * afg + 5 * afg**2) * (1 + 0.7 * Rep**0.2 * Pr**0.33) gas_phase.py:498: RuntimeWarning: invalid value encountered in power + (1.33 - 2.4 * afg + 1.2 * afg**2) * Rep**0.7 * Pr**0.33 gas_phase.py:513: RuntimeWarning: invalid value encountered in power Nuf = 0.023 * ReD**0.8 * Pr**0.4 solid_phase.py:279: RuntimeWarning: invalid value encountered in power Nup = (7 - 10 * afg + 5 * afg**2) * (1 + 0.7 * Rep**0.2 * Pr**0.33) + (1.33 - 2.4 * afg + 1.2 * afg**2) * Rep**0.7 * Pr**0.33 solid_phase.py:362: RuntimeWarning: invalid value encountered in power Nup = (7 - 10 * afg + 5 * afg**2) * (1 + 0.7 * Rep**0.2 * Pr**0.33) + (1.33 - 2.4 * afg + 1.2 * afg**2) * Rep**0.7 * Pr**0.33 solid_phase.py:368: RuntimeWarning: overflow encountered in power qwr = np.pi * Dwi * epb / ((1 - ep) / (ep * epb) + (1 - ew) / ew + 1) * sc * (Tw**4 - Tp**4) solid_phase.py:368: RuntimeWarning: invalid value encountered in subtract qwr = np.pi * Dwi * epb / ((1 - ep) / (ep * epb) + (1 - ew) / ew + 1) * sc * (Tw**4 - Tp**4) solid_phase.py:381: RuntimeWarning: invalid value encountered in power Nuf = 0.023 * ReD**0.8 * Pr**0.4 solid_phase.py:433: RuntimeWarning: invalid value encountered in power 24 / Re_dc * (1 + 8.1716 * Re_dc**(0.0964 + 0.5565 * sfc) * np.exp(-4.0655 * sfc)) solid_phase.py:464: RuntimeWarning: overflow encountered in multiply + Spav[0:Ni] * v[0:Ni] / rhos[0:Ni] solid_phase.py:464: RuntimeWarning: invalid value encountered in add + Spav[0:Ni] * v[0:Ni] / rhos[0:Ni] gas_phase.py:285: RuntimeWarning: invalid value encountered in power 24 / Re_dc * (1 + 8.1716 * Re_dc**(0.0964 + 0.5565 * sfc) * np.exp(-4.0655 * sfc)) gas_phase.py:305: RuntimeWarning: overflow encountered in multiply Smgs = (3 / 4) * rhosbav * (rhog / rhos) * (Cd / ds) * np.abs(-ug - v) kinetics.py:97: RuntimeWarning: overflow encountered in power yv = m0v * Tsv**b0v kinetics.py:99: RuntimeWarning: invalid value encountered in true_divide xv = yv * Mv / np.sum(yv * Mv) The comments below are reposted from an Issue on the SciPy repository about the TR-BDF2 algorithm. The comments are from @laurent90git and are related to code in this repo. I've quickly tested it. adding printouts of the time and time step in solve_ivp shows that the solution is actually advancing. I ran out of patience after a few minutes, having reached t~1ms. As your system is likely very stiff, at least during the initial transient, explicit methods perform very poorly and will explode/overflow if you don't use tight integration tolerances. They will anyway be very inefficient. By the way, I see that you just call solve_ivp without any integration-related parameters (rtol, atol). I think you should take a closer look at these parameters and.their impact on the solution quality. With implicit methods, every integration step is very long, because, as your system is quite nonlinear, the Jacobian of your ODE system must be computed anew very often. As your system is large (1500 ODEs) and scipy does not have any information regarding its sparsity pattern, the solver calls your ODE function dydt at least 1500 times for each Jacobian update... This results in very slow performance. One thing I noticed is that, in your vector of unknowns, you have sorted the variables by type (all discrete rho_xx are contiguous for example). As the different fields interact with one another, this results in a Jacobian sparsity pattern that has components way off the diagonal. For one-dimensional models, it is usually best to sort all the variables according to the index of the mesh cell (or spatial point) that are related to. Then within each cell, the variables may be ordered as you wish, as long as the ordering is constant from one cell to the next. This way, the Jacobian will only have nonzero components very close to the diagonal. Consequently, much more efficient approaches can be used to compute this Jacobian without performing s as many function calls. You have something like 15 fields if I remember correctly, all discretise on the same mesh. In the ideal case, with the previous reordering of the unknowns, you could compute the Jacobian in 3*15=45 calls only, which is way better (I assume your spatial schemes only use a 3-points stencil, for exemple as in the second order centered finite difference for diffusive terms). This assumes that the Jacobian is not formed analytically but by finite-differences. A banded Jacobian may also enable the use of optimized linear solvers for the Newton loop, even though solve_ivp does not offer that level of customisation yet, as far as I am aware. I don't have time this weekend to edit your code to show you what I mean, but I'll try to do that next week! Overall, as far as I've investigated, the issue is not linked to the lack of a performant integration method for stiff equations. I think it's just a combination of stiffness (requiring implicit integration), non-linearity (requiring repeated Jacobian updates) and large system size (inefficient Jacobian evaluation), all of which, added to Python's overall lack of performance (compared to other codes like Fortran or others) result in long computation times. In the article you cited, I don't see any comment on the bad performance of other solvers (compared to ode23tb). Have you read that somewhere else? Hi, I've had some more time to look into this. Regarding the base code, it actually works Setting atol=trol=1e-6, Radau has obtained a solution in ~220 min: ----------------------- Solver Info ------------------------ message: The solver successfully reached the end of the integration interval. success: True nfev: 64002 njev: 251 nlu: 834 ----------------------- Results Info ----------------------- t0 0.0 tf 10.0 N 100 len t 8806 y shape (1500, 8806) Execution time = 216m 44s FYI, nlu represents the number of LU-factorisations of the Radau-method iteration matrix, which involves the Jacobian of dydt and the time step. Hence a new factorisation is performed, roughly, when the Jacobian is updated (because of poor convergence) or when the time step is changed. So here nlu>njev is expected. What's important here: 64002 evaluation of your ODE function dydt, WITHOUT counting the calls to this function that were done to compute the Jacobian (which was updated 251 times). We have seen (see the other issue I opened regarding Jacobian update speed) that each Jacobian evaluation needs 1501 calls to dydt. Hence your dydt was called 251*1501 + 640002 = 440 753 times ! This most likely represents the major part of the computation time ! Factorizing the iteration matrix and solving the non-linear systems arising from the Radau method is likely much faster (the problem size is still relatively small at 1500 unknowns). Here are the output graphs. By the way, I think you should implement a way to automatically backup the solution results, as we currently lose everything once we exit the main() function... It would have been interesting to see what the time step evolution looked like... By the way, I ran a profiler on a single Jacobian update of dydt which took 46s. 36s seconds were spent in your gas phase _calc_mix_props, of which 20s were spent doing sums. Regular memory reallocation does not seem to be a problem, which is good news ! Well, it's good to know that a solution can be obtained but more than 3 hours to get that solution is ridiculous. When I run the Matlab version of the model it provides a solution in about 8 minutes. For now, I will look at the _calc_mix_props function and see if I can get rid of the for-loops. After that I'll work on your other suggestions. Also, what did you use to profile the Python code? Are you sure your Matlab model is exactly the same ? It would be interesting that you access the number of function evaluations and Jacobian evaluations performed by òde23tb. I am not quite sure you can access them via the ODE solver output as in Python. Then maybe you could try using a global variable ncallswhich you increment in your Matlab version ofdydt. Also, make sure your also use àtol=rtol=1e-6so that the result is sort of comparable (it won't be as it is not the same integrator). Do you provide an analytical Jacobian or something similar to Matlab ? Regarding your _calc_mix_props function, I think the for loops are fine (even though you could maybe avoid them). The main issue is the repeated call to np.sum. By the way, the profiler I use is the one including in the "Spyder" IDE, which is similar to the Matlab IDE, but for Python. Actually this gives me an idea: can you profile your Matlab code ? It will then show the number of function calls, and also the inner working of the ode23tb solver, which would be quite interesting ! The Matlab model was used as the basis to develop the Python model. Other than the ODE solver, they are using the same system of equations and performing the same calculations. Below is the Matlab code which calls the ode23tb solver function. I did not provide the solver with a Jacobian matrix. The only inputs to the solver are the function that represents the ODEs, the time span which is labeled as xspan, and the initial conditions x0. tfin = 1000; % total time [s] xspan = [0, tfin]; % time span [s] rhob_b0 = 1e-12; % initial biomass concentration [kg/m³] rhob_c0 = 0; % initial char concentration [kg/m³] rhob_g0 = 0.15; % initial gas concentration [kg/m³] % Initial conditions X0(1:N) = 300; % solid fuel temperature, Ts [K] X0(N+1:2*N) = 1100; % gas temperature, Tg [K] X0(2*N+1:3*N) = rhob_b0; % biomass concentration, rhob_b [kg/m³] X0(3*N+1:4*N) = ugin; % solid fuel velocity, v [m/s] X0(4*N+1:5*N) = 0.2; % gas mass flux, mfg [kg/s-m^2] X0(5*N+1:6*N) = rhob_g0; % bulk gas concentration, rhob_g [kg/m³] X0(6*N+1:7*N) = rhob_g0; % steam concentration, rhob_h2o [kg/m³] X0(7*N+1:7*N+Ni) = 1100; % inert particle (bed) temperature, Tp [K] X0(7*N+Ni+1:8*N) = 0; % inert particle (bed) temperature in freeboard, Tp [K] X0(8*N+1:9*N) = rhob_c0; % char concentration, rhob_c [kg/m³] X0(9*N+1:10*N) = 0; % H2 concentration, rhob_h2 [kg/m³] X0(10*N+1:11*N) = 0.0; % CH4 concentration, rhob_ch4 [kg/m³] X0(11*N+1:12*N) = 0.0; % CO concentration, rhob_co [kg/m³] X0(12*N+1:13*N) = 0.0; % CO2 concentration, rhob_co2 [kg/m³] X0(13*N+1:14*N) = 0.0; % Tar concentration, rhob_t [kg/m³] X0(14*N+1:15*N) = 0.0; % total amount of release char, rhob_ca [kg/m^3] X0(15*N+1:16*N) = 1100; % wall temperature, Tw [K] [t, X] = ode23tb(@bfb_gasifier, xspan, X0); I have compared results from both the Matlab and Python versions after the first couple of time steps and everything looked similar. The trickiest part was making sure the indexing of the arrays was the same in Python as in Matlab. Matlab uses 1 for the first index and Python uses 0 for the first index. I'll try to profile the Matlab code and get more information about what the solver is doing. I ran the Matlab code using the following options for the ode23tb as shown below. It converged to a solution in 1,525 seconds (25 minutes). Notice that this solution is for a time span of 1,000 seconds. The Python version of the model that is in this repo is set to a time span of 10 seconds for the solver. opts = odeset('RelTol', 1e-6, 'AbsTol', 1e-6, 'Stats', 'on'); [t, X] = ode23tb(@bfb_gasifier, xspan, X0, opts); The Matlab ode23tb solver statistics after running the model are shown below. 74347 successful steps 136 failed attempts 272378 function evaluations 56 partial derivatives 1044 LU decompositions 256510 solutions of linear systems Ok, it's funny, ode23tb takes 10 times more steps, but evaluates the Jacobian five times less... Maybe they have a different strategy where the time step is less often varied, but I haven't heard of such a strategy for this method... The fact that your solution goes up to 1000s instead of 10s is not so important, as the system as by then reached steady-state, therefore the time step will quickly increase and it won't cost much more to go from 10s up to 1000s. Could you plot the time step evolution (log(diff(t)) as a function of physical time ? Does your solution look like the Python one ? Also, could you profile the whole Matlab resolution and export the profile result as HTML ? https://fr.mathworks.com/help/matlab/ref/profsave.html Below is a plot of the time steps where the y-axis is log scale. I also attached the profile results from Matlab which are available in the zip file. Open the file named "file0.html" to view the home page of the results. From there you should be able to dig into the functions and see the relevant lines of code. Matlab profile results: profile_results.zip @laurent90git Were you able to view the profile results? I made several changes to my code since this issue was first posted. See the dynamic model content in the README for the latest information about the code. I removed the for-loops for calculating the gas mixture properties which seems to speed up the overall solution process. Thanks @laurent90git for pointing out this issue. I also removed the class objects in favor of using modules that contain just functions. This made it easier for me to code up the differential equations and other calculations. The dynamic model actually reaches a solution now in about 16 minutes on my laptop. However, I get some overflow warnings while the solver is running (see below). Are there any suggestions on how to prevent these warnings from occurring? dyn-bfbgasf/solid_phase.py:133: RuntimeWarning: overflow encountered in power + hps * (Tp - Ts) dyn-bfbgasf/solid_phase.py:309: RuntimeWarning: overflow encountered in power - hps * (Tp - Ts) dyn-bfbgasf/solid_phase.py:393: RuntimeWarning: overflow encountered in power qwr = np.pi * Dwi * epb / ((1 - ep) / (ep * epb) + (1 - ew) / ew + 1) * sc * (Tw**4 - Tp**4) dyn-bfbgasf/gas_phase.py:230: RuntimeWarning: overflow encountered in multiply SmgV = SmgG + Smgs * (ug + v) - (Smgp - Smgg) * ug - SmgF ----------------------- Solver Info ------------------------ message The solver successfully reached the end of the integration interval. success True nfev 63758 njev 291 nlu 1188 ----------------------- Results Info ----------------------- t0 0.0 tf 1000.0 len t 8707 y shape (1600, 8707) elapsed time 16m 22s Hi @wigging sorry for the late reply. REgarding the profiling results, as far as I could tell, Matlab was actually much quicker at evaluating the ODE function, and also used less Jacobian updates, hence the better performance. I am not quite sure there's more to do about your case. I guess you can try using better precompiler for Python (Numba for example), it might improve the overall performance, also in a similar case I personnalyy had no performance improvement... Anyway it seems it is now on paar with Matlab, right ? Regarding the overflow warnings, it is likely due to the fact that the integrator sometimes takes steps that are too large and locally result in an incorrect Newton step which sort of diverges, causing overflow in your function... The Newton loop anyway does not converge so Radau tries the step once again with a lower time step such that this problem does not appear anymore, therefore your solution is not affected by this (just check that you don't have NaNs in your solution history, otherwise your solution is garbage). You could also reduce the likelihood of the problem appearing by lowering the integration tolerances rtol and atol, such that the time steps used are smaller. This will also increase the overall quality of the solution, but at the cost of an increased computational time. You can try playing around with the tolerances. 1e-6 is usually a good all-around value, but you can try lower. Hey @laurent90git. Thank you for replying again. I was worried that you gave up on this issue. I really appreciate your help. And don't worry, I'm almost ready to close out this issue so I won't have to bother you again. Anyway, regarding your last comment... Maybe I misunderstood your comment but it sounds like you tried to use Numba for this code and it didn't give you a performance improvement. Is that correct? I updated the README for running the latest dynamic model if you want to try to run it yourself. Code for the dynamic model is in the dyn-bfbgasf folder. I have compared the results to the Matlab version and everything looks similar. Although the Matlab version does not give the overflow warnings but that's probably because it's using a different ODE solver. And regarding your comment about the overflow warnings, I'm using the Radau solver with rtol=1e-6. With this setting I get the overflow warnings but the final results seem to be fine. I also tried rtol=1e-8, atol=1e-8 but this caused even more overflow warnings. It looks like it also caused some problems related to the Jacobian (see below). I eventually quit the program because it was taking too long to find a solution. dyn-bfbgasf/solid_phase.py:133: RuntimeWarning: overflow encountered in power + hps * (Tp - Ts) dyn-bfbgasf/solid_phase.py:313: RuntimeWarning: overflow encountered in power - hps * (Tp - Ts) dyn-bfbgasf/solid_phase.py:397: RuntimeWarning: overflow encountered in power qwr = np.pi * Dwi * epb / ((1 - ep) / (ep * epb) + (1 - ew) / ew + 1) * sc * (Tw**4 - Tp**4) dyn-bfbgasf/gas_phase.py:230: RuntimeWarning: overflow encountered in multiply SmgV = SmgG + Smgs * (ug + v) - (Smgp - Smgg) * ug - SmgF dyn-bfbgasf/gas_phase.py:197: RuntimeWarning: overflow encountered in multiply + Re_dc * 73.69 / (Re_dc + 5.378 * np.exp(6.2122 * sfc)) * np.exp(-5.0748 * sfc) dyn-bfbgasf/solid_phase.py:245: RuntimeWarning: overflow encountered in multiply + Re_dc * 73.69 / (Re_dc + 5.378 * np.exp(6.2122 * sfc)) * np.exp(-5.0748 * sfc) dyn-bfbgasf/solid_phase.py:126: RuntimeWarning: overflow encountered in true_divide Re_dc = abs(rhog) * abs(-ug - v) * ds / mu dyn-bfbgasf/gas_phase.py:193: RuntimeWarning: overflow encountered in true_divide Re_dc = rhog * np.abs(-ug - v) * ds / mu dyn-bfbgasf/gas_phase.py:196: RuntimeWarning: invalid value encountered in multiply 24 / Re_dc * (1 + 8.1716 * Re_dc**(0.0964 + 0.5565 * sfc) * np.exp(-4.0655 * sfc)) dyn-bfbgasf/gas_phase.py:197: RuntimeWarning: invalid value encountered in true_divide + Re_dc * 73.69 / (Re_dc + 5.378 * np.exp(6.2122 * sfc)) * np.exp(-5.0748 * sfc) dyn-bfbgasf/gas_phase.py:294: RuntimeWarning: overflow encountered in true_divide Re_dc = abs(rhog) * abs(-ug - v) * ds / mu dyn-bfbgasf/solid_phase.py:241: RuntimeWarning: overflow encountered in true_divide Re_dc = abs(rhog) * abs(-ug - v) * ds / mu dyn-bfbgasf/solid_phase.py:244: RuntimeWarning: invalid value encountered in multiply 24 / Re_dc * (1 + 8.1716 * Re_dc**(0.0964 + 0.5565 * sfc) * np.exp(-4.0655 * sfc)) dyn-bfbgasf/solid_phase.py:245: RuntimeWarning: invalid value encountered in true_divide + Re_dc * 73.69 / (Re_dc + 5.378 * np.exp(6.2122 * sfc)) * np.exp(-5.0748 * sfc) /Users/gavinw/miniconda3/lib/python3.7/site-packages/scipy/integrate/_ivp/common.py:336: RuntimeWarning: overflow encountered in multiply new_factor = NUM_JAC_FACTOR_INCREASE * factor[ind] dyn-bfbgasf/solid_phase.py:133: RuntimeWarning: invalid value encountered in add + hps * (Tp - Ts) /Users/gavinw/miniconda3/lib/python3.7/site-packages/scipy/integrate/_ivp/common.py:358: RuntimeWarning: overflow encountered in multiply factor[max_diff < NUM_JAC_DIFF_SMALL * scale] *= NUM_JAC_FACTOR_INCREASE
gharchive/issue
2021-01-29T20:55:50
2025-04-01T06:46:15.075935
{ "authors": [ "laurent90git", "wigging" ], "repo": "wigging/bfb-gasifier", "url": "https://github.com/wigging/bfb-gasifier/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1060784413
Feature request: Save to JSON file Maybe it's good to have a standard data structure like RIPE Atlas. https://atlas.ripe.net/docs/data_struct/#v5000_traceroute e.g. [{ "af": 4, "dst_addr": "202.12.27.33", "dst_name": "202.12.27.33", "endtime": 1637626292, "from": "151.246.166.238", "fw": 5020, "lts": 9, "msm_id": 5006, "msm_name": "Traceroute", "mver": "2.2.1", "paris_id": 10, "prb_id": 1000580, "proto": "UDP", "result": [{ "hop": 1, "result": [{ "from": "172.17.0.1", "rtt": 0.56, "size": 68, "ttl": 64 }, { "from": "172.17.0.1", "rtt": 0.399, "size": 68, "ttl": 64 }, { "from": "172.17.0.1", "rtt": 0.391, "size": 68, "ttl": 64 } ] }, { "hop": 2, "result": [{ "from": "192.168.10.1", "rtt": 2.873, "size": 68, "ttl": 63 }, { "from": "192.168.10.1", "rtt": 2.793, "size": 68, "ttl": 63 }, { "from": "192.168.10.1", "rtt": 2.778, "size": 68, "ttl": 63 } ] }, { "hop": 3, "result": [{ "from": "192.168.1.1", "ittl": 0, "rtt": 6.255, "size": 28, "ttl": 252 }, { "from": "192.168.1.1", "ittl": 0, "rtt": 3.729, "size": 28, "ttl": 252 }, { "from": "192.168.1.1", "ittl": 0, "rtt": 3.408, "size": 28, "ttl": 252 } ] }, { "hop": 4, "result": [{ "from": "94.183.140.6", "rtt": 36.494, "size": 28, "ttl": 252 }, { "from": "94.183.140.6", "rtt": 34.834, "size": 28, "ttl": 252 }, { "from": "94.183.140.6", "rtt": 33.892, "size": 28, "ttl": 252 } ] }, { "hop": 5, "result": [{ "from": "94.183.140.1", "rtt": 34.591, "size": 28, "ttl": 251 }, { "from": "94.183.140.1", "rtt": 34.263, "size": 28, "ttl": 251 }, { "from": "94.183.140.1", "rtt": 33.309, "size": 28, "ttl": 251 } ] }, { "hop": 6, "result": [{ "from": "172.22.133.37", "rtt": 37.557, "size": 68, "ttl": 250 }, { "from": "172.22.133.37", "rtt": 37.491, "size": 68, "ttl": 250 }, { "from": "172.22.133.37", "rtt": 38.068, "size": 68, "ttl": 250 } ] }, { "hop": 7, "result": [{ "from": "172.18.218.145", "rtt": 38.313, "size": 68, "ttl": 249 }, { "from": "172.18.218.137", "rtt": 38.449, "size": 68, "ttl": 249 }, { "from": "172.18.205.249", "rtt": 37.496, "size": 68, "ttl": 249 } ] }, { "hop": 8, "result": [{ "from": "172.18.188.18", "rtt": 38.758, "size": 68, "ttl": 121 }, { "from": "172.18.188.18", "rtt": 38.614, "size": 68, "ttl": 121 }, { "from": "172.18.188.18", "rtt": 38.552, "size": 68, "ttl": 121 } ] }, { "hop": 9, "result": [{ "from": "172.18.188.68", "rtt": 39.842, "size": 68, "ttl": 247 }, { "from": "172.18.188.68", "rtt": 39.143, "size": 68, "ttl": 247 }, { "from": "172.18.188.68", "rtt": 44.919, "size": 68, "ttl": 247 } ] }, { "hop": 10, "result": [{ "from": "172.18.192.17", "rtt": 40.768, "size": 68, "ttl": 246 }, { "from": "172.18.192.25", "rtt": 39.261, "size": 68, "ttl": 246 }, { "from": "172.18.192.13", "rtt": 40.284, "size": 68, "ttl": 246 } ] }, { "hop": 11, "result": [{ "from": "172.18.181.137", "rtt": 40.962, "size": 68, "ttl": 245 }, { "from": "172.16.46.1", "rtt": 41.896, "size": 68, "ttl": 245 }, { "from": "172.16.46.5", "rtt": 47.197, "size": 68, "ttl": 245 } ] }, { "hop": 12, "result": [{ "from": "10.201.177.221", "rtt": 43.408, "size": 68, "ttl": 242 }, { "from": "10.201.177.225", "rtt": 43.689, "size": 68, "ttl": 242 }, { "from": "10.201.177.225", "rtt": 42.649, "size": 68, "ttl": 242 } ] }, { "hop": 13, "result": [{ "from": "10.21.21.10", "rtt": 42.748, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 41.301, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 62.887, "size": 68, "ttl": 241 } ] }, { "hop": 14, "result": [{ "from": "10.21.211.10", "icmpext": { "obj": [{ "class": 0, "type": 0 } ], "rfc4884": 1, "version": 0 }, "ittl": 2, "rtt": 43.338, "size": 140, "ttl": 239 }, { "from": "10.21.211.10", "icmpext": { "obj": [{ "class": 0, "type": 0 } ], "rfc4884": 1, "version": 0 }, "ittl": 2, "rtt": 45.076, "size": 140, "ttl": 239 }, { "from": "10.21.211.10", "icmpext": { "obj": [{ "class": 0, "type": 0 } ], "rfc4884": 1, "version": 0 }, "ittl": 2, "rtt": 41.775, "size": 140, "ttl": 239 } ] }, { "hop": 15, "result": [{ "from": "10.21.21.10", "rtt": 42.107, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 46.849, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 42.075, "size": 68, "ttl": 241 } ] }, { "hop": 16, "result": [{ "from": "185.1.48.16", "rtt": 76.749, "size": 68, "ttl": 49 }, { "from": "185.1.48.16", "rtt": 74.96, "size": 68, "ttl": 49 }, { "from": "185.1.48.16", "rtt": 74.57, "size": 68, "ttl": 49 } ] }, { "hop": 17, "result": [{ "from": "184.104.196.45", "rtt": 84.812, "size": 68, "ttl": 48 }, { "from": "184.104.196.45", "rtt": 82.897, "size": 68, "ttl": 48 }, { "from": "184.104.196.45", "rtt": 83.57, "size": 68, "ttl": 48 } ] }, { "hop": 18, "result": [{ "from": "184.104.196.1", "rtt": 82.066, "size": 28, "ttl": 46 }, { "from": "184.104.196.1", "rtt": 82.954, "size": 28, "ttl": 46 }, { "from": "184.104.196.1", "rtt": 82.795, "size": 28, "ttl": 46 } ] }, { "hop": 19, "result": [{ "from": "184.105.65.133", "rtt": 96.805, "size": 28, "ttl": 46 }, { "from": "184.105.65.133", "rtt": 96.212, "size": 28, "ttl": 46 }, { "from": "184.105.65.133", "rtt": 102.439, "size": 28, "ttl": 46 } ] }, { "hop": 20, "result": [{ "from": "184.105.65.5", "rtt": 115.32, "size": 28, "ttl": 45 }, { "from": "184.105.65.5", "rtt": 111.125, "size": 28, "ttl": 45 }, { "from": "184.105.65.5", "rtt": 113.071, "size": 28, "ttl": 45 } ] }, { "hop": 21, "result": [{ "from": "37.49.237.106", "rtt": 129.242, "size": 28, "ttl": 235 }, { "from": "195.42.144.79", "rtt": 130.602, "size": 28, "ttl": 235 }, { "from": "37.49.237.106", "rtt": 126.654, "size": 28, "ttl": 235 } ] }, { "hop": 22, "result": [{ "from": "202.12.27.33", "rtt": 130.383, "size": 28, "ttl": 233 }, { "from": "202.12.27.33", "rtt": 126.517, "size": 28, "ttl": 233 }, { "from": "202.12.27.33", "rtt": 127.824, "size": 28, "ttl": 233 } ] } ], "size": 40, "src_addr": "172.17.0.2", "stored_timestamp": 1637626336, "timestamp": 1637626288, "type": "traceroute" }, { "af": 4, "dst_addr": "202.12.27.33", "dst_name": "202.12.27.33", "endtime": 1637628093, "from": "151.246.166.238", "fw": 5020, "lts": 14, "msm_id": 5006, "msm_name": "Traceroute", "mver": "2.2.1", "paris_id": 11, "prb_id": 1000580, "proto": "UDP", "result": [{ "hop": 1, "result": [{ "from": "172.17.0.1", "rtt": 0.564, "size": 68, "ttl": 64 }, { "from": "172.17.0.1", "rtt": 0.399, "size": 68, "ttl": 64 }, { "from": "172.17.0.1", "rtt": 0.315, "size": 68, "ttl": 64 } ] }, { "hop": 2, "result": [{ "from": "192.168.10.1", "rtt": 2.748, "size": 68, "ttl": 63 }, { "from": "192.168.10.1", "rtt": 2.916, "size": 68, "ttl": 63 }, { "from": "192.168.10.1", "rtt": 4.468, "size": 68, "ttl": 63 } ] }, { "hop": 3, "result": [{ "from": "192.168.1.1", "ittl": 0, "rtt": 3.946, "size": 28, "ttl": 252 }, { "from": "192.168.1.1", "ittl": 0, "rtt": 3.641, "size": 28, "ttl": 252 }, { "from": "192.168.1.1", "ittl": 0, "rtt": 3.745, "size": 28, "ttl": 252 } ] }, { "hop": 4, "result": [{ "from": "94.183.140.6", "rtt": 34.607, "size": 28, "ttl": 252 }, { "from": "94.183.140.6", "rtt": 34.368, "size": 28, "ttl": 252 }, { "from": "94.183.140.6", "rtt": 33.883, "size": 28, "ttl": 252 } ] }, { "hop": 5, "result": [{ "from": "94.183.140.1", "rtt": 35.126, "size": 28, "ttl": 251 }, { "from": "94.183.140.1", "rtt": 34.16, "size": 28, "ttl": 251 }, { "from": "94.183.140.1", "rtt": 34.08, "size": 28, "ttl": 251 } ] }, { "hop": 6, "result": [{ "from": "172.22.133.37", "rtt": 40.524, "size": 68, "ttl": 250 }, { "from": "172.22.133.37", "rtt": 37.733, "size": 68, "ttl": 250 }, { "from": "172.22.133.37", "rtt": 37.492, "size": 68, "ttl": 250 } ] }, { "hop": 7, "result": [{ "from": "172.18.218.137", "rtt": 38.053, "size": 68, "ttl": 249 }, { "from": "172.18.205.213", "rtt": 38.36, "size": 68, "ttl": 249 }, { "from": "172.18.222.209", "rtt": 37.867, "size": 68, "ttl": 249 } ] }, { "hop": 8, "result": [{ "from": "172.18.188.18", "rtt": 37.903, "size": 68, "ttl": 121 }, { "from": "172.18.188.18", "rtt": 39.561, "size": 68, "ttl": 121 }, { "from": "172.18.188.18", "rtt": 37.919, "size": 68, "ttl": 121 } ] }, { "hop": 9, "result": [{ "from": "172.18.188.68", "rtt": 41.651, "size": 68, "ttl": 247 }, { "from": "172.18.188.68", "rtt": 39.315, "size": 68, "ttl": 247 }, { "from": "172.18.188.68", "rtt": 40.564, "size": 68, "ttl": 247 } ] }, { "hop": 10, "result": [{ "from": "172.18.192.1", "rtt": 38.432, "size": 68, "ttl": 246 }, { "from": "172.18.192.29", "rtt": 39.749, "size": 68, "ttl": 246 }, { "from": "172.18.192.5", "rtt": 53.008, "size": 68, "ttl": 246 } ] }, { "hop": 11, "result": [{ "from": "172.16.46.9", "rtt": 41.886, "size": 68, "ttl": 245 }, { "from": "172.16.46.193", "rtt": 41.042, "size": 68, "ttl": 245 }, { "from": "172.16.46.197", "rtt": 42.388, "size": 68, "ttl": 245 } ] }, { "hop": 12, "result": [{ "from": "10.201.177.225", "rtt": 41.305, "size": 68, "ttl": 242 }, { "from": "10.201.177.221", "rtt": 44.863, "size": 68, "ttl": 242 }, { "from": "10.201.177.225", "rtt": 47.7, "size": 68, "ttl": 242 } ] }, { "hop": 13, "result": [{ "from": "10.21.21.10", "rtt": 42.65, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 42.842, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 42.72, "size": 68, "ttl": 241 } ] }, { "hop": 14, "result": [{ "from": "10.21.211.10", "icmpext": { "obj": [{ "class": 0, "type": 0 } ], "rfc4884": 1, "version": 0 }, "ittl": 2, "rtt": 43.051, "size": 140, "ttl": 239 }, { "from": "10.21.211.10", "icmpext": { "obj": [{ "class": 0, "type": 0 } ], "rfc4884": 1, "version": 0 }, "ittl": 2, "rtt": 43.543, "size": 140, "ttl": 239 }, { "from": "10.21.211.10", "icmpext": { "obj": [{ "class": 0, "type": 0 } ], "rfc4884": 1, "version": 0 }, "ittl": 2, "rtt": 43.603, "size": 140, "ttl": 239 } ] }, { "hop": 15, "result": [{ "from": "10.21.21.10", "rtt": 42.254, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 44.902, "size": 68, "ttl": 241 }, { "from": "10.21.21.10", "rtt": 41.691, "size": 68, "ttl": 241 } ] }, { "hop": 16, "result": [{ "from": "185.1.48.16", "rtt": 77.511, "size": 68, "ttl": 49 }, { "from": "185.1.48.16", "rtt": 74.563, "size": 68, "ttl": 49 }, { "from": "185.1.48.16", "rtt": 76.154, "size": 68, "ttl": 49 } ] }, { "hop": 17, "result": [{ "from": "184.104.196.45", "rtt": 84.355, "size": 68, "ttl": 48 }, { "from": "184.104.196.45", "rtt": 85.82, "size": 68, "ttl": 48 }, { "from": "184.104.196.45", "rtt": 82.819, "size": 68, "ttl": 48 } ] }, { "hop": 18, "result": [{ "from": "184.104.196.1", "rtt": 84.804, "size": 28, "ttl": 46 }, { "from": "184.104.196.1", "rtt": 83.745, "size": 28, "ttl": 46 }, { "from": "184.104.196.1", "rtt": 81.614, "size": 28, "ttl": 46 } ] }, { "hop": 19, "result": [{ "from": "184.105.65.133", "rtt": 96.806, "size": 28, "ttl": 46 }, { "from": "184.105.65.133", "rtt": 95.489, "size": 28, "ttl": 46 }, { "from": "184.105.65.133", "rtt": 96.774, "size": 28, "ttl": 46 } ] }, { "hop": 20, "result": [{ "from": "184.105.65.5", "rtt": 119.542, "size": 28, "ttl": 45 }, { "from": "184.105.65.5", "rtt": 111.951, "size": 28, "ttl": 45 }, { "from": "184.105.65.5", "rtt": 113.013, "size": 28, "ttl": 45 } ] }, { "hop": 21, "result": [{ "from": "195.42.144.79", "rtt": 138.066, "size": 28, "ttl": 235 }, { "from": "37.49.237.106", "rtt": 128.829, "size": 28, "ttl": 235 }, { "from": "37.49.237.106", "rtt": 132.527, "size": 28, "ttl": 235 } ] }, { "hop": 22, "result": [{ "from": "202.12.27.33", "rtt": 129.301, "size": 28, "ttl": 233 }, { "from": "202.12.27.33", "rtt": 124.837, "size": 28, "ttl": 233 }, { "from": "202.12.27.33", "rtt": 125.361, "size": 28, "ttl": 233 } ] } ], "size": 40, "src_addr": "172.17.0.2", "stored_timestamp": 1637628137, "timestamp": 1637628089, "type": "traceroute" }] closed by #4
gharchive/issue
2021-11-23T02:51:44
2025-04-01T06:46:15.099562
{ "authors": [ "xhdix" ], "repo": "wikicensorship/tracevis", "url": "https://github.com/wikicensorship/tracevis/issues/3", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
78101416
Any plans for a Android 5.x release? :grin: Google provides an official Sketch template for Material Design, or do you mean support for more display densities? I did start on a 5.x template in the same vein but abandoned it after Google came out with their own. May revive it if you can convince me it's worth pursuing... :wink: I didn't see that official template from Google. I guess it's better to stick with their template then. Thanks for this anyway!
gharchive/issue
2015-05-19T14:03:22
2025-04-01T06:46:15.103716
{ "authors": [ "gabrielizaias", "wikichen" ], "repo": "wikichen/sketch-android-kit", "url": "https://github.com/wikichen/sketch-android-kit/issues/4", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1308365497
Clicking Party Mode in Table of Contents doesn't bold it Clicking Party mode still highlights Marathon mode. Suggested improvement: add some margin to the bottom so screen can scroll to have Party mode more centered/higher on the screen like when clicking the rest of the options https://user-images.githubusercontent.com/17424008/179579663-f2528887-5563-4d6d-b30f-c50be711bbb1.mov I think this has to do with screen height? For my screen its Marathon mode that doesn't get bolded instead. Not sure what we can do about this
gharchive/issue
2022-07-18T18:37:39
2025-04-01T06:46:15.110886
{ "authors": [ "bricehalder", "mliu59" ], "repo": "wikispeedruns/wikipedia-speedruns", "url": "https://github.com/wikispeedruns/wikipedia-speedruns/issues/408", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
118964457
Fix IE10 error How to reproduce Stickyfill.kill(); $('.sticky').Stickyfill(); The error SCRIPT5007: Unable to get property 'removeChild' of undefined or null reference Can't reproduce it on this page. The check for existance of the parent node could be useful in case the DOM is changed by other scripts, but this doesn't seem to be IE specific bug. Could you show me the page where I can reproduce this bug? I also see this error but only via IE & Edge. Maybe the bug is MS specific but the fix is not, so why do we not merge the fix?
gharchive/pull-request
2015-11-26T03:26:02
2025-04-01T06:46:15.113637
{ "authors": [ "incon", "voku", "wilddeer" ], "repo": "wilddeer/stickyfill", "url": "https://github.com/wilddeer/stickyfill/pull/29", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1470836708
Completion command to generate bash/zsh completion script Resolves https://github.com/wildfly-extras/prospero/issues/205 CC @spyrkob Couple of points to this: The completion of predefined FP names works :+1: . Probably we need to update the main command name, currently it's set to prospero (i.e. this is the command name that trigger the completion in bash). What is the final script name that prospero will use? I remember something like eap-installation-manager.sh? Are going to update this later on, or shall we do it now? The completion in bash only triggers for predefined script names. E.g. with current settings, prospero <TAB><TAB> provides completion, but ./prospero <TAB><TAB> does not. This may be somewhat problematic, because I assume prospero would typically be called by relative path, not as "global" command. The relative path like bin/prospero can be configured too, but it cannot contain ./, as such substrings are sanitized away. If we wanted to support ./prospero or ./bin/prospero to trigger completion, we will probably need to generate the completion script a little differently, as it's not supported out of the box. Examples of help messages: $ ./prospero -h Usage: prospero [-hv] [COMMAND] Options: -h, --help Display this help message. -v, --version Print prospero version and exit. Commands: install Install a new application server instance. update Apply the latest available patches on a server instance. history List previous installation states. revert Reverts to a previous installation state. channel Manage channels the installation is subscribed to. completion Generate bash completion script. (Apply with `source <(prospero completion)`) clone Manage the installation metadata. Exit codes: 0 Successful program execution. 1 Failed operation. 2 Invalid arguments were given. Use `prospero <COMMAND> --help` to show usage information about given command. $ ./prospero completion -h Generate bash completion script. (Apply with `source <(prospero completion)`) Usage: prospero completion [-hV] Generate bash/zsh completion script for prospero. Run the following command to give `prospero` TAB completion in the current shell: source <(prospero completion) Alternatively, you can save the completion script to a file, and evaluate this file from your .bashrc script: prospero completion > ~/prospero_completion echo ". ~/prospero_completion" >> ~/.bashrc Options: -h, --help Display this help message. -V, --version Print prospero version and exit. In fact I don't know if the Commands.MAIN_COMMAND is really used for anything that is visible. It could be in theory removed (maybe). Rebased and updated to respect the prospero.dist.name property. I think it's now ready.
gharchive/pull-request
2022-12-01T07:45:28
2025-04-01T06:46:15.123243
{ "authors": [ "TomasHofman" ], "repo": "wildfly-extras/prospero", "url": "https://github.com/wildfly-extras/prospero/pull/207", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
753424516
DS_Store file in the 1.0.0 zip prevents other feature packs from depending on wildfly-mp-reactive-feature-pack The released 1.0.0 artifact (download from https://repository.jboss.org/nexus/content/groups/public/org/wildfly/extras/reactive/wildfly-microprofile-reactive-feature-pack/1.0.0.Final/wildfly-microprofile-reactive-feature-pack-1.0.0.Final.zip) contains a superfluous file packages/.DS_Store, and when my feature pack declares a dependency on the reactive feature pack 1.0.0, I get this when I try to build my feature pack: [ERROR] Failed to execute goal org.wildfly.galleon-plugins:wildfly-galleon-maven-plugin:4.2.5.Final:build-feature-pack (wildfly-mp-graphql-feature-pack-build) on project wildfly-microprofile-graphql-feature-pack: Failed to process dependencies: /packages/.DS_Store is not a directory -> [Help 1] org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.wildfly.galleon-plugins:wildfly-galleon-maven-plugin:4.2.5.Final:build-feature-pack (wildfly-mp-graphql-feature-pack-build) on project wildfly-microprofile-graphql-feature-pack: Failed to process dependencies at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:156) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:148) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:957) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:289) at org.apache.maven.cli.MavenCli.main (MavenCli.java:193) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:566) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:282) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:225) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:406) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:347) Caused by: org.apache.maven.plugin.MojoExecutionException: Failed to process dependencies at org.wildfly.galleon.maven.WfFeaturePackBuildMojo.doExecute (WfFeaturePackBuildMojo.java:197) at org.wildfly.galleon.maven.AbstractFeaturePackBuildMojo.execute (AbstractFeaturePackBuildMojo.java:178) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:210) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:156) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:148) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:957) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:289) at org.apache.maven.cli.MavenCli.main (MavenCli.java:193) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:566) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:282) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:225) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:406) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:347) Caused by: org.jboss.galleon.ProvisioningDescriptionException: /packages/.DS_Store is not a directory at org.jboss.galleon.layout.FeaturePackDescriber.assertDirectory (FeaturePackDescriber.java:130) at org.jboss.galleon.layout.FeaturePackDescriber.processPackage (FeaturePackDescriber.java:114) at org.jboss.galleon.layout.FeaturePackDescriber.processPackages (FeaturePackDescriber.java:106) at org.jboss.galleon.layout.FeaturePackDescriber.describeFeaturePack (FeaturePackDescriber.java:97) at org.jboss.galleon.layout.FeaturePackDescriber.describeFeaturePackZip (FeaturePackDescriber.java:72) at org.wildfly.galleon.maven.AbstractFeaturePackBuildMojo.processFeaturePackDependencies (AbstractFeaturePackBuildMojo.java:461) at org.wildfly.galleon.maven.WfFeaturePackBuildMojo.doExecute (WfFeaturePackBuildMojo.java:193) at org.wildfly.galleon.maven.AbstractFeaturePackBuildMojo.execute (AbstractFeaturePackBuildMojo.java:178) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:210) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:156) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:148) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:957) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:289) at org.apache.maven.cli.MavenCli.main (MavenCli.java:193) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:566) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:282) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:225) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:406) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:347) This should be fixed in https://github.com/wildfly-extras/wildfly-mp-reactive-feature-pack/releases/tag/1.0.1.Final Works, thanks!!!
gharchive/issue
2020-11-30T12:09:27
2025-04-01T06:46:15.130553
{ "authors": [ "jmartisk", "kabir" ], "repo": "wildfly-extras/wildfly-mp-reactive-feature-pack", "url": "https://github.com/wildfly-extras/wildfly-mp-reactive-feature-pack/issues/33", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
73612818
Add Connection Timeout Parameter Added timeout parameter to set a custom timeout value, as discussed at #38 issue. @miere @jamezp Any chance this can be rebased so it can get merged??? This change would be useful for me right now. Cheers. I've been meaning to process the pull queue and fix this issue, but just haven't had a chance with other deadlines ATM. Maybe I'll get a chance sometime this week or weekend and cut another Alpha release. When could we have release with this change? Really need it. Thanks. I'm hoping to get to it this week or next week. I'm replacing this with https://github.com/wildfly/wildfly-maven-plugin/pull/60 due to the merge conflict.
gharchive/pull-request
2015-05-06T13:03:24
2025-04-01T06:46:15.147635
{ "authors": [ "jamesnetherton", "jamezp", "joeydaowang", "miere" ], "repo": "wildfly/wildfly-maven-plugin", "url": "https://github.com/wildfly/wildfly-maven-plugin/pull/39", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2137327540
WFLY-19040 Configure SNI hostnames per cluster if possible As of Infinispan 14.0.18, HotRod client enables hostname validation by default - so we need to configure SNI hostnames per cluster if possible. Auto-disable hostname validation if unsupported by server configuration. https://issues.redhat.com/browse/WFLY-19040 @jmesnil FYI. upstream PR: https://github.com/wildfly/wildfly/pull/17643 @pferraro thanks!
gharchive/pull-request
2024-02-15T19:53:16
2025-04-01T06:46:15.150060
{ "authors": [ "jmesnil", "pferraro" ], "repo": "wildfly/wildfly", "url": "https://github.com/wildfly/wildfly/pull/17644", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
332356295
Invalid response I'm getting the following error when I run letsencrypt_webfaction run Invalid response from http://www.domain.com/.well-known/acme-challenge/QykNz4N2F5x7V3WxzxsNESfwHEZbumn-sGA_sk52kMI: "<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01//EN" "http://www.w3.org/TR/html4/strict.dtd">` `<html>` `<head>` `<meta http-equi"` `Make sure that you can access http://www.domain/.well-known/acme-challenge/QykNz4N2F5x7V3WxzxsNESfwHEZbumn-sGA_sk52kMI I can find the file via FTP and I can go directly to that URL in my browser. When you go directly to that URL, does it return the acme challenge file? I've run into issues occasionally where IPv6 isn't set up correctly and IPv4 returns the correct file. I get this: QykNz4N2F5x7V3WxzxsNESfwHEZbumn-sGA_sk52kMI.GFXSGlTs2NzaPYiB-J09Dx2LuaoXuzoLvPIwxOOAE8k @jenxi You did edit the domains = array in the letsencrypt_webfaction.toml right? @ericdorsey Yes, I changed the error for posting publicly. I've seen this before. Occasionally, WF will get a server in an inconsistent state where the server will respond correctly to an ipv4 originating request, but will return a 404 to an ipv6 originating request. Most of us issue requests from ipv4, but LE uses ipv6.
gharchive/issue
2018-06-14T11:10:04
2025-04-01T06:46:15.159998
{ "authors": [ "ericdorsey", "jenxi", "will-in-wi" ], "repo": "will-in-wi/letsencrypt-webfaction", "url": "https://github.com/will-in-wi/letsencrypt-webfaction/issues/130", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
987944821
Method calls on non-variables Hey @spencerwi I've been trying to get like 3.to_s or [1,2,3].first to work, and I think what needs to change is property: $ => seq( - field('object', $._variable), + field('object', $._expression), '.', field('name', alias(/[a-z][a-z0-9_]*[\?!]?/, $.identifier)), ), since I think you can call a method on the result of any expression (though this might be wrong). This change causes precedent problems with $assignment , so I'm not sure if I'm on the right track. Do you have ideas on what should be done? Hm, maybe something like this? @@ -383,8 +383,8 @@ module.exports = grammar({ $.index_expression, $.property ); - return prec.right(seq( - commaSep1(field('lhs', target)), + return prec.right(2, seq( + prec.right(1, commaSep1(field('lhs', target))), optional(field('type', $._typeAnnotation)), /=/, field('rhs', choice($._variable, $._expression)) Yeah, function calls are definitely the part that's hardest to work through. I've taken a couple of attempts at trying to support "bare" function calls (like puts "test hello" or getter :my_attr), with no real success so far - precedence conflicts happen like crazy. I've been trying to pin down places where whitespace shouldn't be allowed to try to avoid some of the conflicts I see when I poke at it. I wonder if it's worth trying to adapt the tree-sitter grammar for ruby here...could be worth a shot.
gharchive/issue
2021-09-03T16:44:21
2025-04-01T06:46:15.167044
{ "authors": [ "spencerwi", "will" ], "repo": "will/tree-sitter-crystal", "url": "https://github.com/will/tree-sitter-crystal/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
328918284
Can this module do LBT ? Listen before talk ? It seems that HAL from SEMTECH supports LBT when version is > = 4.1.0, but can AS923 be supported by your board ? (Japan) No, it won't support LBT since it needs a FPGA + SX1278 to scan the channel for LBT. Note that if anyone have the ref V1.5 design sch, I'm willing to add LBT support. 😄 Ok thanks and no I do not have the reference design v1.5 Reopen this issue because it is on my todo list... Schematic can be reversed from this photos: https://fccid.io/AU792U13A16859/Internal-Photos/Internal-Photos-3321124 Almost, but you can't, because it is a multi layer board. I'm hoping to reverse by reversing the FPGA bitstream, but I can't read verilog...... Yep, You are right, but on this one: http://www.embit.eu/wp-content/uploads/fronte-retro1.png The signal layers are top/bottom and inner - are powers. So this is good suggestion to start. Also There is correct FPGA (iCE40LP1K-CM49) to which firmware is avaliable. Very to see to that you think it is a worthy thing to do. Pin mapping of FPGA, correct me if I'm wrong: #SPI interface to HOST set_io SPI_MISO A5 set_io SPI_CS A6 set_io SPI_SCLK A3 #Wire shared with SX1301_SCLK and EEPROM_SCLK set_io SPI_MOSI A4 #Wire shared with SX1301_MOSI and EEPROM_MOSI #SPI slave 0: set_io SX1301_GPIO1 D7 set_io SX1301_GPIO3 A2 set_io SX1301_GPIO4 B2 set_io SX1301_CSN B3 set_io SX1301_MISO B4 set_io SX1301_RADIO_RST F4 #this is input, not used in FPGA fw #SPI slave 1: FPGA itselve #SPI slave 2: set_io EEPROM_CSN A7 set_io EEPROM_MISO C4 #SPI slave 3: set_io SX1272_MOSI A1 set_io SX1272_NSS E2 set_io SX1272_MISO D2 set_io SX1272_SCK D3 set_io SX1272_RESET F2 #Stays in the middle between I2S interface of SX1301 and SX125X (radio A) set_io A_SP_CLK_IN D1 set_io A_QI_TX_IN B1 set_io A_IQ_TX_IN C1 set_io A_SP_CLK_OUT F3 set_io A_IQ_TX_OUT G3 set_io A_QI_TX_OUT G4 There is some small microchip SPI eeprom, but it's not used in software. FPGA can be replaced by more DIY-friendly with QFN48 package, like ICE5LP1K-SG48. Only pin map file must be changed (and recompile firmware with icestorm). Also SX1272 share reference clock with SX125X and receiving RF just after second SAW (have no idea about how impedance matching has been made there). Maybe I can try to it with my lattice ICE40UP5K dev board later... But first need to solder (finalize) my gateway. @piratfm That was awesome! I'm recently doing other projects that uses ICE5LP1K-SG48 (a.k.a learning verilog), so I also have some chip and boards :D I'm glad that I left a bunch of resistor between the SPI lines, it makes it easier to modify. Hi, I have customers that want LoRa gateways. I wonder if you are available to make some V3 boards to sell me on tindie and coach me for pay on setup of all the software to connect to lorawan thethingsnetwork? Later, we can make improved boards, weather enclosures, etc. Interested? Are you trained with a BSEE degree, or hobby? Thanks, John PostScript: I knew some verilog for chip design, but have forgotten most of it. Listen before talk would be a fine feature to brag about! Hope Will would (re)consider to build another batch of boards. After surveying what is low cost and available right now, Lora, in low cost form, is still not so accessible, not much choice and this looks to be the best solution so far. RAK(831) boards are not cheap either. Trying to build an IoT gateway (not using RPI but with STM32/NXP based embedded with SPI) and would like to add in Lora. What would be nice are small companion device development ("thing") boards to connect sensors and actuators for prototype solution purposes. I am in Taipei City as well. @symdeb How is your project to build IoT gateway with STM32/NXP parts coming? Do you use pcb-rnd to make boards? Maybe we can collaborate to get a low cost gateway and sensor boards? I've done work with TinyOS on sensor boards, but that OS stalled. RIOT is the OS I am interested in now.
gharchive/issue
2018-06-04T05:59:42
2025-04-01T06:46:15.175846
{ "authors": [ "4mig4", "jgriessen", "piratfm", "symdeb", "will127534" ], "repo": "will127534/LoRa-concentrator", "url": "https://github.com/will127534/LoRa-concentrator/issues/16", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
933283561
Double click on 8th, then the event is created on 7th great calendar, but this issue is present : Double click on 8th, then the event is created on 7th did try with other dates and the same result Hi, can you tell me what browser you were using? Firefox 89.0.2 From: William Troup @.> Sent: Wednesday, June 30, 2021 2:33 PM To: williamtroup/Calendar.js @.> Cc: Omar R. Tapia @.>; Author @.> Subject: Re: [williamtroup/Calendar.js] Double click on 8th, then the event is created on 7th (#35) Hi, can you tell me what browser you were using? — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHubhttps://github.com/williamtroup/Calendar.js/issues/35#issuecomment-871459246, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AAIX4MS5XCN5RIITIKKSYPLTVMTNJANCNFSM47RPNY3Q. What version of the library are you using? Calendar.js v0.9.7 From: William Troup @.> Sent: Thursday, July 1, 2021 2:29 PM To: williamtroup/Calendar.js @.> Cc: Omar R. Tapia @.>; Author @.> Subject: Re: [williamtroup/Calendar.js] Double click on 8th, then the event is created on 7th (#35) What version of the library are you using? — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHubhttps://github.com/williamtroup/Calendar.js/issues/35#issuecomment-872295622, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AAIX4MQLAMTHBRUNB3VVSLTTVR3VPANCNFSM47RPNY3Q. This will be fixed in v0.9.8, which is coming hopefully in the next week. Thank you for reporting the problem. v0.9.8 is now available. Please let me know if this issue persists. Thanks again. I downloaded the last version today and tried again, still the same problem . I used Firefox and Chrome, even in Chrome I had the same result.
gharchive/issue
2021-06-30T04:03:49
2025-04-01T06:46:15.192588
{ "authors": [ "omarruben", "williamtroup" ], "repo": "williamtroup/Calendar.js", "url": "https://github.com/williamtroup/Calendar.js/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1146929720
Docummenting how to use .get filters It would be really useful to know what's accepted as filters, and what do you pass as results to the function. I tried sending the LibgenSearch event itself, and also LibgenSearch.get_results OrderedDict (and converted to Dict) and it no match was found. Hi @Lamarcke, I really appreciate the comment. I'm currently working on fixing the docs as well as some other patch work. If you could provide the parameters you used when creating the LibgenSearch instance or the steps you did when calling LibgenSearch.get_results that yielded no results that would help me resolve any existing bugs. Thank you! I'm sorry, for some reason i was using LibgenSearch.get_results as an filter function, when in reality you need to pass the parameters to the instance itself. I'm learning quite a lot by studying your code, so i must thank you for your hard work.
gharchive/issue
2022-02-22T13:43:48
2025-04-01T06:46:15.207351
{ "authors": [ "Lamarcke", "willmeyers" ], "repo": "willmeyers/grab-convert-from-libgen", "url": "https://github.com/willmeyers/grab-convert-from-libgen/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2420894567
feat: disable ui in blocked guests Due to the nature of how some *line plugins may work (for me, it's 'heirline.nvim') blocked windows may look like they are interactive which may be confusing. This PR introduces a function that disables UI elements so that blocked windows look like background. Thank you for the PR! I appreciate the contribution but I think this is something that should be left to user configs. You should be able to replicate this behavior with the hooks you can add to the plug-in's configuration.
gharchive/pull-request
2024-07-20T13:13:58
2025-04-01T06:46:15.208682
{ "authors": [ "lostl1ght", "willothy" ], "repo": "willothy/flatten.nvim", "url": "https://github.com/willothy/flatten.nvim/pull/103", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
209540587
[REFACTOR] Build Game Status Query Use a pretty set and not strings to generate status query join would be a good idea for this @emilyseibert is there any way for me to test this? nope, just as long as the games come back for you
gharchive/issue
2017-02-22T18:30:15
2025-04-01T06:46:15.210258
{ "authors": [ "emilyseibert", "morganedmonds" ], "repo": "willowtreeapps/wombats-web-client", "url": "https://github.com/willowtreeapps/wombats-web-client/issues/146", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
488221626
Decide on who's the target audience of the project I post this as a first issue to decide in regard to https://github.com/windelbouwman/ppci-mirror/issues/10 . What is an "improvement" (and what's not) depends largely on who's target audience. For example, it's possible to decide that end users are the target audience. But then the README should not contain "frightening" stuff like: Warning This project is in alpha state and not ready for production use! And it stead should contain stuff along the lines of "Drop GCC, drop LLVM, start use PPCI now, we have cookies!" Or it can be decided that WebAssembly can be a selling point piggybacking on which PPCI could launch into masses. And then README should provide instructions how to build some cute display hack and open it in a popular browser. The examples can continue, let me come up with a specific proposal: target audience should be Python developers. "Developers" definitely, as README itself says that the project is in alpha stage. "Python", because I doubt that many non-Python developers would jump to see another compiler out there. The project should be of most interest to folks who know Python and curious what can be done using their favorite language. However, any instruction in the README should also be friendly to folks who aren't (much) familiar with Python. This is not contradictory goal to the previous paragraph. For example, I'm good enough familiar with Python, but I still don't like to "install" something right away (indeed, that's not related to Python and applies to any software). Fortunately, it seems that PPCI is usable from just a git clone. But instruction should e.g. refer to python3 -m ppci cc instead of ppci-cc, because the latter appears only after "installing" it. Criticism/other ideas are welcome. I anyway wanted to post this, to give a context to other sub-tickets I may post for https://github.com/windelbouwman/ppci-mirror/issues/10. (Again, based on own experience - I get various suggestions for improvements in my projects READMEs, and half of the time I wonder why they think it would be an improvement.) Currently I think the target audience should be somewhat embedded software engineers how are playing with python, and feel brave to try this stuff, or want to script / automate / parse some C code as part of their build process. Another group might be compiler developers who would like to try out a new idea quickly. In fact, I'm unsure who would really use this project for something serious, it's mostly out of curiosity of what can be done with python. Any other thoughts on this? Thanks for the reply here. IMHO, this is one of the most important questions right now, as it sets the further direction of the project, and would make it clear to potential contributors what would construe a useful contribution to the project and what would not. Let me start with commenting to your replies. Currently I think the target audience should be somewhat embedded software engineers That's peculiarly formulated ;-). I guess the main word there is "somewhat embedded". Because one should picture a typical embedded software engineer (as in "engineer of software for deeply embedded systems, i.e. microcontrollers") as a guy sitting on Windows 7 (because Windows 10, Linux, Mac don't support drivers for his hardware probe), and holding his IAR C. In no way this guy will be interested in Python, PPCI, etc. Ok, maybe for "somewhat embedded engineers" it would be different, but there would be very few of such "somewhat embedded engineers", which makes almost zero target audience. Another group might be compiler developers who would like to try out a new idea quickly. Definitely +1, but these are largely served well by LLVM. So, would need to define which subgroup may still be interested in PPCI, and how to make it more interesting to larger number of these folks beyond that. In fact, I'm unsure who would really use this project for something serious, it's mostly out of curiosity of what can be done with python. I kinda suspected it might be like that, and thanks for spelling it out. And there're many projects like that, I dumped https://github.com/windelbouwman/ppci-mirror/issues/33 in preparation to this reply, and that lists only "hobby" compilers in Python, and for C. There're many more compilers in Python for other languages, and more compilers for something written in other languages. But PPCI has got some "problem" - it has too wide scope - multiple source languages, multiple output (byte)codes, SSA, linker, build system, etc. - all packed together. But not only wide, it's also deep: while one needs to do some fiddling, one can see all those pieces actually work (not completely and not fully, but work). And the final "terrible news" is that all this work was done largely by one man - yourself @windelbouwman. Which is again shows the power Python brings to mere people. So, what we have is compiler infra on the brink of being useful, done by mostly one man. Just imagine where that can go if more people get involved, and how it can affect other people. Some people just can't sleep well thinking about all the potential, hence such tickets ;-). In the next comment, let me elaborate a bit on that. So, let's think of who can be target audience of open-source project. While discussing that, I will be giving valuations from the point of view of "community" and from a point of view of "overall progress". Obviously, I'm just a single person, and probably can't represent a "community" (which may not even exist, as in "nobody gives a damn about all this stuff"). A notion of "progress" isn't exactly objective either. With that disclaimer out, let me still try to define: The mission statement: The state of the art in advanced compiler hacking is object-oriented API in C++ with LLVM, hopefully being able to structure that as a "plugin". That's noticeable improvement over previous situation of GCC with much less cleaner C API and explicit political prohibition of plugins. But current state of affairs is still represents "extraordinary effort" for many people who are interested in compiler hacking, but are short of time resources to learn LLVM and maintain C++ stuff. We'd like to improve that largely with compiler infrastructure implement in Python. Ok, types of projects by intended target audience. The target audience is the (usually sole) developer of the project. This is worst case from community perspective. The interest here is not to improve state of the art, but merely learn things which isn't known/familiar to the developer of project. Again, this is the worst case from perspective of community contributors, because any contributors would be ignored, or rejected, because the author isn't really interested in somebody else's ideas, only his own. And these ideas are usually in their own head and he simply don't have enough time and interest to discuss them or other ideas with other people. @windelbouwman, I hope you never wanted to make a project like that. Because otherwise, you failed miserably. Failed by overdoing it - again, your project is on the brink of being useful to a wide community of people. With the second type of project, believe it or not, the crux of project is still main author's own ideas, but target audience is explicitly the community, theatrical style. This is latest fashion, and is great new type of geeky entertainment. One common trait is that these try to be "a whole world from scratch", based on some NIH approach because, of course what we have sucks, and they can do it better. https://github.com/pervognsen/bitwise is the proverbial example of such a project. From README: "I've always been obsessed with how things work under the hood", "Here are some examples of what you will learn to build:". So, the guy is obsessed, but we will learn. Sounds good, bring me popcorn. Turns out, we start learning with inventing own language, Ion, because it ain't cool otherwise. Another example is https://github.com/akkartik/mu . Did you know that you can do structured programming in machine code, and that it's the future of programming? That guy does it. Again, such projects provide a great value for community as entertainment (and education of course). Beyond that, perspectives are mixed. What we can learn from it, that it's great to make NIH tools? Community never had problems with that. @windelbouwman, I hope you smell my fears - I suspect that current PPCI project may fall into this category :-(. Because own build system, XML based in late 2010'es, really? C3-the-your-own-language, really? Then it would be interesting to know whether being in this category was conscious or not conscious. In the first case, everything is doomed. And in the later case, I do all this writing to sway the project in the next category. Projects in this category ration missionary activity. They usually have a specific aim (which can be not explicitly pronounces/conscious too), and don't try to "fix the whole world" (only if by implication, but not rebuilding it from scratch). Quite often a case is "let's do thing like X, but with a), b), c) (finite list) changed". Many projects are like that, e.g., I'd say RustPython is "let's do the same bloat as CPython, just in Rust", MicroPython is "Python is good, but too much bloat, let's pluck it", my own Pycopy is "Hey MicroPython, stop betraying "micro", that's not minimal enough". One good thing about such projects is that their target audience is potentially users of the entire "original" project, and there's a good cross-pollination among them. We would also put LLVM into this category, but the truth is that it hardly ever was in it, and always was in the next category. Corporate open-source projects. These are oftentimes rooted in politics and corporate interests. Let's not skip GCC which is a compiler of GNU-the-corporation with agenda of precluding commercial entities to benefit from community work without returning back to community, and even violating community interests. Next example in queue is LLVM, which is a response from commercial entity to GNU concerns. Etc, etc. These projects are definitely in the interest of (loosely defined) community, as they're well organized and oftetimes funded, so work on "boring" matters like testing, thorough compatibility, horizontal scalability (e.g. many architectures), etc. The community of the projects themselves may be however "astroturfed", and it may be too hard for a "mere human" to contribute to such projects, or make them work in their own (vs corporations') interest. That's it. Don't get me wrong - there're no "bad" types of projects in that list, and no "best". For different projects, one or another model may be suitable, and that may change over time. A professional software developer definitely should try all of them. And again, p.4 is not the "toppest" one. https://github.com/pervognsen of Bitwise mentioned above is a good example, by his account, he had so much fun of corporate programming, that it took a leave on his own to recover a bit of his life and convey a message to community with project of type 2. Anyway, back to PPCI. I would humbly suggest that for PPCI, project type 3 is the best. @windelbouwman, would you agree? I think this project has moved in type 1 and type 2. At first, for me this was just a learning experience, and at some point I put some effort in cleaning the code and creating some documentation, since I realized it might become useful to other people. The way you outlined it, type 3 sounds reasonable, but I would like to keep the scope of the project wide. Not fix the whole world, but have a consistent library which can be used to deal with compilation related problems. I fully understand that the project is wide, and not a classical unix tool ("do one thing, and do it well"). It want this to be a broad project, since them all parts can work nicely together. Off course, this might result in a big ball of mud in which there is a lot of stuff which all does almost work, but I'm willing to bet on this. I've seen too many projects split up into several repo's and subprojects and then configuration management takes over. I would like to think of this project in the lines of sox and netcat, tools which are a sort of swiss army knife for a specific topic. Btw other target audience I thought of (to prevent the target audience from becoming equal to the empty set :)): Software archeoligists, people dealing with really old software for which compilers cannot be used anymore As a bootstrapping tool for either a very old language / compiler or for a new language for which a first bootstrap compiler must be developed. Btw, I'm glad this is not a type 4 project! @windelbouwman, thanks for the reply! The way you outlined it, type 3 sounds reasonable, but I would like to keep the scope of the project wide. Thanks for acking that "type 3 project" sounds good, at least I finally explicated what's the reasoning for the changes I already proposed, and hope to propose even more. And it's absolutely fine for project to be wide - indeed, compilation itself is very wide subject, so a project dealing with it won't be exactly as simple as unix "cat". But already mentioned my concern in https://github.com/windelbouwman/ppci-mirror/issues/29 - you never do "something", you always do "something instead of something else". And that's exactly my idea - to call for down-prioritizing work in some areas and prioritizing work in other areas, often last-mile ones (where simple enough changes can lead to overall vast progress, at least re: community adoptability). Not fix the whole world, but have a consistent library which can be used to deal with compilation related problems. Again, good. But then it's a matter when to set fence posts. For example, for me, build system isn't really related to "compilation problems". It's completely different area of generic dependency tracking and task scheduling. And ppci-build is the part I'm most skeptical about (as in: I'm not interested in using it), and would like to propose to downplay it, and up-play PPCI use with other build systems. Anyway, that's just a specific example. Btw other target audience Software archeoligists, As a bootstrapping tool for either a very old That's still too niche, still not what I have in mind. Let me finally formulate my ideas in the next comment. Ok, as we agreed that "type 3" project makes sense, the main matter is of course implications that has on project approach. Let me summarize what I have in mind: What's one of the most known/used languages? C? Does PPCI support C? Yes? Then let it be that any C programmer be in our potential target audience. (That's mostly "statement of intent", but even it has its implications, we should make thing familiar to C programmers, and not make things not familiar. E.g., C programmers usually don't see error messages like in #9, so we shouldn't have either). What's the mostly widely C compiler? GCC? Then let any GCC user be the actual target audience of the project. Like, and GCC user should be whole-heartedly welcome to try PPCI, and by "whole-heartedly" I mean "we should make their experience positive, and motivating to try further, rather than run away in awe". There're concrete and far-reaching implications of this clause, if following points. E.g. all tools provided by PPCI should be structured following GCC, and have compatible arguments, command-line options, etc. Clang learnt that in a painful way, that few people took them seriously until they acquired most of GCC-compatible command-line options, inline asm quirks, etc. (And they're still not there, and that limits their adoption.) On the other hand, project would rather concentrate on the scope corresponding to GCC and Binutils to avoid stretching too thin and going to no-man's land. One example is build system - it's good that PPCI provides own, but shouldn't be enforced (as in: there should be enough examples of using ppci without ppci-build, and actually such examples should be promoted, to not scare people with NIH things). In general, interoperability is important. One particular area is interoperability with tools similar is scope. E.g., I believe there're some beginnings of dealing with LLVM IR, and that's one important area which fully makes sense to be elaborated further. Similarity, using existing good libraries which adhere to project constraints (like, being pure-Python) may be a good thing, instead of trying to develop everything from scratch (especially in "boring" areas like lexing, parsing...) With all the above in mind - that project should cater for wide audience of C/GCC users, the project is also in unique position to target other audiences, so they shouldn't be overlooked. Python is well-known for its beginner-friendliness, so if someone will want to hack on compilers without knowledge of Python, let's make PPCI to be non-sucky platform to start with. That means things like https://github.com/windelbouwman/ppci-mirror/commit/726dea002d8d54afcf2aa966a130e55fbaac1294 , and just writing docs (at least intros/basic usage) in a way to not rely on intimate Python knowledge. One interesting question re: beginners is terminology and state-of-the-art quirks, inherited from big project like GCC (which in turn inherited it from decades of computing industry history). My IMHO is yes, that we should teach them real terminology, which they can reuse outside the PPCI sandbox. So, linker should be called "ld" and not something else, and uninitialized data segment - BSS. (But having a glossary may be helpful ;-) ). Then all other niche target audiences mentioned above - embedded engineers with non-standard hardware, computer archeologists, etc. They should be covered by: a) being compatible with and interoperable with existing tools, and b) good documentation.
gharchive/issue
2019-09-02T14:38:23
2025-04-01T06:46:15.271439
{ "authors": [ "pfalcon", "windelbouwman" ], "repo": "windelbouwman/ppci-mirror", "url": "https://github.com/windelbouwman/ppci-mirror/issues/11", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
2583815275
🛑 IST - API is down In 0441d6a, IST - API (https://controld-network-api.ncprelief.eu.org?iata_code=IST&status=api) was down: HTTP code: 503 Response time: 107 ms Resolved: IST - API is back up in fae0cd9 after 11 hours, 55 minutes.
gharchive/issue
2024-10-13T09:27:43
2025-04-01T06:46:15.333602
{ "authors": [ "windstatus" ], "repo": "windstatus/Control-D-DNS-status-uptime-monitor", "url": "https://github.com/windstatus/Control-D-DNS-status-uptime-monitor/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
23511114
Doesn't throw error when passed circular structure as meta hope that makes sense? winston = require('winston') winston.add(winston.transports.File, { filename: './app.log' }); // script just stops here, silently, no error message // When I stop the script, it's still running in the background and eating up all my CPU / memory winston.error("bla bla", getCircularStructur()) If I do the same with a JSON.stringify, it throws the error: console.error(JSON.stringify(getCircularStructur())) Closing, see #1248 , in winston@3 you can define custom formatters that do whatever you like with your data -- e.g. managing decycling behavior for circular objects.
gharchive/issue
2013-11-29T22:49:11
2025-04-01T06:46:15.375061
{ "authors": [ "DABH", "gr2m" ], "repo": "winstonjs/winston", "url": "https://github.com/winstonjs/winston/issues/331", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
95359443
Log rotation fails in simple app I tried Hello there, I tried log rotation and it doesn't create new file when maxsize is reached. sample app var winston = require('winston'); winston.add(winston.transports.File, { filename: "log.log", maxsize: 512000, //500KB maxFiles: 1000, tailable : true }); winston.remove(winston.transports.Console); for (i=0; i<100000; i++) { winston.info('0123456789_' + i); } I'm using node -v : v0.10.28 winston : 1.0.1 I tried this on linux and windows and Mac. with same version of node n winston. Same output Looks like a duplicate of #477.
gharchive/issue
2015-07-16T06:12:09
2025-04-01T06:46:15.379330
{ "authors": [ "devsmart", "indexzero" ], "repo": "winstonjs/winston", "url": "https://github.com/winstonjs/winston/issues/666", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1887168389
there's no public directory There's no /public directory, and install doesn't create it, so composer instalation doesn't work "out of the box" By default there's no public directory, but the install should still work out of the box, as the public directory is not a requirement (although it is good practice). If you wish to have a public directory to point the web server to, you simply need to run php artisan winter:mirror public/ after installation. See: https://wintercms.com/docs/v1.2/docs/setup/configuration#using-a-public-folder See https://wintercms.com/docs/v1.2/docs/console/setup-maintenance#mirror-public-files. It's a holdover from the original platform being designed to be installed via a wordpress-like web installer that there isn't a public folder by default as on shared hosts you'd typically be installing the project into the "public" folder already. My personal recommendation and own process has always been to use a public folder though; and the example project now includes that as well: https://github.com/wintercms/wn-blizzard.wintercms.com-site/commit/a5395948b2c54cf27518415a5cdc277ef5c557da @bennothommo too quick on the draw, you beat me too it 😂 GitHub really needs one of those "Someone is typing a comment..." notifications like Facebook.
gharchive/issue
2023-09-08T08:00:25
2025-04-01T06:46:15.388205
{ "authors": [ "LukeTowers", "alagoroChapela", "bennothommo" ], "repo": "wintercms/docs", "url": "https://github.com/wintercms/docs/issues/156", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
528992983
Add a factorial function It would be cool if the calculator provided a way to calculate factorials, e.g. n n! 0 1 1 1 2 2 3 6 4 24 It looks like we just need to expose a new factorial function here: https://github.com/wintron/calculator/blob/a1c403bae367801855a3974226a10c9cf01cfa22/api/controllers/arithmeticController.js#L13-L18 OK, let's estimate how big of a task this will be, using T-shirt sizing. /polls Small Medium Large XL /polls Small Medium Large XL /polls Option1 'Option 2' "Option 3"
gharchive/issue
2019-11-26T21:54:45
2025-04-01T06:46:15.392702
{ "authors": [ "swinton" ], "repo": "wintron/calculator", "url": "https://github.com/wintron/calculator/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2721318597
fix: remove useless escape character in content-collections.mdx Description (required) Same as #10209 but with a different pattern: I haven't noticed this pattern in the previous PR because I was looking for an exact matching... but I checked "**\/*.md" and "**/*.md" are both valid. So, to avoid confusion whether the editor automatically removes the character or ESlint displays a warning, I removed the useless escape character in some loader patterns in content-collections.mdx. Related issues & labels (optional) Suggested label: code snippet update Lunaria Status Overview 🌕 This pull request will trigger status changes. Learn more By default, every PR changing files present in the Lunaria configuration's files property will be considered and trigger status changes accordingly. You can change this by adding one of the keywords present in the ignoreKeywords property in your Lunaria configuration file in the PR's title (ignoring all files) or by including a tracker directive in the merged commit's description. Tracked Files File Note en/guides/content-collections.mdx Source changed, localizations will be marked as outdated. Warnings reference Icon Description 🔄️ The source for this localization has been updated since the creation of this pull request, make sure all changes in the source have been applied.
gharchive/pull-request
2024-12-05T20:23:49
2025-04-01T06:46:15.516992
{ "authors": [ "ArmandPhilippot", "astrobot-houston" ], "repo": "withastro/docs", "url": "https://github.com/withastro/docs/pull/10213", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2165381649
i18n(zh-cn): Update styling.mdx Description (required) Update styling.mdx Related issues & labels (optional) #7126 !coauthor
gharchive/pull-request
2024-03-03T14:19:23
2025-04-01T06:46:15.518999
{ "authors": [ "huyikai", "liruifengv" ], "repo": "withastro/docs", "url": "https://github.com/withastro/docs/pull/7157", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2239117634
i18n(ko-KR): update dev-toolbar-app-reference.mdx Description (required) update dev-toolbar-app-reference.mdx #7846 Related issues & labels (optional) Closes # Suggested label: Lunaria Status Overview 🌕 This pull request will trigger status changes. Learn more By default, every PR changing files present in the Lunaria configuration's files property will be considered and trigger status changes accordingly. You can change this by adding one of the keywords present in the ignoreKeywords property in your Lunaria configuration file in the PR's title (ignoring all files) or by including a tracker directive in the merged commit's description. Tracked Files File Note Locale src/content/docs/ko/reference/dev-toolbar-app-reference.mdx Localization changed, will be marked as complete. ko Warnings reference Icon Description 🔄️ The source for this localization has been updated since the creation of this pull request, make sure all changes in the source have been applied.
gharchive/pull-request
2024-04-12T05:30:57
2025-04-01T06:46:15.526435
{ "authors": [ "astrobot-houston", "jsparkdev" ], "repo": "withastro/docs", "url": "https://github.com/withastro/docs/pull/7875", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2245963726
Update auto.mdx Changed "run dev" command to "astro dev" Description (required) Related issues & labels (optional) Closes # Suggested label: Hello! Thank you for opening your first PR to Astro’s Docs! 🎉 Here’s what will happen next: Our GitHub bots will run to check your changes. If they spot any broken links you will see some error messages on this PR. Don’t hesitate to ask any questions if you’re not sure what these mean! In a few minutes, you’ll be able to see a preview of your changes on Vercel 🥳 One or more of our maintainers will take a look and may ask you to make changes. We try to be responsive, but don’t worry if this takes a few days. Lunaria Status Overview 🌕 This pull request will trigger status changes. Learn more By default, every PR changing files present in the Lunaria configuration's files property will be considered and trigger status changes accordingly. You can change this by adding one of the keywords present in the ignoreKeywords property in your Lunaria configuration file in the PR's title (ignoring all files) or by including a tracker directive in the merged commit's description. Tracked Files File Note Locale src/content/docs/en/install/auto.mdx Source changed, localizations will be marked as outdated. en Warnings reference Icon Description 🔄️ The source for this localization has been updated since the creation of this pull request, make sure all changes in the source have been applied. Hi @suvires ! As Atharva mentioned, we prefer to show the general scripts appropriate for the various package managers. So I'll be closing this PR, but we appreciate your efforts to help us keep the docs updated and correct! Please do make another PR if you spot an error in future! :rocket:
gharchive/pull-request
2024-04-16T12:47:48
2025-04-01T06:46:15.536015
{ "authors": [ "astrobot-houston", "sarah11918", "suvires" ], "repo": "withastro/docs", "url": "https://github.com/withastro/docs/pull/7908", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1740675730
edits to guides content Some text edits to guide content for: components i18n I tried to make sure all apostrophes are not single quotes! :sweat_smile: Feel free to edit as necessary and merge whatever you like, Chris! :raised_hands: size-limit report 📦 Path Size /index.html 13.05 KB (0%) /_astro/*.js 16.51 KB (0%) /_astro/*.css 9.02 KB (0%) Will merge this once the checks pass!
gharchive/pull-request
2023-06-05T00:13:28
2025-04-01T06:46:15.539390
{ "authors": [ "astrobot-houston", "sarah11918" ], "repo": "withastro/starlight", "url": "https://github.com/withastro/starlight/pull/137", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
199132534
Dot-notation for records This PR adds support for code like [#foo].tag in a search section. There is an issue that prevents it from working in an action section (scans should not generate proposals if a dependent variable is covered by some instance of GenerateId provider that hasn't been accepted yet), but I'll work on this in subsequent PR. Let me know if there is anything you'd like me to change in the code. This change is  @joshuafcole @cmontella do either of you have any objections to this? I think I was the main opponent, but I'm not strongly opposed - more just a slight unease. I think it'll make code a little harder to read in some cases, but it seems like a natural thing to want to do. I don't have any specific concerns though, so I think it's fine to allow this. Oh, if you don't want this, I'll be fine without it :) My reasoning was It seemed intuitive that if foo = [#some-record] foo.attribute then [#some-record].attribute is expected to work too. But I admit that it may be far-fetched to apply this traditional reasoning to eve. Especially since my initial use-case was based on incorrect assumption that without this PR there is no way to simplify following code: [#foo bar] (not bar) But I've learned about [#foo not(bar)] since then. The big use case here is for inline queries in maths, e.g. [#person age < 18].height / [#teacher not(#professor)].height In principle it sounds like a good thing to support, but every time I write an example of it I find it relatively hard to read. This is actually something I was advocating for internally earlier, but after trying it out with the logical syntax extension here, I'm not so sure. I can't think of a particularly better syntax either. There's an argument to be made that we should support it anyway for orthogonal completeness, but it doesn't seem super compelling to me. I'd say I'm about a 2/10 against this. If you feel more strongly that we should have it I'm okay with that. On Tue, Jan 10, 2017 at 1:59 PM Ravil Bayramgalin notifications@github.com wrote: Oh, if you don't want this, I'll be fine without it :) My reasoning was It seemed intuitive that if foo = [#some-record] foo.attribute then [#some-record].attribute is expected to work too. But I admit that it may be far-fetched to apply this traditional reasoning to eve. Especially since my initial use-case was based on incorrect assumption that without this PR there is no way to simplify following code: [#foo bar] (not bar) But I've learned about [#foo not(bar)] since then. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/witheve/Eve/pull/722#issuecomment-271711197, or mute the thread https://github.com/notifications/unsubscribe-auth/AATKDrC4Khwbn4Pvx7aaaHZHVw9uV2vGks5rQ_8kgaJpZM4LccJY . Please correct me if I'm wrong in my following assumptions. It seems like the core team does not have strong arguments against this PR, but does have a minor gut feeling against it. And the team is ok to merge it if one person (who wrote the PR and is outside of the core team) strongly desires it. Is this correct? In my opinion, seems like there are better ways to make the decision. A gut feeling of the whole core team should be enough to stop the PR. Community feedback could be gathered if the team is undecided/split. It'll hopefully provide an additional information that could be used by the team to make the final decision. If there is no proponent of PR on the team, then the burden of conviction lies on an outside person (that could include initiating a gathering of community feedback by that person). Until the team is convinced that a PR is' a good idea, it shouldn't be merged. A simple reason to close the PR from the core team like "not convinced it will help" is sufficient. What do you think? @brainopia I think it's a minor enough change that we should just allow it and see how it works. Like you said, it makes sens, and as josh said, it we might want it just as an option for completeness. I say let's merge it and see how it goes. @cmontella It should go the other way around. Anything we do that affects the actual language should be justified. That doesn't mean we won't eventually find a good reason for this, but we don't currently have one and its use does make things harder to read. As @brainopia the gut feeling is this probably shouldn't happen. I opened it up because @joshuafcole has previously expressed a desire for it, but he's lukewarm on it at best. Let's close this for now. If someone comes up with a compelling justification for it, we'll revisit. @ibdknox Sounds good to me, thanks for making a decision :+1: @ibdknox by the way, I have another question. Would you please check two first commits in the PR and tell me if you're interested in merging them. There are no changes in semantics/syntax, but in my opinion, it is a slight improvement of parser by removing a couple of redundant abstractions. Scratch my question. It's irrelevant now :)
gharchive/pull-request
2017-01-06T06:22:21
2025-04-01T06:46:15.553951
{ "authors": [ "brainopia", "cmontella", "ibdknox", "joshuafcole" ], "repo": "witheve/Eve", "url": "https://github.com/witheve/Eve/pull/722", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2140094310
لینک زیرمجموعه گیری سلام خسته نباشید بعد از تغییر متن بنر بازاریابی ، لینک دعوت به صورت کپی شدن در اومده و با کلیک مستقیم به صفحه مورد نظر هدایت نمیشه مشکل از چی میتونه باشه؟ متن به حالت قبلی برگردوندم ولی مشکل همچنان هست!!! درود اگر خودتون تغییر دادین خب برش گردونین به حالت قبل سلام متن حالت قبل تست کردم ولی باز هم لینک دعوت به همین صورت هست فقط متن بنر عوض شده و کار خاصی نکردم خب کاربر کپی کن بعد بزنه روش این درست ولی بعضیا با رسم شکل هم نمیتونن😂 یجورایی حلش میکنیم، ممنون
gharchive/issue
2024-02-17T13:35:19
2025-04-01T06:46:15.603139
{ "authors": [ "AliHemmati79", "wizwizdev" ], "repo": "wizwizdev/wizwizxui-timebot", "url": "https://github.com/wizwizdev/wizwizxui-timebot/issues/853", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
453420295
请教一下数据处理 请问以下代码后面有*****的那一句是什么意思?我不会传图片,不好意思。为什么要设为1呢?另外,看到你代码注释mix up weight default to 1. 这是什么意思? 请楼主有空帮我看看。谢谢了 y_true_13[..., -1] = 1.***** y_true_26[..., -1] = 1.***** y_true_52[..., -1] = 1.***** y_true[feature_map_group][y, x, k, :2] = box_centers[i] y_true[feature_map_group][y, x, k, 2:4] = box_sizes[i] y_true[feature_map_group][y, x, k, 4] = 1. ******* y_true[feature_map_group][y, x, k, 5 + c] = 1. ******* y_true[feature_map_group][y, x, k, -1] = boxes[i, -1] 带 * 号部分分别是指定是不是前/背景,已经分类分到哪一类,是 one-hot 形式。 mix up 使用时,损失函数对应部分也要加权一下。默认1表示没有加权。
gharchive/issue
2019-06-07T09:27:50
2025-04-01T06:46:15.605041
{ "authors": [ "ZH-Lee", "wizyoung" ], "repo": "wizyoung/YOLOv3_TensorFlow", "url": "https://github.com/wizyoung/YOLOv3_TensorFlow/issues/88", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1585680711
support multiple API versions ? Maybe we should think about something "now" to assure future compatibility with new API versions. (while staying at same time backwards compatible on Clojure level) Maybe nothing is needed, just to be sure. I somehow think that Azure will come with a new API version "soon", as current is "rather incomplete" What I mean is, that if we keep the API as it is today; we might come into a situation when the REST API wrapped change in an incompatible way, we cannot support both version of the REST in the same Clojure library. One reason for this are the "function names" in namespace api Imagine that a version v2 of OpenAI REST removes any of the functions. The REST API defined via paths is more "flexible" then the code here. For Azure is see it "likely" that the next version will have more functions (which is less of a problem the less functions, ut a problem as well) I think the only "reasonable way out" and be full future proof, is to give the same flexibility to the library then REST has, which means to remove the api ns, and add 2 more parameters to the params namely version and operation A new API version will most likely mean a new major version of this library to support the new API (and only the new API) depending on what's changed. In the case of removal of an API operation, the library will deprecate the function accordingly. The whole point of the api namespace is to provide a meaningful Clojure wrapper over operations exposed by the API. These named functions provide the developer with docs, code completion and a nicer experience. What you're suggesting almost negates the need for this library completely, as the developer might just as well use Martian directly instead and gain full flexibility. A new API version will most likely mean a new major version of this library to support the new API (and only the new API) depending on what's changed. In the case of removal of an API operation, the library will deprecate the function accordingly. The whole point of the api namespace is to provide a meaningful Clojure wrapper over operations exposed by the API. These named functions provide the developer with docs, code completion and a nicer experience. What you're suggesting almost negates the need for this library completely, as the developer might just as well use Martian directly instead and gain full flexibility. Yes.... I do agree with this... I came as well to a similar conclusion.. Using REST directly from Clojure is nearly as simple as it would be when following my proposal. A new API version will most likely mean a new major version of this library to support the new API (and only the new API) depending on what's changed. In the case of removal of an API operation, the library will deprecate the function accordingly. "New major version" could indeed work. (but now it support 2 REST APIs,,,, with different release cycles.) In this light , maybe my idea of unifying the 2 APIs was not that good. But ok, maybe we could still couple it to version, by just making a table such as, and increase major version, if any of th2 2 changes, whcih should not be too often... Supported versions: libray version OpenAI REST version Azure OpenAI REST version 1.0.0 v1 2022-12-01 Exactly what I was thinking :) On Fri, Feb 17, 2023, 1:02 PM Carsten Behring @.***> wrote: A new API version will most likely mean a new major version of this library to support the new API (and only the new API) depending on what's changed. In the case of removal of an API operation, the library will deprecate https://guide.clojure.style/#deprecated the function accordingly. "New major version" could indeed work. (but now it support 2 REST APIs,,,, with different release cycles.) In this light , maybe my idea of unifying the 2 APIs was not that good. But ok, maybe we could still couple it to version, by just making a table such as, and increase major version, if any of th2 2 changes, whcih should not be too often... Supported versions: libray version OpenAI REST version Azure OpenAI REST version 1.0.0 v1 2022-12-01 — Reply to this email directly, view it on GitHub https://github.com/wkok/openai-clojure/issues/4#issuecomment-1434622617, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABHY7OHLF4HZOJBTYDHIAETWX5ZHZANCNFSM6AAAAAAU4W4CTY . You are receiving this because you commented.Message ID: @.***> Ok, so we can close the issue.
gharchive/issue
2023-02-15T11:19:36
2025-04-01T06:46:15.659221
{ "authors": [ "behrica", "wkok" ], "repo": "wkok/openai-clojure", "url": "https://github.com/wkok/openai-clojure/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2114700983
Override list item styling for select menu items The menu items on the results view ended up showing a margin inherited from app.scss 's style definition for <li> elements (meant for lists like the FAQ and instructions). This PR overrides the inherited style definitions. Bug: T348271 This is already handled as part of #865! closing this
gharchive/pull-request
2024-02-02T11:21:26
2025-04-01T06:46:15.671304
{ "authors": [ "chukarave" ], "repo": "wmde/wikidata-mismatch-finder", "url": "https://github.com/wmde/wikidata-mismatch-finder/pull/873", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
422236756
[typescript] withRR4 type is not undefined Hi @jungpaeng withRr4 was removed from index to prevent error of neededing react router as a dep as it is a peer dependency. https://github.com/wmira/react-sidenav/issues/65 Please see the issue linked above Please see the issue linked above I really want to thank you for your help!
gharchive/issue
2019-03-18T14:06:05
2025-04-01T06:46:15.673731
{ "authors": [ "jungpaeng", "wmira" ], "repo": "wmira/react-sidenav", "url": "https://github.com/wmira/react-sidenav/issues/72", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2198373195
Updated HTML payload step 2 step 5
gharchive/pull-request
2024-03-20T19:19:46
2025-04-01T06:46:15.769961
{ "authors": [ "wolftales" ], "repo": "wolftales/hashicat-app", "url": "https://github.com/wolftales/hashicat-app/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
243062119
expose tapToAutoFocusListener expose tapToAutoFocusListener so that we can observe focus changes Hey @NickPontiff If you are still interested in this being considered, can you take a look at resolving the merge conflict? Thanks for the work here!
gharchive/pull-request
2017-07-14T17:36:31
2025-04-01T06:46:15.774571
{ "authors": [ "NickPontiff", "austinkettner" ], "repo": "wonderkiln/CameraKit-Android", "url": "https://github.com/wonderkiln/CameraKit-Android/pull/134", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1153637482
Provisioning templated dashboard with variables doesn't work When provisioning template dashboard, It outputs the following error: Error updating options datasource named metrics was not found At First, For solve the issue, I tried using environment variables. But, It didn't work. I solved the problem by trying the following: Update Name of 'datasource' in dashboard templated
gharchive/issue
2022-02-28T04:39:59
2025-04-01T06:46:15.938050
{ "authors": [ "woodonggyu" ], "repo": "woodonggyu/system-monitor", "url": "https://github.com/woodonggyu/system-monitor/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1419515099
Use modern error handling and enforce it via lint use error un-/wraping with errors.Is and errors.As we need to make sure this does not introduce regressions Codecov Report Base: 44.74% // Head: 44.80% // Increases project coverage by +0.06% :tada: Coverage data is based on head (974162b) compared to base (f88c70b). Patch coverage: 28.57% of modified lines in pull request are covered. Additional details and impacted files @@ Coverage Diff @@ ## master #1327 +/- ## ========================================== + Coverage 44.74% 44.80% +0.06% ========================================== Files 137 137 Lines 9868 9864 -4 ========================================== + Hits 4415 4420 +5 + Misses 5191 5182 -9 Partials 262 262 Impacted Files Coverage Δ pipeline/frontend/yaml/error.go 0.00% <ø> (ø) server/model/cron.go 0.00% <0.00%> (ø) server/remote/gitea/gitea.go 32.59% <0.00%> (ø) server/remote/gitlab/gitlab.go 21.97% <0.00%> (ø) server/shared/userSyncer.go 0.00% <0.00%> (ø) ...er/store/datastore/migration/000_legacy_to_xorm.go 63.71% <0.00%> (ø) server/store/datastore/migration/common.go 38.94% <0.00%> (ø) server/store/datastore/migration/migration.go 36.00% <0.00%> (ø) server/model/secret.go 49.01% <66.66%> (ø) server/cron/cron.go 36.70% <100.00%> (ø) ... and 2 more Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here. :umbrella: View full report at Codecov. :loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
gharchive/pull-request
2022-10-22T21:19:15
2025-04-01T06:46:15.954073
{ "authors": [ "6543", "codecov-commenter" ], "repo": "woodpecker-ci/woodpecker", "url": "https://github.com/woodpecker-ci/woodpecker/pull/1327", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
548485786
fzf floating window background color https://github.com/wookayin/dotfiles/blob/b5a071c54cabe043ee5d39bd4151fe8992d9f6b5/vim/vimrc#L941 https://github.com/wookayin/dotfiles/blob/b5a071c54cabe043ee5d39bd4151fe8992d9f6b5/vim/vimrc#L964 highlight doesn't work for one of my machines. I fixed this by adding --color=bg:#1a2a31 at the end of the line 941. Do you think it causes some other changes? Seems vim is not supporting 24-bit color. What does the termguicolors variable read? set termguicolors? In addition, how does your environments look like -- vim version, terminal emulator, etc.? echo &termguicolors returns 1 iterms 3.3.7 nvim 0.4.3 Closing as the issue being stale. Please re-open if you can reproduce the problem.
gharchive/issue
2020-01-11T22:03:35
2025-04-01T06:46:15.960474
{ "authors": [ "sungminoh", "wookayin" ], "repo": "wookayin/dotfiles", "url": "https://github.com/wookayin/dotfiles/issues/20", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
717192080
[update] New syscall 'svc_alarm' The new svc_alarm syscall triggers a user handler after some user defined amount of time (in milliseconds) Great ! Thanks @arnauldm !
gharchive/pull-request
2020-10-08T09:48:40
2025-04-01T06:46:15.961645
{ "authors": [ "PThierry", "arnauldm" ], "repo": "wookey-project/ewok-kernel", "url": "https://github.com/wookey-project/ewok-kernel/pull/54", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
230447143
$config group Setting xcode configuration files for each config results in the creation of a top level group in the project called $config, even though none of the files need to be in the list of files included in the project targets: My app: configurations: debug: configs/my-app.xcconfig release: configs/my-app-release.xcconfig Hi @yonaskolb, This is intentional. This is done to allow you to edit the xcconfig files from within Xcode as an alternative to using the grouped build settings view. Ok. If multiple targets have config files though, only those for the first target are added to that group. Would configurations be a better name than $config? I guess you were trying to avoid name clashes Hi @yonaskolb, I've just tested this on a sample project with two targets that have their own unique xcconfig files: targets: My app: type: ":application" sources: sources configurations: debug: config-target.xcconfig release: config-target-release.xcconfig My other app: type: ":application" sources: sources configurations: debug: config-target-2.xcconfig release: config-target-2-release.xcconfig With this configuration, all xcconfig files are present: Yeah, I just tested this again. It does work as you say, but only if those config files are included in a target's sources. Makes sense. And if I didn't want to see the in the groups, I can leave them out of sources 👍
gharchive/issue
2017-05-22T16:15:10
2025-04-01T06:46:16.075968
{ "authors": [ "lyptt", "yonaskolb" ], "repo": "workshop/struct", "url": "https://github.com/workshop/struct/issues/130", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1303678212
Add apache commons-math3 for discrete Laplace random value This change is  This is for PR https://github.com/world-federation-of-advertisers/cross-media-measurement/pull/619
gharchive/pull-request
2022-07-13T16:30:26
2025-04-01T06:46:16.079304
{ "authors": [ "iverson52000" ], "repo": "world-federation-of-advertisers/common-jvm", "url": "https://github.com/world-federation-of-advertisers/common-jvm/pull/137", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1863366894
Tefera gh pages Help files (help_file.html and help_file.md) on PIP Stata page have been updated. Hi @Tefera19 , Thank you. PLease make sure you use the most recent version of pip. Right now, it is 0.10.7. This should have fixed the issue of the server. To see which is the most recent version, go here: https://github.com/worldbank/pip/tags
gharchive/pull-request
2023-08-23T13:42:01
2025-04-01T06:46:16.081294
{ "authors": [ "Tefera19", "randrescastaneda" ], "repo": "worldbank/pip", "url": "https://github.com/worldbank/pip/pull/82", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1088869236
Fix function _getfilepath This function is only looking at the string, and not checking whether the folder path actually exists. So if someone simply write the name of a do-file without the file format .do, it will not understand that it's meant to be a do-file. [ ] Switch all slashes to forward slashes [ ] Check if input contains any dots [ ] If it does, extract the last one and consider it the file format [ ] Test that the file format is .do [ ] If it is, that's it [ ] If it is not, return an error [ ] If it does not, test if it is a folder (that is, check that a folder with that name exists) [ ] If it is, consider that's it [ ] If it is not, test that a do file with that name + . do exists [ ] If it does, return the name with .do in the end [ ] If it does not, return an error Just a quick comment, for the second point, instead of finding any dots, python's os package has "os.path.basename" function to return the file name (I guess if it's a folder, it returns nothing). Or, more directly, you may be able to use "os.path.splitext" to get the file extension and see if it's a do file or not. Thanks, Hiro! This is great, but this part of the code is implemented in Stata, which unfortunately does not have this functionality automated... oh no! sorry then never mind
gharchive/issue
2021-12-27T00:06:58
2025-04-01T06:46:16.086620
{ "authors": [ "luizaandrade", "mizuhirosuzuki" ], "repo": "worldbank/stata-linter", "url": "https://github.com/worldbank/stata-linter/issues/29", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2952505
Bugfix/last child on delete /cc @joshdrake @adamfast this will make django-threadedcomments require Django>=1.3. Let me know if that's ok.
gharchive/issue
2012-01-24T16:10:06
2025-04-01T06:46:16.087772
{ "authors": [ "fcurella" ], "repo": "worldcompany/django-threadedcomments", "url": "https://github.com/worldcompany/django-threadedcomments/issues/1", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1718300535
🛑 Wouter Planet Cloud is down In 3b21aa1, Wouter Planet Cloud (https://cloud.wouterplanet.com) was down: HTTP code: 523 Response time: 2716 ms Resolved: Wouter Planet Cloud is back up in 4f9699f.
gharchive/issue
2023-05-20T21:46:32
2025-04-01T06:46:16.116948
{ "authors": [ "woeterman94" ], "repo": "wouterplanet/monitoring", "url": "https://github.com/wouterplanet/monitoring/issues/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
373981179
Fix golint import path I am a bot. Please reach out to @azillion if you have any issues, or just close the PR. Coverage increased (+26.0%) to 88.889% when pulling 55c4f0e12b0d7117352a89e200ea3aa71c989522 on golint-fixer:master into 66144498bd20e404a6fe506ee80834e2d360f3af on writescript:master.
gharchive/pull-request
2018-10-25T14:38:24
2025-04-01T06:46:16.241712
{ "authors": [ "coveralls", "golint-fixer" ], "repo": "writescript/writescript", "url": "https://github.com/writescript/writescript/pull/12", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
267485880
Postgres 10.0 Error Error running query on database: pg_stat_replication pq: function pg_current_xlog_location() does not exist\n" source="postgres_exporter.go:873 Presumably caused by the rename of pg_xlog to pg_wal https://git.postgresql.org/gitweb/?p=postgresql.git;a=commit;h=f82ec32ac30ae7e3ec7c84067192535b2ff8ec0e Could there either be an option to disable the built in queries or just consider removing them all together and go entirely with an external file to supply the queries to run? When the internal ones break like this due to catalog changes, it causes issues that are hard to then fix without hacking on the go source code and recompiling. That's a good idea actually. PG 10s release has coincided unfortunately worth a busy time for me so I've not been on the ball with updating for it. I need to get to it soon. On 23 October 2017 2:30:41 pm AEDT, Keith Fiske notifications@github.com wrote: Could there either be an option to disable the built in queries or just consider removing them all together and go entirely with an external file to supply the queries to run? When the internal ones break like this due to catalog changes, it causes issues that are hard to then fix without hacking on the go source code and recompiling. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/wrouesnel/postgres_exporter/issues/119#issuecomment-338539046 -- Sent from my Android device with K-9 Mail. Please excuse my brevity. Removing the built in queries would certainly cut down on the required maintenance of the core code. Allowing multiple query files to be defined would be nice as well. I tried defining -extend.query-path more than once, but it only uses the last one defined. Would allow a core set of queries to use the same file everywhere and add additional ones as needed on certain systems.
gharchive/issue
2017-10-22T18:09:09
2025-04-01T06:46:16.247933
{ "authors": [ "Zolmeister", "keithf4", "wrouesnel" ], "repo": "wrouesnel/postgres_exporter", "url": "https://github.com/wrouesnel/postgres_exporter/issues/119", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
731703570
ResetAllEspmNames keyword missing In the farmhouseChimneys.txt script, there is a line: ResetAllEspmNames Which is parsed as an Identifier in an expr instead of a Keyword. Okay, I'm stupid, I thought this was some kind of priority issue but it's just the keyword missing in the Keyword rule... Here is the list of keyword I use in my modified wizard.g41: Keyword: 'DeSelectAll' | 'DeSelectAllPlugins' | 'DeSelectAllEspms' | 'DeSelectPlugin' | 'DeSelectEspm' | 'DeSelectSubPackage' | 'Note' | 'RenamePlugin' | 'RenameEspm' | 'RequireVersions' | 'ResetPluginName' | 'ResetEspmName' | 'ResetAllEspmNames' | 'ResetAllPlugins' | 'SelectAll' | 'SelectAllPlugins' | 'SelectAllEspms' | 'SelectPlugin' | 'SelectEspm' | 'SelectSubPackage'; 1 I currently keep a custom version in my repository but I'll remove it when this version get everything fixed. NB: ResetAllPlugins does not exist, the right keyword is ResetAllPluginNames. NB: ResetAllPlugins does not exist, the right keyword is ResetAllPluginNames. Thanks, looks like I never encountered it in all the mods I tested 🙃 I added the *Plugin* versions of all keywords in https://github.com/wrye-bash/wrye-bash/commit/3426384083bd5d7c61d42730ed7fa9c5629bc2db, so they haven't been out for very long. Especially since almost no new wizards get written these days. Ah, nice to know. Shame this is not use more... There are a few things I found annoying in the scripting language but overall it's pretty nice. I've made a (I think) fully functional installer for Mod Organizer 2 based on my interpreter and a slightly modified version of your grammar which might be included in the next MO2 release, so maybe people will start to write script again 🙃 It's definitely superior to FOMOD, simply by virtue of having a specification, buggy and incomplete as it is :stuck_out_tongue:
gharchive/issue
2020-10-28T18:49:13
2025-04-01T06:46:16.254095
{ "authors": [ "Holt59", "Infernio" ], "repo": "wrye-bash/wizparse", "url": "https://github.com/wrye-bash/wizparse/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
459696265
Cannot provide scaling thresholds for zero scaling configurations with KNative Currently we support zero scaling configurations as below. .. autoscaling: { policy: { minReplicas: 0, maxReplicas: 2 }, overridable: true } .... But, we should be able to provide the request numbers threshold or something similar in which the scaling should be performed. We need to investigate more on this, and providing this support. Current auto scaling policy configuration is tighly bound to cpuUtilization with below syntax, we should support number of request if the reported problem can be solved. cellery:AutoScalingPolicy scalePolicy1 = { minReplicas: 0, maxReplicas: 2, cpuPercentage: { percentage: 10 } };
gharchive/issue
2019-06-24T05:34:47
2025-04-01T06:46:16.274600
{ "authors": [ "sinthuja" ], "repo": "wso2-cellery/sdk", "url": "https://github.com/wso2-cellery/sdk/issues/419", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1889982907
500 error occurs when refreshing Business Info and API definition pages Description 500 error occurs when refreshing Business Info and API definition pages. Steps to Reproduce Login to APIM publisher portal Deploy an API Navigate to Business Info/ API definition pages Refresh the page. Affected Component APIM Version 4.2.0 Environment Details (with versions) No response Relevant Log Output [2023-09-11 14:05:05,973] ERROR - AuthenticationValve Error while normalizing the request URI to process the authentication: java.net.URISyntaxException: Illegal character in path at index 61: /publisher/apis/52dd7d90-1b70-44cd-bc62-197071453c83/business info at java.net.URI$Parser.fail(URI.java:2913) ~[?:?] at java.net.URI$Parser.checkChars(URI.java:3084) ~[?:?] at java.net.URI$Parser.parseHierarchical(URI.java:3166) ~[?:?] at java.net.URI$Parser.parse(URI.java:3125) ~[?:?] at java.net.URI.<init>(URI.java:600) ~[?:?] at org.wso2.carbon.identity.auth.service.util.AuthConfigurationUtil.getNormalizedRequestURI(AuthConfigurationUtil.java:346) ~[org.wso2.carbon.identity.auth.service_1.7.1.2.jar:?] at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:105) ~[org.wso2.carbon.identity.auth.valve_1.7.1.2.jar:?] at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:106) ~[org.wso2.carbon.tomcat.ext_4.8.1.jar:?] at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49) ~[org.wso2 Related Issues No response Suggested Labels No response This issue is not reproduced in APIM 4.2.0 or 4.3.0.
gharchive/issue
2023-09-11T08:39:36
2025-04-01T06:46:16.279203
{ "authors": [ "HeshanSudarshana", "chamilaadhi" ], "repo": "wso2/api-manager", "url": "https://github.com/wso2/api-manager/issues/2161", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2170615236
Change gateway name Purpose $subject Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Project coverage is 80.21%. Comparing base (248df53) to head (68cfc18). Report is 15 commits behind head on main. Additional details and impacted files @@ Coverage Diff @@ ## main #2090 +/- ## =========================================== + Coverage 26.09% 80.21% +54.12% =========================================== Files 47 9 -38 Lines 9205 551 -8654 Branches 0 246 +246 =========================================== - Hits 2402 442 -1960 + Misses 6672 109 -6563 + Partials 131 0 -131 Flag Coverage Δ adapter ? common-controller ? idp-domain-service 80.21% <ø> (?) Flags with carried forward coverage won't be shown. Click here to find out more. :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2024-03-06T04:31:50
2025-04-01T06:46:16.287012
{ "authors": [ "Tharsanan1", "codecov-commenter" ], "repo": "wso2/apk", "url": "https://github.com/wso2/apk/pull/2090", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1161362455
Websocket port not getting updated with port offset Location : https://apim.docs.wso2.com/en/latest/use-cases/streaming-usecase/create-streaming-api/create-a-websocket-streaming-api/ I am using the GA version and the websocket port does not get updated on the UIs when a port offset is given The following configuration should be provided in the deployment.toml file in order to do this [[apim.gateway.environment]] ws_endpoint = "ws://localhost:9190" Note: 9190 = default port 9090 + port_offset @Mariangela let's add this as a note to the docs as discussed offline
gharchive/issue
2022-03-07T12:34:27
2025-04-01T06:46:16.314981
{ "authors": [ "Shenavi", "senthuran16" ], "repo": "wso2/docs-apim", "url": "https://github.com/wso2/docs-apim/issues/5403", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
878367774
Remove gerund in left nav titles Purpose Changed Performing Root Cause Analysis to Perform Root Cause Analysis because we have avoided gerunds in all other left nav titles. @Nashaath Pls check this. Shall we fix all to be consistent? https://github.com/wso2-enterprise/choreo/issues/3786 I sent the page name changes via https://github.com/wso2/docs-choreo-dev/pull/58
gharchive/pull-request
2021-05-07T02:57:15
2025-04-01T06:46:16.316998
{ "authors": [ "Nashaath", "anupama-pathirage" ], "repo": "wso2/docs-choreo-dev", "url": "https://github.com/wso2/docs-choreo-dev/pull/57", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
516979432
Add Ingress support for WSO2 Open Banking Business Intelligence Dashboard Description: Add Ingress support for the relevant WSO2 Open Banking Business Intelligence Dashboad services, to expose them outside of the Kubernetes cluster. It is expected to use the NGINX Ingress Controller [1], for this purpose. [1]: NGINX Ingress Controller maintained by Kubernetes Fixed with #26
gharchive/issue
2019-11-04T06:39:24
2025-04-01T06:46:16.318808
{ "authors": [ "kaumini" ], "repo": "wso2/kubernetes-open-banking", "url": "https://github.com/wso2/kubernetes-open-banking/issues/18", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2431732689
The property "Reason" in Fault mediator has to be required. Description: Currently in property Reason is not required in Fault mediator. But creating a Fault mediator without the Reason property cause below error when running the project. [2024-07-26 14:10:45,925] ERROR {ProxyServiceDeployer} - ProxyService Deployment from the file : /Users/chathurangaj/Downloads/wso2mi-4.3.0/tmp/carbonapps/-1234/1721983245891FaultMediatorExample_1.0.0.car/test_1.0.0/test-1.0.0.xml : Failed. org.apache.synapse.SynapseException: The fault reason is a required attribute for the makefault mediator unless it is a pox fault Suggested Labels: Suggested Assignees: Affected Product Version: v1.0.0 OS, DB, other environment details and versions: Steps to reproduce: Create a Fault mediator without Reason field. Run the project. Related Issues: The latest release v1.0.1 has resolved the issue. Please reopen if the problem persists.
gharchive/issue
2024-07-26T08:42:19
2025-04-01T06:46:16.322747
{ "authors": [ "chathuranga-jayanath-99", "thuva9872" ], "repo": "wso2/mi-vscode", "url": "https://github.com/wso2/mi-vscode/issues/230", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1032166490
A new resource to Gateway REST API as a health check API to verify whether all the APIs are deployed or not Describe your problem(s) With the new Artifact Synchronization architecture since WSO2 APIM 3.2.0, APIs will be deployed from a storage instead of the file system. If there are several APIs waiting to be deployed at the startup, it will take some time to get all the APIs deployed in the gateway. There is no proper way to verify whether all the APIs are deployed in the gateway or not before routing the traffic to the gateway. Describe your solution A new resource to the Gateway Rest API, which will return the status of API deployment in the gateway. If all the APIs are deployed in the gateway, this resource should send 200 OK, and if there are APIs that are missing or waiting to be deployed, the resource should return the list of these APIS. After having a discussion with @tharindu1st , we will have it as just a health check endpoint. This will return 200 OK if all the apis are deployed and 403, if at least one API is not deployed. it will return 403. We can use the /apis resource of the Gateway REST API to get the list of APIs deployed in the gateway. https://apim.docs.wso2.com/en/latest/install-and-setup/setup/deployment-best-practices/basic-health-checks/
gharchive/issue
2021-10-21T07:43:59
2025-04-01T06:46:16.326277
{ "authors": [ "1akshitha", "Krishanx92", "lakshithagunasekara" ], "repo": "wso2/product-apim", "url": "https://github.com/wso2/product-apim/issues/11884", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
448755851
APIAuthenticationHandler throws java.lang.NumberFormatException when DEBUG enabled in a distributed setting Description: I have set up a distributed deployment of WSO2 APIM 2.1.0 in my local machine running each profile in separate ports. I have enabled debug for org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler class from log4.properties. When I create a simple API and tries to invoke it, it throws the following exception, java.lang.NumberFormatException: null at java.lang.Long.parseLong(Long.java:552) at java.lang.Long.parseLong(Long.java:631) at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.logMessageDetails(APIAuthenticationHandler.java:268) at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest(APIAuthenticationHandler.java:123) at org.apache.synapse.rest.API.process(API.java:325) at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:90) at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:69) at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:304) at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:78) at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180) at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:325) at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:371) at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:151) at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:748) This occurs inside APIAuthenticationHandler's handleRequest method. when debug is enabled it tries to execute logMessageDetails method. Inside the logMessageDetails method, the exception throws due to parsing of null object returned from Axis2MessageContext as "wso2statistics.request.received.time". However it works fine and no exception is thrown when we run the single pack of api manager. My requirement is to enable existing debug log for APIAuthenticationHandler class to log the elapsed time to authentication response. Since it throws this exception in debug mode, I am unable to put any new debug logs or use existing debug logs within the class for the purpose. By @dushaniw from https://wso2.org/jira/browse/APIMANAGER-6062 null check is available in [1]. [1] https://github.com/wso2/carbon-apimgt/blob/3463a1ee657bdbd078b34ebf35ad76df0aad85be/components/apimgt/org.wso2.carbon.apimgt.gateway/src/main/java/org/wso2/carbon/apimgt/gateway/handlers/security/APIAuthenticationHandler.java#L618
gharchive/issue
2019-05-27T09:27:03
2025-04-01T06:46:16.330630
{ "authors": [ "isharac", "tgtshanika" ], "repo": "wso2/product-apim", "url": "https://github.com/wso2/product-apim/issues/4863", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
538212525
After adding a documentation "file" type, Publisher does not redirect to documentation list Create an API and add a documentation type "File" After successfully adding the doc, instead of redirecting to the documentation list, Publisher shows the following UI. Fixed with https://github.com/wso2/carbon-apimgt/pull/7899 in 3.1.0 @CrowleyRajapakse FYI
gharchive/issue
2019-12-16T06:30:34
2025-04-01T06:46:16.333054
{ "authors": [ "dushaniw", "kavishkafernando" ], "repo": "wso2/product-apim", "url": "https://github.com/wso2/product-apim/issues/7155", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
941932960
SAML Session index/assertion query gives success even after logout SAML Sessinindex/assertion query gives success even after logout: Wso2 identity server SAML session index/assertion query woeks even after logout. Because even after logout or session timeout the session index or the record which persisted in the DB won't delete. So, every time if we query for session existance it will check for session as even after session expiry/logout also record is present every time it will give sucess. How to reproduce: I have followed the following ink for configuring and querying SAML session existance https://docs.wso2.com/display/IS530/Querying+SAML+Assertions Expected behavior: In database even after logout session will be available Environment information (Please complete the following information; remove any unnecessary fields) : Product Version: IS 5.8.0, IS 5.10.0 OS: Ubuntu16 Database: H2, MSSQL Userstore: LDAP This issue is being closed due to extended inactivity. Please feel free to reopen it if further attention is needed. Thank you for helping us keep the issue list relevant and focused!
gharchive/issue
2021-07-12T10:56:29
2025-04-01T06:46:16.350021
{ "authors": [ "isharak", "rahul123-dotcom" ], "repo": "wso2/product-is", "url": "https://github.com/wso2/product-is/issues/12090", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1401969031
[Testing] Guides on verification for updated user attributes The two guides related to enabling verification for updated user attributes [1] should be tested. [1] https://is.docs.wso2.com/en/latest/guides/identity-lifecycles/enable-email-account-verification-for-an-updated-email-address/ This issue is being closed due to extended inactivity. Please feel free to reopen it if further attention is needed. Thank you for helping us keep the issue list relevant and focused!
gharchive/issue
2022-10-08T14:46:36
2025-04-01T06:46:16.351731
{ "authors": [ "isharak", "nilminiwso2" ], "repo": "wso2/product-is", "url": "https://github.com/wso2/product-is/issues/15014", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1982080353
Assigned roles of the group need to show role's audience Describe the issue: Create roles with the same name in different audiences (eg: role1 in organization audience, role1 in different application audience) Create a group Assign the group to roles with the same name in different audiences. Group-to-role assignments will be successful Go to the Group's assigned roles view Only the assigned roles names are shown But the role audience is also required to distinguish the assigned roles. (API returns the audience data) https://github.com/wso2/product-is/assets/25483865/29dcf047-af25-40d1-8f6d-7a474224c6f6 There is also a backend issue. In the new authorization runtime, when retrieving a group's roles, roles with similar names are assigned the same audience (You can see this in the screen shot). Fix for backend issue : https://github.com/wso2-extensions/identity-inbound-provisioning-scim2/pull/509
gharchive/issue
2023-11-07T19:31:55
2025-04-01T06:46:16.355352
{ "authors": [ "AnuradhaSK", "shashimalcse" ], "repo": "wso2/product-is", "url": "https://github.com/wso2/product-is/issues/17652", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2043316239
[Custom Connector] [SAML] : HTTP Binding is allowed to save empty Describe the issue: When creating a SAML authenticator in the custom connector, we are allowed to save the authenticator without the HTTP Protocol binding field since it is non mandatory How to reproduce: Create a custom connector Click on the Settings Navigate to create a SAML authenticator and verify the HTTP Protocol Binding type Expected behavior: The field should be mandatory with a default selected protocol binding type Environment information (Please complete the following information; remove any unnecessary fields) : Product Version: [e.g., IS 5.10.0, IS 5.9.0] OS: [e.g., Windows, Linux, Mac] Database: [e.g., MySQL, H2] Userstore: [e.g., LDAP, JDBC] Optional Fields Related issues: Suggested labels: Default value issue in the dropdown is fixed with https://github.com/wso2/identity-apps/pull/5132 PR. However, if we need to make the HTTP Binding as a required field we need to update the meta data coming from the backend. As there is a default values for the HTTP binding field. Therefore no need to make it mandatory. Hence closing the issue.
gharchive/issue
2023-12-15T09:44:13
2025-04-01T06:46:16.361398
{ "authors": [ "DilshanSenarath", "GoushiRam", "Thisara-Welmilla" ], "repo": "wso2/product-is", "url": "https://github.com/wso2/product-is/issues/18623", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }