id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1031178743
是否有方法更新combo的大小(不是固定大小fixSize) [ ] I have searched the issues of this repository and believe that this is not a duplicate. What problem does this feature solve? combo在折叠后,如果移动了combo位置,再展开后,combo的大小会变成默认的大小,而不是通过combo内节点的位置计算的大小,需要用户手动移动combo中的node,宽高才会更新。 What does the proposed API look like? 能够让开发者手动调用一个函数来实现combo的大小更新呢? 目前已试过combo.refresh()是达不到我想要的效果。 用: combo.update({size:[width,height]})解决了 我也遇见过这个问题,收起的时候可以控制大小,但是展开的时候有时候不会包裹住内部的节点,不知道楼主有没有遇见过这个问题
gharchive/issue
2021-10-20T09:24:38
2025-04-01T04:33:26.082706
{ "authors": [ "user-lihai", "yangmiao13160" ], "repo": "antvis/G6", "url": "https://github.com/antvis/G6/issues/3263", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2059138955
item.update无法更新type G6 Version 4.x Describe the bug `const model = { id: 'node', type: 'rect', label: 'node', style: { fill: 'red', }, }; item.update(model); item.refresh(); ` 原来的圆形节点未变成矩形 Your Example Website or App 无 Steps to Reproduce the Bug or Issue 无 Expected behavior 无 Screenshots or Videos No response Platform OS: [e.g. macOS, Windows, Linux] Browser: [e.g. Chrome, Safari, Firefox] Version: [e.g. 91.1] Additional context No response 已解决,4.8.23版本有此问题,将版本后正常
gharchive/issue
2023-12-29T07:03:55
2025-04-01T04:33:26.086955
{ "authors": [ "HogoZhang" ], "repo": "antvis/G6", "url": "https://github.com/antvis/G6/issues/5308", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1247335577
【仕様変更提案】tagsがなくても投稿できるようにしたい 現在、以下のように if (!tags) { console.log( '\n' + emoji.get('disappointed') + ' 選択した記事にタグが設定されていません.\n記事を投稿するには一つ以上タグが設定されている必要があります.\n' ); return -1; } tags が設定されていないとエラーが出力されて投稿することができません。 記事を投稿する場合にはタグを設定しない場合もあるので、tagsが入っていなければエラーにして落とすのではなく空の配列にして投稿できるようにする形にした方がいいのかなと思いました。 APIの仕様上できそうになかったのでこのissueはclose とします
gharchive/issue
2022-05-25T01:53:03
2025-04-01T04:33:26.089919
{ "authors": [ "TakuKobayashi" ], "repo": "antyuntyuntyun/qiita-cli", "url": "https://github.com/antyuntyuntyun/qiita-cli/issues/44", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2613717308
Update the interface of About us section , making it more responsive and good looking. I will update the interface of about us section. about us page is ok . but still if u have a better UI design for the page , then raise pr . check the pages in profile dropdown , they need major improvement . their footer and navbar is random . u can redesign the pages to match with the theme ( colors ) of the website .
gharchive/issue
2024-10-25T10:30:59
2025-04-01T04:33:26.127801
{ "authors": [ "abhi03ruchi", "prakharpande04" ], "repo": "anuragverma108/SwapReads", "url": "https://github.com/anuragverma108/SwapReads/issues/3994", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2353402835
Added the author filter for book recommendation Related Issue [Cite any related issue(s) this pull request addresses. If none, simply state “None”] Fixes: #1634 Description [Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.] Added a new filter of Author for the book recommendation system. If the user enters only genre then it will recommend books of that genre. If the user enter genre and author then it will recommend books of that author in that specified genre. If such book is not present in the system then it will notify that book is not found. Type of PR [ ] Bug fix [X] Feature enhancement [ ] Documentation update [ ] Other (specify): _______________ Screenshots / videos (if applicable) [Attach any relevant screenshots or videos demonstrating the changes. Make sure to attach before & after screenshots in your PR.] Books recommended on basis of genre ONLY. Books recommended based on genre and author BOTH. Even if the entire author name is not added it still suggests appropriate book. NO BOOK IS FOUND for that genre and author Checklist: [X] I have made this change from my own. [X] I have taken help from some online resources. [X] My code follows the style guidelines of this project. [X] I have performed a self-review of my own code. [X] I have commented my code, particularly in hard-to-understand areas. [X] I have made corresponding changes to the documentation. [X] My changes generate no new warnings. [X] I have tested the changes thoroughly before submitting this pull request. [X] I have provided relevant issue numbers and screenshots after making the changes. Conflicting files assets/html/book_recommend.html Please see the conflicts highlighted in the image above. Resolve the conflicts in the mentioned files. Ensure you pull all the latest changes before pushing your feature/bug fixes to the main branch. After resolving the conflicts, please open a new PR for review. Thank you!
gharchive/pull-request
2024-06-14T13:34:52
2025-04-01T04:33:26.136000
{ "authors": [ "Archiesachin", "anuragverma108" ], "repo": "anuragverma108/SwapReads", "url": "https://github.com/anuragverma108/SwapReads/pull/1852", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2606135794
Extended FAQs Section for Improved Buyer Communication Related Issue [Cite any related issue(s) this pull request addresses. If none, simply state “None”] Fixes: #3808 Description [Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.] Type of PR [ ] Bug fix [x] Feature enhancement [ ] Documentation update [ ] Other (specify): _______________ Screenshots / videos (if applicable) [Attach any relevant screenshots or videos demonstrating the changes. Make sure to attach before & after screenshots in your PR.] Checklist: [x] I have made this change from my own. [x] My code follows the style guidelines of this project. [x] I have performed a self-review of my own code. [x] My changes generate no new warnings. [x] I have tested the changes thoroughly before submitting this pull request. [x] I have provided relevant issue numbers and screenshots after making the changes. @atmajaa could you please merge this???
gharchive/pull-request
2024-10-22T18:06:47
2025-04-01T04:33:26.141321
{ "authors": [ "sriharsha0x1" ], "repo": "anuragverma108/SwapReads", "url": "https://github.com/anuragverma108/SwapReads/pull/3889", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2657027187
Tony/21_chrome_support Pull Request Details What issue does this PR address Fixed the issue that video cannot be played on Chrome. Link to associated issue: #21 Checklist Complete the check-list below to ensure your branch is ready for PR. Flutter Style Guide: https://survivor.togaware.com/gnulinux/flutter-style.html [x] Screenshots included in linked issue [ ] Changes adhere to the team style and coding guideline [x] No confidential information [x] No duplicated content [x] No lint check errors related to your changes (make prep or flutter analyze lib) [ ] Pre-exisiting lint errors noted: [HERE] [ ] Tested on at least one device [ ] Android Phone [ ] Android Emulator [ ] Chrome on Android [x] Chrome [ ] iOS [ ] Linux [x] MacOS [ ] Windows [x] Added 2 reviewers (or 1 for private repositories then they add another) Finalising Once PR discussion is complete and 2 reviewers have approved: [ ] Merge dev into the branch [ ] Resolve any conflicts [ ] Add one line summary into CHANGELOG.md [ ] Bump appropriate version number in pubspec.yaml [ ] Push to git repository and review [ ] Merge PR into dev I am getting a black window on Linux. When I click Play the audio is heard but no video. I am getting a black window on Linux. When I click Play the audio is heard but no video. Thanks @gjwgit. I'll check it on Linux. If my black screen on chrome is not seen on other OS then please merge this PR and have the Linux issue as a separate issue for resolution later on. If my black screen on chrome is not seen on other OS then please merge this PR and have the Linux issue as a separate issue for resolution later on. Hi @gjwgit, I will fix it in #30. This issue will be merged first.
gharchive/pull-request
2024-11-13T22:40:05
2025-04-01T04:33:26.152743
{ "authors": [ "gjwgit", "tonypioneer" ], "repo": "anusii/markdown_widgets", "url": "https://github.com/anusii/markdown_widgets/pull/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2284435674
SOLID POD: Secure Key request when entered and press ENTER it should be same as OK User Story After I have typed the key here, as a user I am inclined to press ENTER. This should be the same a tapping the OK button. Implemented in pull request https://github.com/anusii/solidpod/pull/164 To test, go to the keypod demo page and Click "Forget Security Key" Turn on the "Encrypt Data" switch Click "Key Value Table Demo" Add a key-value pair and click "Submit" In the security key input widget, type your security key and press the enter key. Known issue: If you move the input focus off the text input after typing your security key, then press the enter key seems to do nothing (haven't been able to find a fix for this so far). Tested and works as expected
gharchive/issue
2024-05-08T00:18:37
2025-04-01T04:33:26.157253
{ "authors": [ "cdawei", "gjwgit" ], "repo": "anusii/solidpod", "url": "https://github.com/anusii/solidpod/issues/129", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1566244375
Non numpy vectorization doc Being paranoidly explicit about appropriate format of the out object expected from the _evaluate method, per discussion in #372 You may want to do the merge with fastforward/squash, which apparently I can't request on my side. Thanks again! Thanks for putting the effort into improving the documentation. I used squash before merging indeed. The paragraph will be available once I release a new version. You are most welcome, that was a tiny contribution back to into an excellent project which is giving a good help for my research!
gharchive/pull-request
2023-02-01T14:45:26
2025-04-01T04:33:26.159289
{ "authors": [ "blankjul", "davide-q" ], "repo": "anyoptimization/pymoo", "url": "https://github.com/anyoptimization/pymoo/pull/375", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2554014912
[maketmpl] make template builder it builds the zip file, the ray-app.json metadata file, and various flavors of readme markdown files.. cc @vigneshka let me know if this PR is too long. I can break it into smaller ones.
gharchive/pull-request
2024-09-28T05:11:54
2025-04-01T04:33:26.160655
{ "authors": [ "aslonnie" ], "repo": "anyscale/templates", "url": "https://github.com/anyscale/templates/pull/367", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
453770781
Lab 3 Fixes #295 Review of colleague's PR # Changes proposed in this PR: Complete Lab 3 Add letters function Add sortLetters function Add Unit Tests for 100% test coverage Codecov Report Merging #296 into master will not change coverage. The diff coverage is 100%. @@ Coverage Diff @@ ## master #296 +/- ## ===================================== Coverage 100% 100% ===================================== Files 18 19 +1 Lines 303 322 +19 ===================================== + Hits 303 322 +19 Impacted Files Coverage Δ 03_letters/joshcarp/letters.go 100% <100%> (ø) Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 15b200b...dd0e191. Read the comment docs. The go-course is now closed. Thank you very much for participating.
gharchive/pull-request
2019-06-08T08:27:51
2025-04-01T04:33:26.168686
{ "authors": [ "codecov-io", "joshcarp", "juliaogris" ], "repo": "anz-bank/go-course", "url": "https://github.com/anz-bank/go-course/pull/296", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
204317347
ViewControllers not loaded if addViewController not at init Hi sir, firstly thank you for your great lib :) I dynamically load the tabs based on server data. So i call http first to get my tab list. Then i construct using [self addViewController:]; It works, i can see the tab bar and the titles. But the view controllers is not showing (viewDidLoad is not called). only after i drag to slide (even small drag) then the viewDidLoad triggered and then i can see the content. Any idea how to solve this? Thank you Hi sletith, I have reproduced the bug you mentioned above. there are two problems I found. I have pushed a new branch [hotfix] trying to fix it. You can try the new code. if the update can solve your problem, I will release a new version. if it can't, I think you'd better show me some code then I can locate the problem more clearly . Thank you. thanks sir. it works ^^
gharchive/issue
2017-01-31T14:42:21
2025-04-01T04:33:26.187373
{ "authors": [ "aokizen", "sleith" ], "repo": "aokizen/SPSlideTabBarController", "url": "https://github.com/aokizen/SPSlideTabBarController/issues/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
714188901
Bugfix: Use upstream instead of target branch I'm pulling from upstream/master to a branch with a different name (ansible-st2-tracking). That branch is not in the upstream repo, so this gave the error: New commits being pulled: fatal: ambiguous argument 'upstream/ansible-st2-tracking': unknown revision or path not in the working tree. Use '--' to separate paths from revisions, like this: 'git <command> [<revision>...] -- [<file>...]' Yep, thanks. :)
gharchive/pull-request
2020-10-03T22:35:39
2025-04-01T04:33:26.204177
{ "authors": [ "aormsby", "cognifloyd" ], "repo": "aormsby/Fork-Sync-With-Upstream-action", "url": "https://github.com/aormsby/Fork-Sync-With-Upstream-action/pull/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
679199745
Added sprig support for Go template Integrated Sprig shall widen the usage of add command. Tested on 1GB+ CSV with adequate complex usage of add command and Sprig function with decent performance. I believe this integration shall improve the usage of gocsv. I have also updated the README for better usage. @aotimme - I mistakenly pushed the gocsv and gocsv.docx, therefore I closed this PR. I have raised another one, https://github.com/aotimme/gocsv/pull/28
gharchive/pull-request
2020-08-14T14:35:34
2025-04-01T04:33:26.207190
{ "authors": [ "munim" ], "repo": "aotimme/gocsv", "url": "https://github.com/aotimme/gocsv/pull/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
85899975
aozorahack用のアイコン aozorahack のアイコンが欲しいです! slackのものと同一で問題ないでしょうか? slackで使ってるアイコンの青空写真はパブリックドメインのものなのでご自由にどうぞ!ATNDのトップと同じです。 2015年6月7日日曜日、genneinotifications@github.comさんは書きました: aozorahack のアイコンが欲しいです! slackのものと同一で問題ないでしょうか? — Reply to this email directly or view it on GitHub https://github.com/aozorahack/aozorahack/issues/19. -- iPhoneより送信 :bow:
gharchive/issue
2015-06-07T12:59:12
2025-04-01T04:33:26.217719
{ "authors": [ "gennei", "kskktk" ], "repo": "aozorahack/aozorahack", "url": "https://github.com/aozorahack/aozorahack/issues/19", "license": "unlicense", "license_type": "permissive", "license_source": "bigquery" }
233640155
previous doesn't work with :e buffer This is probably user error, and not a bug, but when I have buffers open, :bprev and :bnext work as expected, but when I do :e to get a file list, the file list becomes a buffer that I can not :prev. i.e. I can cycle past it using next, but previous doesn't work. Also, the "tab" at the top isn't highlighted when I'm viewing the ":e" buffer. Any ideas on how I'm supposed to be handling this situation? Thanks. That’s due to how the Netrw plugin works (which is what Vim uses to handle directory listings), unfortunately. In short, it puts its directory listing in a nameless hidden buffer. This is what ultimately drove me to write the readdir plugin. Trying to do it differently from Netrw then taught me why Netrw is written as it is – both approaches have their problems. I’m not sure what to do here really. I can’t tell everyone to install Readdir just to fix this, nor should they have to, nor would I recommend it to everyone any more. And in the end the UX is suboptimal either way. Kinda crappy situation. Ok, thanks for the information about the issue. I can live with it, and I'll try readdir to see if that works for me.
gharchive/issue
2017-06-05T16:23:03
2025-04-01T04:33:26.220267
{ "authors": [ "ap", "equant" ], "repo": "ap/vim-buftabline", "url": "https://github.com/ap/vim-buftabline/issues/31", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1205978926
On Singleplayer it says "You can not have a origin" I've seen people with this issue but with them it happens on servers, mine is happening on singleplayer-- It didn't give me a prompt to pick origin and I cant even use the orb or commands because it comes up with this. I own an SMP where we use origins and we're updating to 1.18.2 soon and we're stumped as to why this is happening. The jar file is labelled "Origins-1.18.2-1.4.1" if that helps. UPDATE: There is an incompatibility with Charm, is there a way for this patched so we can have both? Thanks for figuring out which mod causes the incompatibility. I'm currently out of town, so can't work on this, but I'll see what I can do when I'm back next week. I experienced the same bug today and was able to trace it to an incompatibility with DeepslateCutting (v1.3.2). UPDATE: There is an incompatibility with Charm, is there a way for this patched so we can have both? Also experience this with Charm, disabled Charm, works perfectly. Do any of you experiencing this bug have the mod OK Zoomer installed by any chance? Do any of you experiencing this bug have the mod OK Zoomer installed by any chance? I have both mods active and have had no problems. I have both mods active and have had no problems. By both mods you mean OK Zoomer and DeepslateCutting? If you haven't had any problems, why have you reported experiencing this bug? Or did you mean Charm and OK Zoomer aren't causing problems for you, and they only happen once you had DeepslateCutting? By both mods you mean OK Zoomer and DeepslateCutting? No, I meant OK Zoomer and Origins. Sorry about not being clear. For me it is only Deepslate Cutting and Origins that are causing the error if both mods are active. Even without OK Zoomer active? My questions are: Can you run Origins and DeepslateCutting alone? Can you run Origins and OK Zoomer alone? You answered yes. Can you run Origins and DeepslateCutting and OK Zoomer together? Can you run Origins and DeepslateCutting alone? Can you run Origins and OK Zoomer alone? You answered yes. Can you run Origins and DeepslateCutting and OK Zoomer together? Did some testing on a fresh instance with newly downloaded mods. Here are my results of testing: Origins + DeepslateCutting -> Origins selectable Origins + OK Zoomer -> Origins selectable Origins + DeepslateCutting + OK Zoomer -> Origins not selectable So it seems I did not test the first configuration before and must have thought the problem being with DeepslateCutting alone and not in combination with another mod. Sorry about that. Mod versions used: deepslatecutting-1.18.2-1.3.2 fabric-api-0.51.1+1.18.2 okzoomer-5.0.0-beta.6+1.18.2 Origins-1.18.2-1.4.1 Fabric Loader 0.13.3 (OK Zoomer causes crash on startup with Fabric Loader 0.14.0) Having this problem with 1.20.1 and none of the aforementioned mods
gharchive/issue
2022-04-16T01:41:31
2025-04-01T04:33:26.230513
{ "authors": [ "Wahngrok", "Whocraf", "apace100", "skxrlxt", "wiryfuture" ], "repo": "apace100/origins-fabric", "url": "https://github.com/apace100/origins-fabric/issues/523", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
239266879
Dynamic Allocation: Too many executors than desired? On running a Spark Pi with dynamic allocation enabled, I see the following somewhat consistently.. When I run the job with these arguments bin/spark-submit --deploy-mode cluster --class org.apache.spark.examples.SparkPi --master k8s://https://192.168.6.154:6443 --kubernetes-namespace default --conf spark.dynamicAllocation.enabled=true --conf spark.shuffle.service.enabled=true --conf spark.kubernetes.shuffle.namespace=default --conf spark.kubernetes.shuffle.labels=app=spark-shuffle-service,spark-version=2.1.0 --conf spark.app.name=spark-pi --conf spark.kubernetes.driver.docker.image=docker:5000/spark-driver:varun_latest31 --conf spark.kubernetes.executor.docker.image=docker:5000/spark-executor:varun_latest31 --conf spark.kubernetes.initcontainer.docker.image=docker:5000/spark-init-container:varun_latest31 local:///opt/spark/examples/jars/spark-examples_2.11-2.1.0-k8s-0.2.0-SNAPSHOT.jar 61000 The number of pods launched seem to be a much higher number than what is reported in the logs.. .. spark-pi-1498671730530-driver 1/1 Running 0 2m spark-pi-1498671730530-exec-1 1/1 Running 0 2m spark-pi-1498671730530-exec-10 1/1 Running 0 1m spark-pi-1498671730530-exec-11 1/1 Running 0 1m spark-pi-1498671730530-exec-12 1/1 Running 0 1m spark-pi-1498671730530-exec-13 1/1 Running 0 1m spark-pi-1498671730530-exec-14 1/1 Running 0 1m spark-pi-1498671730530-exec-15 1/1 Running 0 1m spark-pi-1498671730530-exec-16 1/1 Running 0 1m spark-pi-1498671730530-exec-17 1/1 Running 0 59s spark-pi-1498671730530-exec-18 1/1 Running 0 59s spark-pi-1498671730530-exec-19 1/1 Running 0 59s spark-pi-1498671730530-exec-2 1/1 Running 0 2m spark-pi-1498671730530-exec-20 1/1 Running 0 59s spark-pi-1498671730530-exec-21 1/1 Running 0 59s spark-pi-1498671730530-exec-22 0/1 Pending 0 38s spark-pi-1498671730530-exec-23 0/1 Pending 0 38s spark-pi-1498671730530-exec-24 0/1 Pending 0 38s spark-pi-1498671730530-exec-25 0/1 Pending 0 38s spark-pi-1498671730530-exec-26 0/1 Pending 0 38s spark-pi-1498671730530-exec-3 1/1 Running 0 2m spark-pi-1498671730530-exec-4 1/1 Running 0 2m spark-pi-1498671730530-exec-5 1/1 Running 0 2m spark-pi-1498671730530-exec-6 1/1 Running 0 2m spark-pi-1498671730530-exec-7 1/1 Running 0 1m spark-pi-1498671730530-exec-8 1/1 Running 0 1m spark-pi-1498671730530-exec-9 1/1 Running 0 1m .. In the driver logs I see the following, I see only upto 5 launched ... └─[0] <git:(master 29acb7e✱✈) > kubectl logs spark-pi-1498671730530-driver | grep -i backend 2017-06-28 17:42:30 INFO KubernetesClusterSchedulerBackend:54 - SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8 2017-06-28 17:42:35 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 1 2017-06-28 17:42:46 INFO KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint:54 - Registered executor NettyRpcEndpointRef(null) (10.38.0.3:51308) with ID 1 2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 2 2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 3 2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 4 2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 5 ... I tried using the published images from: https://apache-spark-on-k8s.github.io/userdocs/running-on-kubernetes.html My logs say: ~/g/src k logs spark-pi-1498677689366-driver | grep 'Requesting a new executor, total executors' 2017-06-28 19:21:52 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 1 2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 2 2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 3 2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 4 2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 5 2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 6 2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 7 2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 8 2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 9 2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 10 2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 11 2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 12 2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 13 2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 14 2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 15 2017-06-28 19:22:21 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 16 2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 17 2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 18 2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 19 2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 20 2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 21 2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 22 2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 23 2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 24 2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 25 2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 26 2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 27 2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 28 2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 29 2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 30 2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 31 2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 32 2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 33 2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 34 2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 35 2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 36 2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 37 2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 38 2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 39 2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 40 2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 41 2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 42 2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 43 2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 44 2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 45 2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 46 Can you try with the published images? Are your images built from HEAD? These are built from the HEAD...I will check once, if there is some user/operator error on my part and report.. @varunkatta can you try with static allocation? This might be a bug with our images in general - this wouldn't look specific to dynamic allocation, we build the pod's arguments identically. And also use kubectl describe pod to get the details of the executor pods. I double checked our Dockerfiles on HEAD and they seem sane. Here's the command: CMD SPARK_CLASSPATH="${SPARK_HOME}/jars/*" && \ if ! [ -z ${SPARK_MOUNTED_CLASSPATH}+x} ]; then SPARK_CLASSPATH="$SPARK_MOUNTED_CLASSPATH:$SPARK_CLASSPATH"; fi && \ if ! [ -z ${SPARK_EXECUTOR_EXTRA_CLASSPATH+x} ]; then SPARK_CLASSPATH="$SPARK_EXECUTOR_EXTRA_CLASSPATH:$SPARK_CLASSPATH"; fi && \ if ! [ -z ${SPARK_MOUNTED_FILES_DIR} ]; then cp -R "$SPARK_MOUNTED_FILES_DIR/." .; fi && \ exec /sbin/tini -- ${JAVA_HOME}/bin/java -Dspark.executor.port=$SPARK_EXECUTOR_PORT -Xms$SPARK_EXECUTOR_MEMORY -Xmx$SPARK_EXECUTOR_MEMORY -cp $SPARK_CLASSPATH org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url $SPARK_DRIVER_URL --executor-id $SPARK_EXECUTOR_ID --cores $SPARK_EXECUTOR_CORES --app-id $SPARK_APPLICATION_ID --hostname $SPARK_EXECUTOR_POD_IP So we're probably seeing something where SPARK_EXECUTOR_MEMORY isn't being set properly, but looking at KubernetesClusterSchedulerBackend I can't quite see how this would happen. Would be good to inspect the pod to get more insight. on describe pod of executor, this is what I see. I will try digging more information, if possible. ..... Volume Mounts: /tmp from tmp (rw) /var/run/secrets/kubernetes.io/serviceaccount from default-token-2jnd3 (ro) Environment Variables: SPARK_EXECUTOR_PORT: 10000 SPARK_DRIVER_URL: spark://CoarseGrainedScheduler@10.45.128.3:46339 SPARK_EXECUTOR_CORES: 1 SPARK_EXECUTOR_MEMORY: 2g SPARK_APPLICATION_ID: spark-5fefae5c99a04ca4b40249e078a653a1 SPARK_EXECUTOR_ID: 1 SPARK_MOUNTED_CLASSPATH: /var/spark-data/spark-jars/* SPARK_EXECUTOR_POD_IP: (v1:status.podIP) K8s versions: Client Version: version.Info{Major:"1", Minor:"4", GitVersion:"v1.4.6+e569a27", GitCommit:"e569a27d02001e343cb68086bc06d47804f62af6", GitTreeState:"not a git tree", BuildDate:"2016-11-12T09:29:14Z", GoVersion:"go1.7.3", Compiler:"gc", Platform:"darwin/amd64"} Server Version: version.Info{Major:"1", Minor:"5", GitVersion:"v1.5.4", GitCommit:"7243c69eb523aa4377bce883e7c0dd76b84709a1", GitTreeState:"clean", BuildDate:"2017-03-07T23:34:32Z", GoVersion:"go1.7.4", Compiler:"gc", Platform:"linux/amd64"} This issue is stale now.
gharchive/issue
2017-06-28T19:16:07
2025-04-01T04:33:26.242472
{ "authors": [ "foxish", "mccheah", "varunkatta" ], "repo": "apache-spark-on-k8s/spark", "url": "https://github.com/apache-spark-on-k8s/spark/issues/360", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
263282421
Delete Kubernetes resources when the client waits for and sees app completion What changes were proposed in this pull request? This PR fixes #519 for the case where the submission client waits for the submitted application to finish. Upon completion of the application, the submission client deletes all Kubernetes resources created for the application to run. rerun integration tests please I want to re-iterate on this issue/PR. If we have concern around losing some objects like the ConfigMap for setting up the init-container, as I said above, we could log information stored in it for debugging purpose. This, IMO, is better than making the ConfigMap stick around just for debugging. Thoughts? @mccheah @foxish Any thoughts on this ? It would be good to cleanup resources after completion. In a normal scenario, this is filling up a lot of services in completed state for example. Thanks @felixcheung for jumping on this :) hey where are we on this? and how about going upstream? @felixcheung Yes, I think we should go upstream. I created https://issues.apache.org/jira/browse/SPARK-23571. Sorry, didn't see this before. Same comment as in https://github.com/apache/spark/pull/20722#discussion_r171968410. Why not do this during driver shutdown - that way, 1) we lose the driver and k8s garbage collection cleans up 2) driver terminates and we clean up executors as well as auxiliary resources like configmaps etc. I want to re-iterate on this issue/PR. If we have concern around losing some objects like the ConfigMap for setting up the init-container, as I said above, we could log information stored in it for debugging purpose. This, IMO, is better than making the ConfigMap stick around just for debugging. Thoughts? I agree. We can dump all k8s objects. My hunch is that it's not that useful, given it's a pretty deeply buried implementation detail. As discussed in https://github.com/apache/spark/pull/20722, we think the right solution is move resource management into the driver pod. This way, cleanup of auxiliary resources upon completion is guaranteed regardless of which deployment mode is used and whether the client waits for application to complete or not.
gharchive/pull-request
2017-10-05T22:11:13
2025-04-01T04:33:26.248502
{ "authors": [ "dharmeshkakadia", "felixcheung", "foxish", "liyinan926" ], "repo": "apache-spark-on-k8s/spark", "url": "https://github.com/apache-spark-on-k8s/spark/pull/520", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
805347634
feat(legacy-preset-chart-nvd3): subject NVD3 Bar chart sort by Add control for sort by "asc" or "desc" Associated with: https://github.com/apache/superset/pull/13049 AFTER https://user-images.githubusercontent.com/8277264/107492589-50831880-6b95-11eb-8042-c94c3b21500e.mp4 Does this re-order bars as well? I'm worried there may be confusion if it doesn't. Does this re-order bars as well? I'm worried there may be confusion if it doesn't. @ktmud Actually the sort is working with a row limit, for example, you have a limit of 50, and sort descendent can be different values Yes, I understand how it works, but users might expect it to sort the x-axis. How about we change "Sort by" to "Sort rows by"? cc @junlincc I think we should default to no ordering, and only opt-in to sorting, both ascending and descending. Perhaps have a "sort bars" control and another "sort descending" one? in this case, I agree we should leave it to "no default" since it ties to row limit. We also discussed about having a 'switch' for ascending and descending. @mihir174 How about we change "Sort by" to "Sort rows by"? I see your point @ktmud Introducing a new label to one specific chart might create more design debt. let's add tooltip "Sort rows by. changing row limit may affect sorting resulting" something like that? tag it for design-revisit
gharchive/pull-request
2021-02-10T09:44:33
2025-04-01T04:33:26.255455
{ "authors": [ "junlincc", "ktmud", "maloun96" ], "repo": "apache-superset/superset-ui", "url": "https://github.com/apache-superset/superset-ui/pull/947", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
562213768
ARTEMIS-2617 Improve AMQP Journal loading I've opened this PR for discussion. Currently it brings 4 improvements: enable G1GC string deduplication uses CoreMessageObjectPools on journal loading to save allocations allows AMQP lazy message data scanning on journal loading half-brute force search x-opt-delivery-time to avoid message data scanning @tabish121 @gemmellr @clebertsuconic @michaelandrepearce This PR contains many changes and I know there are parts that could be simplified a lot. The most important bits that can really be a game-changer for AMQP on journal loading are on https://github.com/apache/activemq-artemis/pull/2975/commits/fc77a546ca1ee047218e938f0413c55ed8c838c6. So please review me especially that part. An important note: I've found that AMQPMessagePersisterV2::decode does something weird re AMQPMessage::setAddress: AMQPMessagePersister::decode can AMQPMessage::setAddress, allocating AMQPMessage::extraProperties, but AMQPMessagePersisterV2::decode can decode a new extraProperties totally overwriting the existing one. Is it a bug or it's suppose to work like that? I see that this behaviour has been introduced on ARTEMIS-1858 in 1ae2784dc6075875b18780fa8ba40f86cb895f7b with this comment: /** * This will set the address on CoreMessage. * * Note for AMQPMessages: * in AMQPMessages this will not really change the address on the message. Instead it will add a property * on extraProperties which only transverse internally at the broker. * Whatever you change here it won't affect anything towards the received message. * * If you wish to change AMQPMessages address you will have to do it directly at the AMQP Message, however beware * that AMQPMessages are not supposed to be changed at the broker, so only do it if you know what you are doing. * @param address * @return */ Message setAddress(SimpleString address); I've tried to mitigate this by using CoreMessageObjectPools to save SimpleString allocations here and by allowing TypedProperties::decode to append/replace any existing extraProperty, saving an HashMap allocation, but IMO it could be addressed by properly fixing it, wdyt? any idea how? @michaelandrepearce I suppose you are interested in numbers: in my syntethic tests a 2 GB journal with very small messages has wasn't able to be loaded by a broker with 4GB heap size, always leading to a OOM: now it can load without relevant GC pauses eg We're not quite yet at the same level of Core messages, but very near: I think that the duplicate address is responsible of the difference here. The commit that allows to half-brute force search x-opt-delivery-time is what I'm not sure about: that method should be optimized and performed optimally by proton itself IMO (nudge nudge @tabish121 @gemmellr), but I would be super happy to provide a PR for it currently it can give rare false positive (causing message scanning when not necessary) I've yet to verify how it behave with many message annotations and longer messages If @michaelandrepearce has some real journal to test with and compare with master it would be great, given that this change can affect positively your usage of the broker ;) In addition, I hope I haven't missed to force message data scanning on getter of other properties eg routingType. @tabish121 I've tried to improve the fail-fast behaviour for both the search algorithm and the MessageAnnotations section search. The latter by using an IdentityHasMap to save checking the N possible message sections that would terminate the algorithm: let me know if that makes sense according to the AMQP spec :+1: I need this merged, as this will impact the changes I'm making for large messages. I see that you still need other work done, such as adding more tests.. on which case can you please add any further changes into a new PR? thanks a lot.
gharchive/pull-request
2020-02-09T17:53:51
2025-04-01T04:33:26.264663
{ "authors": [ "clebertsuconic", "franz1981" ], "repo": "apache/activemq-artemis", "url": "https://github.com/apache/activemq-artemis/pull/2975", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
897389870
ARTEMIS-3311 - ensure visibility of error state on operation context … …callback registration, fix and test did you run a whole test suite.. if you did it LGTM yep. full tests look good.
gharchive/pull-request
2021-05-20T20:13:06
2025-04-01T04:33:26.266401
{ "authors": [ "clebertsuconic", "gtully" ], "repo": "apache/activemq-artemis", "url": "https://github.com/apache/activemq-artemis/pull/3592", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
216221166
APEXCORE-676 Show description for DefaultProperties in get-app-packag… …e-info command only when user requests it by providing --with-description flag Output without --with-description flag Command : get-app-package-info {apa-file} "defaultProperties": { "dt.application..operator..attr.JVM_OPTIONS": "-Xmx128M", "dt.application..operator..attr.MEMORY_MB": "256" } Output with --with-description flag Command : get-app-package-info {apa-file} --with-description "defaultProperties": { "dt.application..operator..attr.JVM_OPTIONS": { "value": "-Xmx128M", "description": null } , "dt.application..operator..attr.MEMORY_MB": { "value": "256", "description": null } } Invalid argument get-app-package-info {apa-file} --with-descripti Output: com.datatorrent.stram.cli.ApexCli$CliException: Invalid parameter --with-descripti at com.datatorrent.stram.cli.ApexCli$GetAppPackageInfoCommand.execute(ApexCli.java:3482) at com.datatorrent.stram.cli.ApexCli$3.run(ApexCli.java:1519) @sandeshh @vrozov Kindly review @vrozov Incorporated your comments Jenkins build failed. Closing and reopening the request.
gharchive/pull-request
2017-03-22T21:58:48
2025-04-01T04:33:26.368227
{ "authors": [ "ajaygit158" ], "repo": "apache/apex-core", "url": "https://github.com/apache/apex-core/pull/493", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1527746035
feat: support disablestatus Type of change: [ ] Bugfix [ ] New feature provided [ ] Improve performance [ ] Backport patches What this PR does / why we need it: Pre-submission checklist: [ ] Did you explain what problem does this PR solve? Or what new features have been added? [ ] Have you added corresponding test cases? [ ] Have you modified the corresponding document? [ ] Is this PR backward compatible? If it is not backward compatible, please discuss on the mailing list first Codecov Report Merging #1595 (46269c2) into master (486b46a) will increase coverage by 0.26%. The diff coverage is 45.58%. :exclamation: Current head 46269c2 differs from pull request most recent head cc876d7. Consider uploading reports for the commit cc876d7 to get more accurate results @@ Coverage Diff @@ ## master #1595 +/- ## ========================================== + Coverage 41.44% 41.70% +0.26% ========================================== Files 87 87 Lines 7420 7457 +37 ========================================== + Hits 3075 3110 +35 - Misses 3990 3992 +2 Partials 355 355 Impacted Files Coverage Δ pkg/config/config.go 63.54% <ø> (ø) pkg/providers/ingress/ingress.go 4.05% <0.00%> (-0.04%) :arrow_down: pkg/providers/utils/ingress_status.go 31.34% <77.77%> (+31.34%) :arrow_up: pkg/providers/utils/status.go 30.43% <77.77%> (+30.43%) :arrow_up: cmd/ingress/ingress.go 82.90% <100.00%> (+0.14%) :arrow_up: pkg/providers/apisix/translation/apisix_route.go 31.21% <100.00%> (+0.81%) :arrow_up: :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
gharchive/pull-request
2023-01-10T17:16:51
2025-04-01T04:33:26.436524
{ "authors": [ "AlinsRan", "codecov-commenter" ], "repo": "apache/apisix-ingress-controller", "url": "https://github.com/apache/apisix-ingress-controller/pull/1595", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2166265414
fix: rat check error in vscode ide Which issue does this PR close? Closes #160 . Rationale for this change What changes are included in this PR? How are these changes tested? @viirya cc Thanks @thexiay @snmvaughan
gharchive/pull-request
2024-03-04T08:36:44
2025-04-01T04:33:26.448357
{ "authors": [ "thexiay", "viirya" ], "repo": "apache/arrow-datafusion-comet", "url": "https://github.com/apache/arrow-datafusion-comet/pull/161", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1672964330
Improve RoundRobin RepartitionExec Describe the bug RoundRobin repartitioning currently does not distribute the input tasks evenly over the output channels, causing the work to be not distributed evenly. To Reproduce When loading the data in memory in the TPC-H benchmark, this can be seen in the number of batches in MemoryExec (which uses RoundRobin partitioning). MemoryExec: partitions=32, partition_sizes=[32, 32, 32, 32, 32, 32, 32, 32, 26, 26, 26, 25, 25, 25, 25, 25, 25, 25, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16], metrics=[] It has a bias for the first output partitions/channels. Expected behavior Batches should be distributed more evenly over output channels. Additional context No response Batches should be distributed more evenly over output channels. Seems to be a load balancing issue. If you could count the number of batches already distributed to each channel and not completed then the classical The Power of Two Choices in Randomized Load Balancing algorithm could be evaluated. @cristian-ilies-vasile yes, instead of round-robin repartitioning an improved scheme could be implemented based on number of buffered batches. One good article describing this technique can be read here: Deterministic Aperture: A distributed, load balancing algorithm https://blog.twitter.com/engineering/en_us/topics/infrastructure/2019/daperture-load-balancer
gharchive/issue
2023-04-18T11:57:25
2025-04-01T04:33:26.452277
{ "authors": [ "Dandandan", "cristian-ilies-vasile" ], "repo": "apache/arrow-datafusion", "url": "https://github.com/apache/arrow-datafusion/issues/6043", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1988120175
Fix: Do not try and preserve order when there is no order to preserve in RepartitionExec Which issue does this PR close? Closes https://github.com/apache/arrow-datafusion/issues/8043 Rationale for this change We encountered problem is our downstream tests that RepartitionExec was trying to preserve an order even when there is no order to preserve. This was because one codepath (RepartitionExec::new_with_children) it set the preserve_order flag to true when it creates a new RepartitionExec even if the new child had no order to preserve. During execution, this cause the code to try and merge a stream with no sort exprs, which causes an internal error What changes are included in this PR? Update RepartitionExec::with_preserve_order to avoid Improve documentation add test Are these changes tested? Yes, though I am not thrilled with the test (I will comment inline) Are there any user-facing changes? Is it possible to move this test to the same file with struct RepartitionExec as a unit test? yes, that is a good idea -- I did so in ee55ec3d016574851aa2e77d0cf773990bc7aa32 I also see one call with_preserve_order(false) which is unnecessary and confusing in enforce_distribution.rs as well as one such call in replace_with_order_preserving_variants.rs. Let's remove them. That is an excellent idea, I did so in 0c492e16fc7e0db5b17395559d878e8cd38caeba and I think the logic is much clearer now Thanks @alamb for this PR. Having this check at initialization is great, and doesn't lead to unexpected behavior.
gharchive/pull-request
2023-11-10T18:14:22
2025-04-01T04:33:26.457761
{ "authors": [ "alamb", "mustafasrepo" ], "repo": "apache/arrow-datafusion", "url": "https://github.com/apache/arrow-datafusion/pull/8127", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
867664618
Added support to length of Binary and List. Note: migrated from original JIRA: https://issues.apache.org/jira/browse/ARROW-11420 None I guess we could close this issue now.
gharchive/issue
2021-04-26T12:46:15
2025-04-01T04:33:26.459258
{ "authors": [ "HaoYang670", "alamb" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/issues/172", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1519442146
Release Arrow 30.0.1 (maintenance release for 30.0.0) Is your feature request related to a problem or challenge? Please describe what you are trying to do. We recently released arrow 30.0.0: https://github.com/apache/arrow-rs/issues/3336 and found two issues: [ ] The upgrade to multiversion https://github.com/apache/arrow-rs/pull/3396 means now all crates that use arrow-rs require rebuild on every compilation (see https://github.com/calebzulawski/target-features/issues/1) [x] https://github.com/apache/arrow-rs/issues/3454 Describe the solution you'd like A new release version 30.0.1 with just the fixes for: [ ] https://github.com/apache/arrow-rs/pull/3452 [ ] https://github.com/apache/arrow-rs/pull/3451 Describe alternatives you've considered Possibly release 30.1.0 (need dependency)? Just wait until the release of 31.0.0 #3415 Additional context Found while trying to update datafusion. This may block the upgrade and possibly delay the release https://github.com/apache/arrow-datafusion/issues/4776 cc @andygrove Release thread - https://lists.apache.org/thread/p5dqfopymxw0or7hqh1tl5npcj4338qw Shall this be closed since the release has already taken place?
gharchive/issue
2023-01-04T19:02:34
2025-04-01T04:33:26.465677
{ "authors": [ "alamb", "iajoiner", "tustvold" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/issues/3455", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1688007803
object_store: delete multiple Problem I want to delete multiple objects atomically. Solution Add a new method to ObjectStore. It would follow the API similar to S3 DeleteObjects Alternatives Multiple calls to delete aren't atomic. Discussion Not all implementations of ObjectStore support this. We could have an extension trait, ObjectStoreAtomic: ObjectStore. Would you consider accepting a PR that added this? Any thoughts on implementation? I believe this is a duplicate of #2615 FWIW S3 DeleteObjects is not atomic, it can partially fail - https://docs.aws.amazon.com/AmazonS3/latest/API/API_DeleteObjects.html Agreed. And thanks, good to know that about S3.
gharchive/issue
2023-04-28T06:56:24
2025-04-01T04:33:26.468979
{ "authors": [ "Kinrany", "tustvold" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/issues/4154", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1040107683
Fix 876, allow null array to be cast from and to other types Which issue does this PR close? Closes #876 Depends on #591 Rationale for this change What changes are included in this PR? Are there any user-facing changes? Looks like the tests are failing with dictionaries: Test casting Dictionary(Int8, Int32) --> Null thread 'compute::kernels::cast::tests::test_can_cast_types' panicked at 'not implemented: Take not supported for data type Null', arrow/src/compute/kernels/take.rs:283:14 failures: compute::kernels::cast::tests::test_can_cast_types Codecov Report Merging #884 (5143f63) into master (898924f) will increase coverage by 0.02%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## master #884 +/- ## ========================================== + Coverage 82.45% 82.47% +0.02% ========================================== Files 168 168 Lines 48231 48232 +1 ========================================== + Hits 39767 39778 +11 + Misses 8464 8454 -10 Impacted Files Coverage Δ arrow/src/compute/kernels/cast.rs 94.63% <100.00%> (+<0.01%) :arrow_up: arrow/src/datatypes/datatype.rs 65.36% <0.00%> (-0.44%) :arrow_down: parquet/src/arrow/array_reader.rs 77.72% <0.00%> (-0.09%) :arrow_down: arrow/src/ipc/writer.rs 84.28% <0.00%> (+0.18%) :arrow_up: arrow/src/array/data.rs 74.93% <0.00%> (+0.81%) :arrow_up: arrow/src/array/array.rs 85.71% <0.00%> (+2.33%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 898924f...5143f63. Read the comment docs.
gharchive/pull-request
2021-10-30T10:14:04
2025-04-01T04:33:26.482838
{ "authors": [ "Jimexist", "alamb", "codecov-commenter" ], "repo": "apache/arrow-rs", "url": "https://github.com/apache/arrow-rs/pull/884", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1614833496
[Python] Make ChunkedArray & Array iterable Describe the enhancement requested This would allow for more Pythonic patterns in downstream libraries. >>> next(pa.chunked_array([[1,2,3]])) Traceback (most recent call last): File "<stdin>", line 1, in <module> TypeError: 'pyarrow.lib.ChunkedArray' object is not an iterator Component(s) Python Can you give an example of in what kind of pattern you would want to use this? And for a ChunkedArray, do you expect that it gives a chunk or a scalar?
gharchive/issue
2023-03-08T08:13:42
2025-04-01T04:33:26.486503
{ "authors": [ "Fokko", "jorisvandenbossche" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/issues/34495", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1767775375
[Python] Python to array conversion sometimes ignores "safe=False" or is using unsafe instead of safe cast Describe the bug, including details regarding any error messages, version, and platform. There are two cases where the array constructor taking the python conversion path (python_to_arrow.cc) doesn't handle the safe keyword properly. The safe keyword is set to True by default and is ignored if passed One example where python list is taking python_to_arrow.cc code path are Decimals (which are python objects). Here the conversion from Decimal to int does not overflow by default and one can’t turn it off with safe=False: >>> import pyarrow as pa >>> pa.array([Decimal('1234')]).cast(pa.int8(), safe=False) <pyarrow.lib.Int8Array object at 0x7efbbd18e4b0> [ -46 ] >>> pa.array([Decimal('1234')], pa.int8(), safe=False) Traceback (most recent call last): ... ArrowInvalid: Value 1234 too large to fit in C integer type Another example is JSON data with nested data (list type). The conversion is taking python_to_arrow.cc code path where safe keyword is also ignored. See example in https://github.com/apache/arrow/issues/31402. In some cases the safe keyword is ignored and does unsafe conversions Nested case >>> import pyarrow as pa >>> pa.array(np.array([[1.5], [2.5, 3.5]], dtype=object), type=pa.list_(pa.int64()), safe=True) <pyarrow.lib.ListArray object at 0x7f004fc74700> [ [ 1 ], [ 2, 3 ] ] Primitive array case. Numpy array and python list take a different code path (numpy_to_array.cc vs python_to_arrow.cc): >>> import pyarrow as pa >>> pa.array(np.array([1.5, 2.5]), type=pa.int64(), safe=True) ... ArrowInvalid: Float value 1.5 was truncated converting to int64 vs >>> import pyarrow as pa >>> pa.array([1.5, 2.5], type=pa.int64(), safe=True) <pyarrow.lib.Int64Array object at 0x7f004fc72c40> [ 1, 2 ] Another example of wrong handling of safe keyword is when using nested data in pandas (an object thus taking the python_to_arrow.cc code path): >>> import pandas as pd >>> import pyarrow as pa >>> int_dataframe = pd.DataFrame({"array": [[1, 2]]}) >>> float_dataframe = pd.DataFrame({"array": [[1.5, 2.3]]}) >>> int_table = pa.Table.from_pandas(int_dataframe) >>> table = pa.Table.from_pandas(float_dataframe, schema=int_table.schema) >>> table pyarrow.Table array: list<item: int64> child 0, item: int64 ---- array: [[[1,2]]] cc @jorisvandenbossche @dane Component(s) Python @AlenkaF I may not be the Dane you're looking for, but I can confirm this is Python. Best of luck on addressing the bug! Oh sorry about that! :)
gharchive/issue
2023-06-21T14:44:15
2025-04-01T04:33:26.492985
{ "authors": [ "AlenkaF" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/issues/36202", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2368041912
[C++][Parquet] Add a RowGroup reader interface class Currently the logic for interacting with row group metadata and constructing column decoders is embedded in the parquet_reader.cc executable here: https://github.com/apache/parquet-cpp/blob/master/example/parquet_reader.cc With PARQUET-434, we have a file reader container, which can then provide a row group reader container, something like RowGroupReader* group_reader = file_reader->row_group(i); Reporter: Wes McKinney / @wesm Assignee: Wes McKinney](https://issues.apache.org/jira/secure/ViewProfile.jspa?name=wesm) / @wesm Note: This issue was originally created as PARQUET-451. Please see the migration documentation for further details. Wes McKinney / @wesm: This is done in https://github.com/apache/parquet-cpp/pull/23 The API currently returns std::shared_ptr instances owned by the parent file reader. Feedback on this API is requested. std::shared_ptr<RowGroupReader> reader = file_reader->RowGroup(i); Nong Li / @nongli: Issue resolved by pull request 23 https://github.com/apache/parquet-cpp/pull/23
gharchive/issue
2016-01-22T19:43:20
2025-04-01T04:33:26.498877
{ "authors": [ "asfimport" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/issues/42307", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1119989343
ARROW-3016: [Docs][C++] Memory profiling with perf Put together some scripts and an example for using perf to understand memory allocations in an Arrow C++ program. Thanks, this is pretty slick. Just two nits. Well the commands you provided earlier help me get unstuck on learning how to use perf probes, so thank you @lidavidm! 🙏
gharchive/pull-request
2022-01-31T22:12:46
2025-04-01T04:33:26.500721
{ "authors": [ "wjones127" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/12307", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
363411168
[GLib] Import Parquet GLib It's based on https://github.com/red-data-tools/parquet-glib at 32a2b2099f90c19d28322693dd853dc8643c55ec. See also: https://github.com/red-data-tools/parquet-glib/commit/32a2b2099f90c19d28322693dd853dc8643c55ec I started discussion at dev@parquet.apache.org: https://lists.apache.org/thread.html/72b2bf04d8035d854a44487d956322cbe6c8629488128bc326613fb1@<dev.parquet.apache.org> Codecov Report Merging #2622 into master will decrease coverage by 0.01%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #2622 +/- ## ========================================== - Coverage 87.17% 87.15% -0.02% ========================================== Files 380 380 Lines 59037 59037 ========================================== - Hits 51466 51456 -10 - Misses 7501 7507 +6 - Partials 70 74 +4 Impacted Files Coverage Δ go/arrow/math/int64_avx2_amd64.go 0% <0%> (-100%) :arrow_down: go/arrow/memory/memory_avx2_amd64.go 0% <0%> (-100%) :arrow_down: go/arrow/math/float64_avx2_amd64.go 0% <0%> (-100%) :arrow_down: go/arrow/math/uint64_avx2_amd64.go 0% <0%> (-100%) :arrow_down: go/arrow/memory/memory_amd64.go 28.57% <0%> (-14.29%) :arrow_down: go/arrow/math/math_amd64.go 31.57% <0%> (-5.27%) :arrow_down: go/arrow/math/float64_amd64.go 33.33% <0%> (ø) :arrow_up: go/arrow/math/int64_amd64.go 33.33% <0%> (ø) :arrow_up: go/arrow/math/uint64_amd64.go 33.33% <0%> (ø) :arrow_up: go/arrow/math/float64_sse4_amd64.go 100% <0%> (+100%) :arrow_up: ... and 3 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update a20d281...43ae5c5. Read the comment docs. Thanks @kou -- we should go through the IP clearance after having a vote on dev@parquet.a.o. Your ICLA should be sufficient for this (I see there is one other contributor with a 2 line change)? we should go through the IP clearance after having a vote on dev@parquet.a.o. OK. Your ICLA should be sufficient for this (I see there is one other contributor with a 2 line change)? Yes. The other contributor has one commit: https://github.com/red-data-tools/parquet-glib/commit/8091c812aedd4c94af98d1e664b8bb424c36b992 But the change isn't included in this pull request because we don't need the change by merging this project to the Arrow repository: https://github.com/apache/arrow/pull/2622/files#diff-88333024fdffef0c691b5aa164aa1a37R110 Great work @kou . I looked through the code and everything looks fine. We should do the formal vote and than we can merge this. Thanks for reviewing this. I'll create a PARQUET JIRA ticket when the vote is passed. @wesm Thanks for creating a JIRA ticket. Is it OK to use ARROW ticket instead of PARQUET ticket for Parquet GLib bindings? If it's OK, I'll also use ARROW ticket for new features. OK. I'll also use ARROW project.
gharchive/pull-request
2018-09-25T05:03:14
2025-04-01T04:33:26.519873
{ "authors": [ "codecov-io", "kou", "wesm", "xhochy" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/2622", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
386721068
ARROW-3916: [Python] Add support for filesystem kwarg in ParquetWriter Implements ARROW 3916. I'll resolve the flake8 failures this evening. Anyone know if I should be worried about the failing Java micro benchmark? Codecov Report Merging #3070 into master will increase coverage by 1.11%. The diff coverage is 95%. @@ Coverage Diff @@ ## master #3070 +/- ## ========================================== + Coverage 87.06% 88.18% +1.11% ========================================== Files 489 431 -58 Lines 68974 65330 -3644 ========================================== - Hits 60055 57610 -2445 + Misses 8818 7720 -1098 + Partials 101 0 -101 Impacted Files Coverage Δ python/pyarrow/parquet.py 93.89% <100%> (ø) :arrow_up: python/pyarrow/filesystem.py 52.91% <100%> (+2.08%) :arrow_up: python/pyarrow/tests/test_parquet.py 97.29% <92.85%> (-0.11%) :arrow_down: cpp/src/plasma/fling.cc 67.18% <0%> (-17.26%) :arrow_down: cpp/src/arrow/util/thread-pool-test.cc 98.91% <0%> (-0.55%) :arrow_down: cpp/src/plasma/thirdparty/dlmalloc.c 47.16% <0%> (-0.38%) :arrow_down: cpp/src/arrow/util/parsing.h 95.6% <0%> (ø) :arrow_up: go/arrow/array/table.go go/arrow/math/uint64_amd64.go go/arrow/internal/testing/tools/bool.go ... and 62 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 98bdde8...9f70cfc. Read the comment docs. Are you able to rebase this?
gharchive/pull-request
2018-12-03T09:10:08
2025-04-01T04:33:26.533909
{ "authors": [ "codecov-io", "macks22", "wesm" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/3070", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1897179622
GH-34105: [R] Provide extra output for failed builds Rationale for this change This is a replacement for the previous PR https://github.com/apache/arrow/pull/37698. The rationale for this PR is providing extra output for R package builds where the C++ build fails What changes are included in this PR? Update the system call to save output when building Arrow C++ from the R package and output it if it's failed Are these changes tested? No automated tests but the changes have been tested manually. Are there any user-facing changes? Yes, but only for users building the R package from source which is hopefully not common. Closes: #34105 I've updated the PR to address points brought up in https://github.com/apache/arrow/pull/37727#discussion_r1326477485 so this is ready for another look.
gharchive/pull-request
2023-09-14T19:30:40
2025-04-01T04:33:26.537223
{ "authors": [ "amoeba" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/37727", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2291901373
GH-41608: [C++][Python] Extends the add_key_value to parquet::arrow and PyArrow Rationale for this change The previous pr ( https://github.com/apache/arrow/pull/34889 ) add a AddKeyValueMetadata to FileWriter. And now we should export it to Parquet Arrow and Python API. What changes are included in this PR? Add AddKeyValueMetadata in parquet::arrow Add add_key_value_metadata in pyarrow testing Are these changes tested? Yes Are there any user-facing changes? New api allowing add key-value metadata to Parquet file GitHub Issue: #41608 I've no idea why R language CI failed @AlenkaF @jorisvandenbossche @pitrou Would you mind take a look? Migrate to test_parquet_writer.py now. @pitrou would you mind revisit C++ part or merge this? @pitrou comment fixed @mapleFU Feel free to merge if CI is fine. CI failed is unrelated, merge
gharchive/pull-request
2024-05-13T06:04:57
2025-04-01T04:33:26.542632
{ "authors": [ "mapleFU", "pitrou" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/41633", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
445576055
ARROW-5360: [Rust] Update rustyline to fix build This is placeholder to check if the pending update to rustyline will fix the build issue. I'll update with the correct crate once it's published @wesm @sunchao PTAL, I'll wait to see if the build passes before I go to bed.
gharchive/pull-request
2019-05-17T19:00:34
2025-04-01T04:33:26.543685
{ "authors": [ "nevi-me" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/4337", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
587985757
ARROW-8220: [Python] Make dataset FileFormat objects serializable Also did some refactoring for a more pleasant user API. Also did some refactoring for a more pleasant user API. I also don't like the ParquetFileFormatReaderOptions very much as user API, but, I am not sure we can just pass them all to ParquetFileFormat, since we are going to use that for both reading and writing, and mixing keywords for those all in a single constructor is going to get confusing. I think we should rather give a better API in a parquet / reading specific API like parquet.read_table or ParquetDataset. @jorisvandenbossche agree. May we defer your suggestion to a follow-up? Well, my comment is kind of: we need to keep ParquetFileFormatReaderOptions, so since you are removing that, I would rather not defer that to a follow-up (but you don't need to agree with keeping it, of course :-)) ParquetFileFormatReaderOptions was still bound to the ParquetFormat, your proposal is more about making the reader and writer options independent from the ParquetFormat. So this PR doesn't change that dependency. I can wire these options, but it's not entirely clear because we don't have a read() method on the datasets. Once we add support for writing we can refine the API. ParquetFileFormatReaderOptions was still bound to the ParquetFormat, your proposal is more about making the reader and writer options independent from the ParquetFormat. So this PR doesn't change that dependency. Yes, it is still bound to the format, but it splits its keywords in two groups: format = ParquetFileFormat(reader_options=dict(...), writer_options=dict(...)) it's not entirely clear because we don't have a read() method on the datasets I think to_table is the "read" method? Once we add support for writing we can refine the API. Yeah, I fully agree much of this discussion is a bit "up in the air", since we don't yet have writing, so don't yet know how we would want to make the API for writing. But it's for that reason that I commented to keep it as is, as there is also no clear reason yet for changing IMO, since we don't know the final API with writing (but it was an explicit decision, at least on the C++ side, to have this a separate set of options instead of direct ParquetFileFormat options). But OK, since it is easy to put it back later, I won't block removing it if you prefer that :) @jorisvandenbossche updated as you requested Build failure is unrelated.
gharchive/pull-request
2020-03-25T20:56:05
2025-04-01T04:33:26.549545
{ "authors": [ "jorisvandenbossche", "kszucs" ], "repo": "apache/arrow", "url": "https://github.com/apache/arrow/pull/6720", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
202608973
[BEAM-1292] PubSubIO: fix and improve testing for DisplayData Also adds better type/nullability checking in the code. R: @tgroh Coverage decreased (-0.04%) to 69.743% when pulling fcee4e4d434d18e719a5c729a2712cbf57f423d4 on dhalperi:pubsub-display-data into 9db5f746ae81c4fa755a9a8f2c4888759c8e7042 on apache:master.
gharchive/pull-request
2017-01-23T18:09:46
2025-04-01T04:33:26.557067
{ "authors": [ "coveralls", "dhalperi" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/1817", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1260178897
Add mockito configuration to support static mocking without mockito-i… …nline Following the tips here: https://stackoverflow.com/questions/65986197/difference-between-mockito-core-vs-mockito-inline Removing extra dep added here: https://github.com/apache/beam/pull/17753/ Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Choose reviewer(s) and mention them in a comment (R: @username). [ ] Add a link to the appropriate issue in your description, if applicable. This will automatically link the pull request to the issue. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. hm this seems to definitely have broken something, hah.
gharchive/pull-request
2022-06-03T17:38:03
2025-04-01T04:33:26.564412
{ "authors": [ "pabloem" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/18385", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
207300666
[BEAM-1471]: Add cross-sdk tests for IterableCoder Be sure to do all of the following to help us incorporate your contribution quickly and easily: [ ] Make sure the PR title is formatted like: [BEAM-<Jira issue #>] Description of pull request [ ] Make sure tests pass via mvn clean verify. (Even better, enable Travis-CI on your fork and ensure the whole test matrix passes). [ ] Replace <Jira issue #> in the title with the actual Jira issue number, if there is one. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. R: @dhalperi @robertwb Coverage decreased (-0.003%) to 69.713% when pulling 965160c0c0bada4500d51ca44c6c8e5e1c0476bf on vikkyrk:common_iterable_coder into 30cb93ced35494db79c81c49e07a298f560c3905 on apache:master. retest this please Changes Unknown when pulling cab5e6347f3fcece2cac4819e268488d7ce66830 on vikkyrk:common_iterable_coder into ** on apache:master**.
gharchive/pull-request
2017-02-13T18:28:38
2025-04-01T04:33:26.569698
{ "authors": [ "coveralls", "dhalperi", "vikkyrk" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/1996", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
214208676
[BEAM-1721] Do not shift Timestamps forwards in Reshuffle Be sure to do all of the following to help us incorporate your contribution quickly and easily: [x] Make sure the PR title is formatted like: [BEAM-<Jira issue #>] Description of pull request [x] Make sure tests pass via mvn clean verify. (Even better, enable Travis-CI on your fork and ensure the whole test matrix passes). [x] Replace <Jira issue #> in the title with the actual Jira issue number, if there is one. [x] If this contribution is large, please file an Apache Individual Contributor License Agreement. Timestamps can be shifted forwards after the fact, but cannot generally be shifted backwards. Because reshuffle outputs "as quickly as possible", only elements that arrive approximately simulatenously with each other will have their timestamps shifted. There is currently no way to output all input elements with their original timestamps without explicitly reifying those timestamps and reassigning them on the output elements. Coverage decreased (-0.003%) to 70.177% when pulling 347250b80fafecbdb12059233255e526370a1623 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master. R: @jkff Coverage decreased (-0.008%) to 70.173% when pulling da185f577c8193ea2dd9071c183b385fe6f83560 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master. Coverage decreased (-0.01%) to 70.17% when pulling 2a7a4adc36c058f966ba235ae2da01e0796e96c7 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master. Coverage decreased (-0.001%) to 70.179% when pulling 2a7a4adc36c058f966ba235ae2da01e0796e96c7 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master. Coverage increased (+0.002%) to 70.183% when pulling 2a7a4adc36c058f966ba235ae2da01e0796e96c7 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master. Started to explicitly reify timestamps in Reshuffle and restore them after the fact. CC: @dpmills retest this please retest this please Coverage decreased (-0.05%) to 70.137% when pulling 4763f74d325a5c22e82236a8225ec5248ba2a5cb on tgroh:reshuffle_output_time_fn into 49c29f1faebddebedb151013cc18c4c4f2237644 on apache:master. retest this please Coverage decreased (-0.04%) to 70.156% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master. Coverage decreased (-0.04%) to 70.156% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master. Coverage decreased (-0.03%) to 70.159% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master. retest this please Coverage decreased (-0.04%) to 70.156% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master. retest this please Coverage decreased (-0.03%) to 70.159% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master.
gharchive/pull-request
2017-03-14T21:14:26
2025-04-01T04:33:26.587758
{ "authors": [ "coveralls", "jkff", "tgroh" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/2246", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1412365582
GA Migration Java Precommit IO AWS, AWS2, Azure, Cassandra, CDAP ( Java Precommit Sharding) This PR is part of the migration from jenkins to github actions. This PR is part of the Precommit Java Sharding This PR contains this tests Precommit Java IO AWS Precommit Java IO AWS2 Precommit Java IO Azure Precommit Java IO Cassandra Precommit Java IO CDAP 4 Files were added to the workflows folder of github actions job-precommit-java-io-aws.yml job-precommit-java-io-aws2.yml job-precommit-java-io-azure.yml job-precommit-java-io-cassandra.yml job-precommit-java-io-cdap.yml CI.md was modified adding the new GA workflows Waiting for this PR before merge https://github.com/apache/beam/pull/22703 Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Choose reviewer(s) and mention them in a comment (R: @username). [ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. R: @damccorm This PR was missing in the document with all the other PR's I'll add it ( sorry for that). It's also part of the Java Precommit Sharding tests.
gharchive/pull-request
2022-10-17T23:29:34
2025-04-01T04:33:26.597782
{ "authors": [ "andoni-guzman" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/23683", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1433143103
[Task]: PR Bot will push commits only if they are non-empty fixes #23851 Currently a number of non-empty commits are pushed for the PR bot. This pr provides for modifying the bot such that it commits & pushes only if there are any changes made. Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Choose reviewer(s) and mention them in a comment (R: @username). [ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI. R: @pabloem This looks great! Thanks for sticking with this one! Thanks for helping me so much . I also learnt about exitcodes in git and how they can be used with OR operator by using your idea. 🙂
gharchive/pull-request
2022-11-02T13:24:16
2025-04-01T04:33:26.606169
{ "authors": [ "ayushthe1" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/23937", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1846665445
[Website] add octo case-study add octo case-study with related files before: after: Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI or the workflows README to see a list of phrases to trigger workflows. R: @aaltay, could you kindly, review current PR?
gharchive/pull-request
2023-08-11T11:30:50
2025-04-01T04:33:26.613966
{ "authors": [ "bullet03" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/27967", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1953289679
[YAML] Improved pipeline schema definition. Slightly stricter definitions for catching more errors, as well as avoding the use of anyOf which often makes it difficult to deduce what the true error is. This does mean a pipeline must have a transform (or source/sink) block rather than simply be itself a list of transforms. Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily: [ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead. [ ] Update CHANGES.md with noteworthy changes. [ ] If this contribution is large, please file an Apache Individual Contributor License Agreement. See the Contributor Guide for more tips on how to make review process smoother. To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md GitHub Actions Tests Status (on master branch) See CI.md for more information about GitHub Actions CI or the workflows README to see a list of phrases to trigger workflows. R: @damccorm
gharchive/pull-request
2023-10-20T00:54:00
2025-04-01T04:33:26.621287
{ "authors": [ "robertwb" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/29083", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
228432462
[BEAM-1345] Remove FileSystems.setDefaultConfigInWorkers since Dataflow no longer depends on this Be sure to do all of the following to help us incorporate your contribution quickly and easily: [x] Make sure the PR title is formatted like: [BEAM-<Jira issue #>] Description of pull request [ ] Make sure tests pass via mvn clean verify. [x] Replace <Jira issue #> in the title with the actual Jira issue number, if there is one. [x] If this contribution is large, please file an Apache Individual Contributor License Agreement. R: @dhalperi Dataflow worker for master still needs to be updated, this is to be held till that occurs LGTM, once Dataflow worker for master is updated and examples precommit is green. Coverage decreased (-0.002%) to 70.669% when pulling e6650d9d4a5fb68bc45acc9286bec52c44a83a4f on lukecwik:cleanup into 49245080ad6393b5678077caa047f7a6ef0efffc on apache:master. retest this please Coverage increased (+0.005%) to 70.676% when pulling e6650d9d4a5fb68bc45acc9286bec52c44a83a4f on lukecwik:cleanup into 49245080ad6393b5678077caa047f7a6ef0efffc on apache:master.
gharchive/pull-request
2017-05-12T23:46:53
2025-04-01T04:33:26.626852
{ "authors": [ "coveralls", "dhalperi", "lukecwik" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/3134", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2489574014
Support ordered list states in python sdk and fnapi runner This is a follow-up PR of #30317, where we support ordered list state in java sdk. Here we implement ordered list state for python sdk. We also add the support of ordered list state into Python FnApi runner for debug and testing purpose. R: @robertwb @acrites @robertwb Could you please take another look? Thanks!
gharchive/pull-request
2024-08-27T14:36:46
2025-04-01T04:33:26.628308
{ "authors": [ "shunping" ], "repo": "apache/beam", "url": "https://github.com/apache/beam/pull/32326", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
800358958
Build failure when artifact cache is incorrectly set up See original issue on GitLab In GitLab by [Gitlab user @raoul].hidalgocharman on Nov 14, 2018, 15:46 Summary If you specify an artifact cache incorrectly, you eventually get something along these lines [00:00:08][f0c3789e][build:hello.bst] FAILURE push_directory was called, but no remote artifact servers are configured as push remotes. Either more useful information or a warning that continues despite not successfully pushing to a directory would be preferable. Steps to reproduce Add an artifact cache in project.conf that isn't actually set up, and a remote-execution service (but it doesn't need to be configured correctly). For example, provided you don't have an artifact server running locally on 50052: artifacts: - url: https://localhost:50052 push: true remote-execution: url: http://remote-execution-server:50051 Then try and build an artifact. What is the current bug behavior? It warns at the beginning that it failed to initialise the remote: [--:--:--][][] WARNING Failed to initialize remote http://localhost:50052: Connect Failed Then when trying to push an artifact after building it, the entire build fails. What is the expected correct behavior? Highlighting that it failed to initialise the remote in the build failure would be more useful , but it doesn't seem appropriate that an incorrectly configure artifact cache (or one that is currently down) should cause an entire build to fail. When remote execution isn't specified the build will still continue, and it should also do this when remote execution is specified. Relevant logs and/or screenshots Using the following project.conf for the autotools example project, with artifacts and remote-execution specified: # Unique project name name: autotools # Required BuildStream format version format-version: 18 # Subdirectory where elements are stored element-path: elements # Define some aliases for the tarballs we download aliases: alpine: https://gnome7.codethink.co.uk/tarballs/ gnu: http://ftpmirror.gnu.org/gnu/automake/ artifacts: - url: http://localhost:50052 push: true remote-execution: url: http://remote-execution-server:50051 When trying to build you get the following failure message on hello.bst: [00:00:06][f0c3789e][build:hello.bst ] FAILURE Running configure-commands [--:--:--][f0c3789e][build:hello.bst ] START Caching artifact [00:00:00][f0c3789e][build:hello.bst ] SUCCESS Caching artifact [00:00:09][f0c3789e][build:hello.bst ] FAILURE push_directory was called, but no remote artifact servers are configured as push remotes. Printing the last 20 lines from log file: /home/raoulhidalgo/.cache/buildstream/logs/autotools/hello/f0c3789e-build.26301.log ====================================================================== [--:--:--] START hello.bst: Integrating sandbox [00:00:00] SUCCESS hello.bst: Integrating sandbox [--:--:--] START hello.bst: Staging sources [00:00:00] SUCCESS hello.bst: Staging sources [--:--:--] START hello.bst: Running configure-commands [--:--:--] STATUS hello.bst: Running configure-commands export NOCONFIGURE=1; if [ -x ./configure ]; then true; elif [ -x ./autogen ]; then ./autogen; elif [ -x ./autogen.sh ]; then ./autogen.sh; elif [ -x ./bootstrap ]; then ./bootstrap; elif [ -x ./bootstrap.sh ]; then ./bootstrap.sh; else autoreconf -ivf .; fi [00:00:06] FAILURE hello.bst: Running configure-commands [--:--:--] START hello.bst: Caching artifact [00:00:00] SUCCESS hello.bst: Caching artifact [00:00:09] FAILURE hello.bst: push_directory was called, but no remote artifact servers are configured as push remotes. ====================================================================== Build failure on element: hello.bst Possible fixes The exception is raised in buildstream._artifactcache.artifact_cache.py:735. Maybe make sure it doesn't attempt to use push_directory when no artifact caches get successfully set up, or make push_directory silently pass. This needs to be re-verified, I'm quite sure this is not the case anymore.
gharchive/issue
2021-02-03T14:03:05
2025-04-01T04:33:26.637109
{ "authors": [ "BuildStream-Migration-Bot", "gtristan" ], "repo": "apache/buildstream", "url": "https://github.com/apache/buildstream/issues/766", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1160648302
Some logging improvements See individual commits for details This looks like a good improvement to me, and I've run the tests locally for good measure.
gharchive/pull-request
2022-03-06T16:05:51
2025-04-01T04:33:26.638251
{ "authors": [ "abderrahim", "gtristan" ], "repo": "apache/buildstream", "url": "https://github.com/apache/buildstream/pull/1611", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2342774658
TLS support of Knative Requirement Knative Eventing has recently released TLS support for event delivery, which is currently in the alpha stage. Adding support for this feature in Camel-K would ensure encrypted event delivery. The annoucement link can be found here The corresponding github issue can be found here The public design proposal can be found here Knative Eventing TLS project board can be found here. Problem Currently, Camel-K does not support the newly released TLS feature for Knative Eventing components. This limitation prevents users from leveraging encrypted event delivery within their clusters and third-party sinks when they enabled transport-encryption feature flag in Knative. Proposal No response Open questions No response Thanks for the feature request. Just a little notice though. The Knative feature is still marked as "experimental". Any contribution is welcome as usual anyway. /remove-stale
gharchive/issue
2024-06-10T03:47:11
2025-04-01T04:33:26.642210
{ "authors": [ "Leo6Leo", "squakez" ], "repo": "apache/camel-k", "url": "https://github.com/apache/camel-k/issues/5607", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2346790751
Trait annotations saved on the IntegrationKit resource What happened? The IntegrationKit resource seems to save the trait annotations that have been used to build the kit. Some trait annotation configs do not influence the kit and get saved, too. This is a problem because subsequent integrations that resolve to that IntegrationKit inherit the trait annotation config and this influences the integration runtime. As an example you can use trait.camel.apache.org/knative-service.enabled: 'false' annotation. When the integration that triggers the kit build uses this annotation it gets saved to the IntegrationKit resource: apiVersion: camel.apache.org/v1 kind: IntegrationKit metadata: annotations: camel.apache.org/operator.id: camel-k trait.camel.apache.org/knative-service.enabled: 'false' This is a problem because every integration that uses this IntegrationKit will also inherit the annotation trait config leading to the Knative service trait being disabled for the all integrations that use the kit. Steps to reproduce Start an integration that uses trait.camel.apache.org/knative-service.enabled: 'false' Make sure that the IntegrationKit is build for this integration and see the Knative service condition being disabled explicitly status: conditions: - lastTransitionTime: '2024-06-11T15:59:40Z' lastUpdateTime: '2024-06-11T15:59:40Z' message: explicitly disabled reason: KnativeServiceNotAvailable status: 'False' type: KnativeServiceKnativeServiceAvailable Start an integration that matches the previously built kit but does not use the trait annotation at all See the Knative service trait also being disabled for this 2nd integration because it is explicitly disabled by trait config inherited from the integration kit We would expect the Knative service being enabled for the 2nd integration because this is the default behavior when Knative serving is available on the cluster Relevant log output No response Camel K version 2.3.3 Yeah, that's a bad side effect of using trait into annotations. Probably those annotations should be completely filtered out. If you notice, the IntegrationKit, by design has only a subset of traits: https://github.com/apache/camel-k/blob/489f42ede99ccd91b64bb9efdc28e9593d3770f0/pkg/apis/camel/v1/integrationkit_types.go#L82 and they are populated in the Quarkus trait when creating a new IntegrationKit. I'm working on the fix, and I've realized the usage of traits in annotations can be simplified a lot in general. However, I am wondering if for Integrations, IntegrationKits and IntegrationPlatforms it does make sense to have the possibility to use also annotations to provide such a configuration. I think it's fine in Pipes, but cannot figure it out the use case for Integrations (ie, if the user can create an annotation, it should be able to edit easily the Integration spec as well). @lburgazzoli I see this was introduced to support Pipes (well, KameletBindings [1]) but was moved automatically to the others custom resources. I'd be inclined to deprecate, unless there is any valid reason to keep it. [1] https://github.com/apache/camel-k/pull/2338 I'm working on the fix, and I've realized the usage of traits in annotations can be simplified a lot in general. However, I am wondering if for Integrations, IntegrationKits and IntegrationPlatforms it does make sense to have the possibility to use also annotations to provide such a configuration. I think it's fine in Pipes, but cannot figure it out the use case for Integrations (ie, if the user can create an annotation, it should be able to edit easily the Integration spec as well). @lburgazzoli I see this was introduced to support Pipes (well, KameletBindings [1]) but was moved automatically to the others custom resources. I'd be inclined to deprecate, unless there is any valid reason to keep it. No, I don't think there is any more reason to use traits as annotations on resources other than Pipes
gharchive/issue
2024-06-11T16:01:30
2025-04-01T04:33:26.649843
{ "authors": [ "christophd", "lburgazzoli", "squakez" ], "repo": "apache/camel-k", "url": "https://github.com/apache/camel-k/issues/5620", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1695981970
Upgraded snmp4j to another provider to avoid flaky tests follows https://github.com/apache/camel-quarkus/pull/4870 follows https://github.com/apache/camel-quarkus/pull/4867#issuecomment-1534290658 Camel in 2.13.x brings org.apache.servicemix.bundles.snmp4j which causes random errors in snmp tests (with low resources). For example: java.nio.BufferOverflowException java.util.ConcurrentModificationException The actual length of the SEQUENCE object org.snmp4j.PDUv1 is 11, but 14 was expected Solution is to switch to org.snmp4j.snmp4j (by forcing in CQ BOM - which is purpose of this PR). There is different method signature (during call of SecurityModels.getInstance().addSecurityModel(this.usm);, therefore both producers have to be recompiled in CQ. Because both classes contains a lot of private fileds, it was not possible to override only 1 method. Therefore both classes are copied from the Camel (without any change). If Camel 3.x in future contains org.snmp4j.snmp4j. This PR could be reverted. @ppalaga @jamesnetherton FYI Great work, @JiriOndrusek! I am totally happy that we can make this work also on CQ 2.13. I have a couple of questions/suggestions: Could you please check what is the impact on the generated catalog in catalog/target/classes/org/apache/camel/catalog/quarkus after building the whole source tree before and after the change? Ideally, there should be no changes. Esp. snmp.json should still be there and the metadata in it should mirror the real state. We should inform users in the docs that Camel Quarkus uses a different protocol implementation than plain Camel 3.18. Explain reasons, possible gotchas (if any) and explain that org.snmp4j.snmp4j will be used in Camel 4 and CQ 3 anyway. Practical: Could please the code copied from Camel keep the original java source file structure, so that it is easier to compare what changes have been done between the two and so that it is easier to port any future changes from Camel 3.18? Question: so now the SNMP tests in CQ 2.13.x and main are 100% identical, right? @ppalaga thanks for the suggestions! 1 - There is no difference (I used diff with the generated stuff before this PR and after PR) 2,3- Following PR (https://github.com/apache/camel-quarkus/pull/4873) adds doc, refactors class files from Camel (I had to change the name of the files to avoid duplicate java class error) 4 - PR for 2 and 3 with a small upstream change (https://github.com/apache/camel-quarkus/pull/4872), makes tests in CQ 2.13.x and main 100% identical (of course there is a difference javax vs jakarta) @JiriOndrusek thanks for adressing the suggestions! I dared to fixup some small glitches in this PR by creating a new PR to avoid another review round: https://github.com/apache/camel-quarkus/pull/4876 First, snm in the org.apache.camel.quarkus.component.snm.graal package name looked like a typo. I changed it to simply org.apache.camel.quarkus.component.snmp because we tend to use graal packages for substitutions. I also moved the camel classes to org.apache.camel.quarkus.component.snmp because I think there is no reason to introduce a duplicate package. Other than that, i rephrased the docs text a bit. I hope I have not broken anything.
gharchive/pull-request
2023-05-04T13:09:00
2025-04-01T04:33:26.659448
{ "authors": [ "JiriOndrusek", "ppalaga" ], "repo": "apache/camel-quarkus", "url": "https://github.com/apache/camel-quarkus/pull/4871", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
176924254
Camel 10319: SNMP Producer https://issues.apache.org/jira/browse/CAMEL-10319 Hello @Fabryprog Can you please amend the commit messages with a reference to the issue CAMEL-10319? This way we can always track the history of an issue. Please, run mvn -Psourcecheck too oscerd@ghost:~/workspace/apache-camel/camel/components/camel-snmp$ mvn -Psourcecheck [WARNING] [WARNING] Some problems were encountered while building the effective settings [WARNING] 'profiles.profile.id' must be unique but found duplicate profile with id jboss-public-repository @ /opt/apache-maven-3.3.9/conf/settings.xml [WARNING] 'profiles.profile.id' must be unique but found duplicate profile with id redhat-earlyaccess-repository @ /opt/apache-maven-3.3.9/conf/settings.xml [WARNING] [INFO] Scanning for projects... [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Camel :: SNMP 2.18.0-SNAPSHOT [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-enforcer-plugin:1.4.1:enforce (enforce-maven) @ camel-snmp --- [INFO] [INFO] --- maven-bundle-plugin:3.2.0:cleanVersions (versions) @ camel-snmp --- [INFO] [INFO] --- maven-checkstyle-plugin:2.17:checkstyle (validate) @ camel-snmp --- [INFO] Starting audit... [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpEndpoint.java:114:5: File contains tab characters (this is the first instance). [FileTabCharacter] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:2: Line does not match expected header line of ' * Licensed to the Apache Software Foundation (ASF) under one or more'. [Header] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:106:17: '{' at column 17 should have line break after. [LeftCurly] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:106:50: 'catch' is not followed by whitespace. [WhitespaceAround] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:106:65: '{' is not followed by whitespace. [WhitespaceAround] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:107:17: '{' at column 17 should have line break after. [LeftCurly] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:107:45: 'catch' is not followed by whitespace. [WhitespaceAround] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:107:60: '{' is not followed by whitespace. [WhitespaceAround] [ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/test/java/org/apache/camel/component/snmp/ProducerTest.java:2: Line does not match expected header line of ' * Licensed to the Apache Software Foundation (ASF) under one or more'. [Header] Audit done. [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 2.998 s [INFO] Finished at: 2016-09-14T17:06:47+02:00 [INFO] Final Memory: 30M/319M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-checkstyle-plugin:2.17:checkstyle (validate) on project camel-snmp: An error has occurred in Checkstyle report generation. Failed during checkstyle execution: There are 9 errors reported by Checkstyle 6.17 with camel-checkstyle.xml ruleset. -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException Maybe, squash all the commit into one only if possible. Thanks. NOW SOURCE CHECK IS OK! Can you please amend the commit messages with a reference to the issue CAMEL-10319? Or squash all the commit into only one? Thanks! @lburgazzolii have fix variable into consumer! I meant if the initialization of targetAddress, address and usm can be moved in onStart and related cleanup onStop , in general the number of object created should be reduced to the min. @lburgazzoli I entered your valuable guidance Can you please squash commit once again? Thanks. After that I'll merge. If there will be any change you can simply track your update on a new JIRA and open a PR. :-) @oscerd is it mandatory? :-) i merged the master to test it and now is a very tedious work! :-1: It isn't mandatory but read "merge branch 'master'" in the git log is so ugly... Merged! :-) Many thanks! Can you close this PR? Thanks :100: !!!!
gharchive/pull-request
2016-09-14T14:34:56
2025-04-01T04:33:26.667462
{ "authors": [ "Fabryprog", "lburgazzoli", "oscerd" ], "repo": "apache/camel", "url": "https://github.com/apache/camel/pull/1177", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
316782457
[CARBONDATA-2387]Primitive AVRO Datatype Checking Test case added to check all the primitive datatype support of AVRO and its corresponding mapping with carbon datatypes. Supported Primitive Datatype. 1. Boolean 2. Int 3. long 4. float -> To carbon Internally it is double. 5. double 6. String Not Supported 1.NULL Datatype 2.Bytes [x] Any interfaces changed? [x] Any backward compatibility impacted? [x] Document update required? [x] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. [x] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4160/ SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/4485/ Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5344/
gharchive/pull-request
2018-04-23T12:03:51
2025-04-01T04:33:26.673196
{ "authors": [ "CarbonDataQA", "ravipesala", "sounakr" ], "repo": "apache/carbondata", "url": "https://github.com/apache/carbondata/pull/2214", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
345113831
[wip]disable local dictionary by default Be sure to do all of the following checklist to help us incorporate your contribution quickly and easily: [ ] Any interfaces changed? [ ] Any backward compatibility impacted? [ ] Document update required? [ ] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. [ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7555/ Build Failed with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6309/ Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7564/ Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7568/ SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6033/ retest this please retest sdv please Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7603/ retest this please Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7627/ retest this please retest this please retest this please Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7639/ Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6364/ SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6070/ Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6380/ retest sdv please SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6077/
gharchive/pull-request
2018-07-27T07:04:12
2025-04-01T04:33:26.684042
{ "authors": [ "CarbonDataQA", "akashrn5", "brijoobopanna", "ravipesala" ], "repo": "apache/carbondata", "url": "https://github.com/apache/carbondata/pull/2570", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
392083144
[CARBONDATA-3161]Pipe dilimiter is not working for streaming table Be sure to do all of the following checklist to help us incorporate your contribution quickly and easily: [Issue]: During data load in streaming scenario the delimiter was taking from CSVInputformat but it was not set in CarbonAppendableStreamSink Class. [Solution]: Set parameters like Delimiters, Escape char, Max columns in CSVInputFormat for streaming also [No] Any interfaces changed? [No] Any backward compatibility impacted? [No ] Document update required? [Yes] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. [No] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. Build Success with Spark 2.3.2, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/10082/ Build Success with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/2035/ Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/1827/ LGTM
gharchive/pull-request
2018-12-18T10:11:45
2025-04-01T04:33:26.689096
{ "authors": [ "CarbonDataQA", "QiangCai", "brijoobopanna" ], "repo": "apache/carbondata", "url": "https://github.com/apache/carbondata/pull/2997", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1093249249
[WIP] Fix clean files removing wrong delta files Why is this PR needed? In the case where there are multiple delete delta files in a partition in a partition table, some delta files were being ignored and deleted, thus changing the value during the query What changes were proposed in this PR? Fixed the logic which checks which delta file to delete. Now checking the deltaStartTime and comparing it with deltaEndTime to check consider all the delta files during clean files. Does this PR introduce any user interface change? No Is any new testcase added? Yes, one test case has been added. Build Failed with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4449/ Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6193/ Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/584/ Build Success with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6194/ Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/585/ retest this please Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/586/ Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6195/ Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4451/ retest this please Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4452/ Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/587/ Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6197/ Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4453/ Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/588/ Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6198/ Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4454/ Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/589/ Build Failed with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4455/ Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6199/ Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/590/ retest this please Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6200/ Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4457/ Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/592/ failure is not related to PR LGTM
gharchive/pull-request
2022-01-04T11:12:05
2025-04-01T04:33:26.705789
{ "authors": [ "CarbonDataQA2", "akashrn5", "kunal642", "vikramahuja1001", "ydvpankaj99" ], "repo": "apache/carbondata", "url": "https://github.com/apache/carbondata/pull/4246", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2210701406
[CELEBORN-1359] Support Netty Logging at the network layer What changes were proposed in this pull request? Supporting Netty level logging at the network layer. To configure Netty level logging a LogHandler must be added to the channel pipeline. NettyLogger is introduced as a new class which is able to construct a log handler depending on the log level: In case of : a custom log handler is created which does not dump the message contents. This way the log is a bit more compact. Moreover when network level encryption is switched on this level might be sufficient. In case of : Netty's own log handler is used which dumps the message contents. Otherwise (when the logger is not TRACE or DEBUG) the pipeline does not contain a log handler (there is no runtime penalty for the default setting but a long running app/service must be restarted along with the new log level to have an effect). Backport: [SPARK-36719][CORE] Supporting Netty Logging at the network layer [SPARK-45377][CORE] Handle InputStream in NettyLogger Why are the changes needed? This level of logging proved to be sufficient during debugging some external shuffle related problem. Compared with the tcpdump this log lines can be more easily correlated with the Celeborn internal calls. Moreover the log layout can be configured to contain the thread names that way for a timeout a busy thread could be identified. Does this PR introduce any user-facing change? No. How was this patch tested? Local manually test. Code change LGTM, please update the PR description, we are using LOG4J2 with log4j2.xml rahter than LOG4J1 with log4j.properties log4j.logger.XXX @pan3793, I have updated the description of this pull request. PTAL.
gharchive/pull-request
2024-03-27T12:42:08
2025-04-01T04:33:26.712589
{ "authors": [ "SteNicholas", "pan3793" ], "repo": "apache/celeborn", "url": "https://github.com/apache/celeborn/pull/2423", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
528004028
creating template from snapshot does not work from snapshot list If you go to STORAGE (main section menu on the left), then select VOLUMES from the dropdown list, select a particular volume, than see snapshots, then select a snaphost - then clicking on button WILL work fine and popup window is show as follows: But if you go to STORAGE (main section menu on the left), then select SNAPSHOTS from the dropdown list and find one and click the button - then an exception happens in the javascript code: ISSUE TYPE Bug Report COMPONENT NAME UI CLOUDSTACK VERSION 4.13 CONFIGURATION OS / ENVIRONMENT SUMMARY STEPS TO REPRODUCE EXPECTED RESULTS ACTUAL RESULTS during the latest update of jquery the createTemplate functions for volumes and snapshots were unified into a shared function without fully abstracting out the entities 'volumes' and 'snapshots' We can duplicate the function as a quick-fix or put effort into the abstraction. I am choosing the first as the current UI is to be phased out. @andrijapanicsb @rhtyd @rafaelweingartner sounds good @DaanHoogland I've also seen (probably NOT related, but to mention...) that when you register the template from URL, previously there was a spinning circle just until the template was added to the DB - then you can actually click on it and see it being downloaded etc. In current master when you register the template, the circle is spinning all the time, until the template is actually completely downloaded, which is insane. - not sure if that's also a quick fix... actually @andrijapanicsb half an hour extra search gave me a location to add a conditional, testing is probably going to be more effort. What you are saying seems a different location in the code. i'll keep it in mind. Did you create an issue for that already? No @DaanHoogland, I was lazy...
gharchive/issue
2019-11-25T10:49:55
2025-04-01T04:33:26.727905
{ "authors": [ "DaanHoogland", "andrijapanicsb" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/issues/3716", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
228969233
[4.9+] CLOUDSTACK-9860: Power off VMs when force stop is provided The 'force' option provided with the stopVirtualMachine API command is often assumed to be a hard shutdown sent to the hypervisor, when in fact it is for CloudStacks' internal use. CloudStack should be able to send the 'hard' power-off request to the hosts. When forced parameter on the stopVM API is true, power off (hard shutdown) a VM. This uses initial changes from #1635 to pass the forced parameter to hypervisor plugin via the StopCommand, and fixes force stop (poweroff) handling for KVM, VMware and XenServer. This will speed up VM stops with forced=true. Useful for CI-like environments. The global setting vm.destroy.forcestop is now also honoured to force stop VMs while destroying them, if it is set to true. Pinging for review @DaanHoogland @abhinandanprateek @nathanejohnson @kiwiflyer @wido @karuturi and others. @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos6 ✔centos7 ✔debian. JID-727 @blueorangutan test @rhtyd a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests Nice @rhtyd thanks, will test it soon Trillian test result (tid-1099) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 29534 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1099-kvm-centos7.zip Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py Test completed. 47 look ok, 2 have error(s) Test Result Time (s) Test File test_04_rvpc_privategw_static_routes Failure 376.63 test_privategw_acl.py test_02_list_snapshots_with_removed_data_store Error 0.04 test_snapshots.py test_01_vpc_site2site_vpn Success 180.96 test_vpc_vpn.py test_01_vpc_remote_access_vpn Success 71.17 test_vpc_vpn.py test_01_redundant_vpc_site2site_vpn Success 265.85 test_vpc_vpn.py test_02_VPC_default_routes Success 269.30 test_vpc_router_nics.py test_01_VPC_nics_after_destroy Success 588.43 test_vpc_router_nics.py test_05_rvpc_multi_tiers Success 527.95 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Success 1306.50 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Success 574.37 test_vpc_redundant.py test_02_redundant_VPC_default_routes Success 786.83 test_vpc_redundant.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Success 1327.73 test_vpc_redundant.py test_09_delete_detached_volume Success 156.58 test_volumes.py test_08_resize_volume Success 156.41 test_volumes.py test_07_resize_fail Success 161.51 test_volumes.py test_06_download_detached_volume Success 156.39 test_volumes.py test_05_detach_volume Success 150.71 test_volumes.py test_04_delete_attached_volume Success 151.24 test_volumes.py test_03_download_attached_volume Success 156.31 test_volumes.py test_02_attach_volume Success 89.42 test_volumes.py test_01_create_volume Success 712.21 test_volumes.py test_deploy_vm_multiple Success 318.15 test_vm_life_cycle.py test_deploy_vm Success 0.03 test_vm_life_cycle.py test_advZoneVirtualRouter Success 0.02 test_vm_life_cycle.py test_10_attachAndDetach_iso Success 36.71 test_vm_life_cycle.py test_09_expunge_vm Success 125.29 test_vm_life_cycle.py test_08_migrate_vm Success 40.96 test_vm_life_cycle.py test_07_restore_vm Success 0.13 test_vm_life_cycle.py test_06_destroy_vm Success 130.90 test_vm_life_cycle.py test_03_reboot_vm Success 125.87 test_vm_life_cycle.py test_02_start_vm Success 10.20 test_vm_life_cycle.py test_01_stop_vm Success 35.30 test_vm_life_cycle.py test_CreateTemplateWithDuplicateName Success 95.77 test_templates.py test_08_list_system_templates Success 0.03 test_templates.py test_07_list_public_templates Success 0.04 test_templates.py test_05_template_permissions Success 0.06 test_templates.py test_04_extract_template Success 5.16 test_templates.py test_03_delete_template Success 5.12 test_templates.py test_02_edit_template Success 90.16 test_templates.py test_01_create_template Success 105.85 test_templates.py test_10_destroy_cpvm Success 191.71 test_ssvm.py test_09_destroy_ssvm Success 168.56 test_ssvm.py test_08_reboot_cpvm Success 101.57 test_ssvm.py test_07_reboot_ssvm Success 133.60 test_ssvm.py test_06_stop_cpvm Success 136.74 test_ssvm.py test_05_stop_ssvm Success 133.67 test_ssvm.py test_04_cpvm_internals Success 1.20 test_ssvm.py test_03_ssvm_internals Success 3.34 test_ssvm.py test_02_list_cpvm_vm Success 0.12 test_ssvm.py test_01_list_sec_storage_vm Success 0.13 test_ssvm.py test_01_snapshot_root_disk Success 16.18 test_snapshots.py test_04_change_offering_small Success 211.01 test_service_offerings.py test_03_delete_service_offering Success 0.05 test_service_offerings.py test_02_edit_service_offering Success 0.08 test_service_offerings.py test_01_create_service_offering Success 0.11 test_service_offerings.py test_02_sys_template_ready Success 0.14 test_secondary_storage.py test_01_sys_vm_start Success 0.23 test_secondary_storage.py test_09_reboot_router Success 40.36 test_routers.py test_08_start_router Success 35.36 test_routers.py test_07_stop_router Success 10.17 test_routers.py test_06_router_advanced Success 0.06 test_routers.py test_05_router_basic Success 0.04 test_routers.py test_04_restart_network_wo_cleanup Success 5.80 test_routers.py test_03_restart_network_cleanup Success 65.55 test_routers.py test_02_router_internal_adv Success 1.11 test_routers.py test_01_router_internal_basic Success 0.52 test_routers.py test_router_dns_guestipquery Success 77.20 test_router_dns.py test_router_dns_externalipquery Success 0.08 test_router_dns.py test_router_dhcphosts Success 272.78 test_router_dhcphosts.py test_router_dhcp_opts Success 21.81 test_router_dhcphosts.py test_01_updatevolumedetail Success 5.15 test_resource_detail.py test_01_reset_vm_on_reboot Success 181.44 test_reset_vm_on_reboot.py test_createRegion Success 0.05 test_regions.py test_create_pvlan_network Success 5.35 test_pvlan.py test_dedicatePublicIpRange Success 0.67 test_public_ip_range.py test_03_vpc_privategw_restart_vpc_cleanup Success 510.53 test_privategw_acl.py test_02_vpc_privategw_static_routes Success 446.30 test_privategw_acl.py test_01_vpc_privategw_acl Success 97.33 test_privategw_acl.py test_01_primary_storage_nfs Success 35.86 test_primary_storage.py test_createPortablePublicIPRange Success 10.19 test_portable_publicip.py test_createPortablePublicIPAcquire Success 15.55 test_portable_publicip.py test_isolate_network_password_server Success 60.31 test_password_server.py test_UpdateStorageOverProvisioningFactor Success 0.12 test_over_provisioning.py test_oobm_zchange_password Success 30.72 test_outofbandmanagement.py test_oobm_multiple_mgmt_server_ownership Success 16.37 test_outofbandmanagement.py test_oobm_issue_power_status Success 5.26 test_outofbandmanagement.py test_oobm_issue_power_soft Success 15.35 test_outofbandmanagement.py test_oobm_issue_power_reset Success 15.37 test_outofbandmanagement.py test_oobm_issue_power_on Success 15.40 test_outofbandmanagement.py test_oobm_issue_power_off Success 10.39 test_outofbandmanagement.py test_oobm_issue_power_cycle Success 10.38 test_outofbandmanagement.py test_oobm_enabledisable_across_clusterzones Success 82.73 test_outofbandmanagement.py test_oobm_enable_feature_valid Success 5.16 test_outofbandmanagement.py test_oobm_enable_feature_invalid Success 0.10 test_outofbandmanagement.py test_oobm_disable_feature_valid Success 5.25 test_outofbandmanagement.py test_oobm_disable_feature_invalid Success 0.14 test_outofbandmanagement.py test_oobm_configure_invalid_driver Success 0.12 test_outofbandmanagement.py test_oobm_configure_default_driver Success 0.13 test_outofbandmanagement.py test_oobm_background_powerstate_sync Success 23.52 test_outofbandmanagement.py test_extendPhysicalNetworkVlan Success 15.47 test_non_contigiousvlan.py test_01_nic Success 439.34 test_nic.py test_releaseIP Success 192.59 test_network.py test_reboot_router Success 403.51 test_network.py test_public_ip_user_account Success 10.28 test_network.py test_public_ip_admin_account Success 40.29 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Success 66.95 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Success 61.78 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Success 125.04 test_network.py test_delete_account Success 318.06 test_network.py test_02_port_fwd_on_non_src_nat Success 55.71 test_network.py test_01_port_fwd_on_src_nat Success 111.87 test_network.py test_nic_secondaryip_add_remove Success 298.28 test_multipleips_per_nic.py test_list_zones_metrics Success 0.27 test_metrics_api.py test_list_volumes_metrics Success 5.56 test_metrics_api.py test_list_vms_metrics Success 191.85 test_metrics_api.py test_list_pstorage_metrics Success 0.40 test_metrics_api.py test_list_infrastructure_metrics Success 0.55 test_metrics_api.py test_list_hosts_metrics Success 0.36 test_metrics_api.py test_list_clusters_metrics Success 0.38 test_metrics_api.py login_test_saml_user Success 19.17 test_login.py test_assign_and_removal_lb Success 133.74 test_loadbalance.py test_02_create_lb_rule_non_nat Success 187.41 test_loadbalance.py test_01_create_lb_rule_src_nat Success 219.22 test_loadbalance.py test_03_list_snapshots Success 0.08 test_list_ids_parameter.py test_02_list_templates Success 0.04 test_list_ids_parameter.py test_01_list_volumes Success 0.05 test_list_ids_parameter.py test_07_list_default_iso Success 0.06 test_iso.py test_05_iso_permissions Success 0.07 test_iso.py test_04_extract_Iso Success 5.19 test_iso.py test_03_delete_iso Success 95.17 test_iso.py test_02_edit_iso Success 0.06 test_iso.py test_01_create_iso Success 21.06 test_iso.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Success 228.94 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Success 148.70 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Success 556.22 test_internal_lb.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Success 431.23 test_internal_lb.py test_dedicateGuestVlanRange Success 10.28 test_guest_vlan_range.py test_UpdateConfigParamWithScope Success 0.14 test_global_settings.py test_rolepermission_lifecycle_update Success 6.14 test_dynamicroles.py test_rolepermission_lifecycle_list Success 6.01 test_dynamicroles.py test_rolepermission_lifecycle_delete Success 5.85 test_dynamicroles.py test_rolepermission_lifecycle_create Success 5.89 test_dynamicroles.py test_rolepermission_lifecycle_concurrent_updates Success 6.02 test_dynamicroles.py test_role_lifecycle_update_role_inuse Success 5.89 test_dynamicroles.py test_role_lifecycle_update Success 10.99 test_dynamicroles.py test_role_lifecycle_list Success 5.92 test_dynamicroles.py test_role_lifecycle_delete Success 10.95 test_dynamicroles.py test_role_lifecycle_create Success 5.93 test_dynamicroles.py test_role_inuse_deletion Success 5.92 test_dynamicroles.py test_role_account_acls_multiple_mgmt_servers Success 8.13 test_dynamicroles.py test_role_account_acls Success 8.28 test_dynamicroles.py test_default_role_deletion Success 6.05 test_dynamicroles.py test_04_create_fat_type_disk_offering Success 0.11 test_disk_offerings.py test_03_delete_disk_offering Success 0.05 test_disk_offerings.py test_02_edit_disk_offering Success 0.07 test_disk_offerings.py test_02_create_sparse_type_disk_offering Success 0.10 test_disk_offerings.py test_01_create_disk_offering Success 0.12 test_disk_offerings.py test_deployvm_userdispersing Success 20.59 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userconcentrated Success 50.78 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_firstfit Success 60.71 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userdata_post Success 10.39 test_deploy_vm_with_userdata.py test_deployvm_userdata Success 70.84 test_deploy_vm_with_userdata.py test_02_deploy_vm_root_resize Success 5.96 test_deploy_vm_root_resize.py test_01_deploy_vm_root_resize Success 6.00 test_deploy_vm_root_resize.py test_00_deploy_vm_root_resize Success 283.01 test_deploy_vm_root_resize.py test_deploy_vm_from_iso Success 217.56 test_deploy_vm_iso.py test_DeployVmAntiAffinityGroup Success 76.08 test_affinity_groups.py test_03_delete_vm_snapshots Skipped 0.00 test_vm_snapshots.py test_02_revert_vm_snapshots Skipped 0.00 test_vm_snapshots.py test_01_create_vm_snapshots Skipped 0.00 test_vm_snapshots.py test_06_copy_template Skipped 0.00 test_templates.py test_static_role_account_acls Skipped 0.02 test_staticroles.py test_01_scale_vm Skipped 0.00 test_scale_vm.py test_01_primary_storage_iscsi Skipped 0.06 test_primary_storage.py test_06_copy_iso Skipped 0.00 test_iso.py test_deploy_vgpu_enabled_vm Skipped 0.01 test_deploy_vgpu_enabled_vm.py Looks good @rhtyd ! LGTM based on the code @rhtyd In cloudstack api doc, forced means "Force stop the VM (vm is marked as Stopped even when command fails to be send to the backend). The caller knows the VM is stopped." If you change the process, the api doc needs to be changed as well @ustcweizhou thanks, fixed the api doc for stopVirtualMachine API and updated the JIRA id. @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos6 ✔centos7 ✔debian. JID-729 @blueorangutan test centos7 vmware-55u3 @rhtyd a Trillian-Jenkins test job (centos7 mgmt + vmware-55u3) has been kicked to run smoke tests code LGTM. Trillian test result (tid-1100) Environment: vmware-55u3 (x2), Advanced Networking with Mgmt server 7 Total time taken: 55717 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1100-vmware-55u3.zip Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermitten failure detected: /marvin/tests/smoke/test_routers.py Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py Intermitten failure detected: /marvin/tests/smoke/test_ssvm.py Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermitten failure detected: /marvin/tests/smoke/test_vpc_vpn.py Test completed. 44 look ok, 5 have error(s) Test Result Time (s) Test File test_01_vpc_remote_access_vpn Failure 961.01 test_vpc_vpn.py test_04_cpvm_internals Failure 0.65 test_ssvm.py test_04_rvpc_privategw_static_routes Failure 899.97 test_privategw_acl.py test_01_vpc_privategw_acl Failure 977.00 test_privategw_acl.py test_04_rvpc_network_garbage_collector_nics Error 961.96 test_vpc_redundant.py test_02_list_snapshots_with_removed_data_store Error 86.03 test_snapshots.py test_02_list_snapshots_with_removed_data_store Error 91.13 test_snapshots.py test_01_vpc_site2site_vpn Success 376.46 test_vpc_vpn.py test_01_redundant_vpc_site2site_vpn Success 624.65 test_vpc_vpn.py test_02_VPC_default_routes Success 340.94 test_vpc_router_nics.py test_01_VPC_nics_after_destroy Success 726.70 test_vpc_router_nics.py test_05_rvpc_multi_tiers Success 689.97 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Success 716.86 test_vpc_redundant.py test_02_redundant_VPC_default_routes Success 730.98 test_vpc_redundant.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Success 1421.53 test_vpc_redundant.py test_09_delete_detached_volume Success 31.03 test_volumes.py test_06_download_detached_volume Success 70.73 test_volumes.py test_05_detach_volume Success 100.25 test_volumes.py test_04_delete_attached_volume Success 10.22 test_volumes.py test_03_download_attached_volume Success 15.34 test_volumes.py test_02_attach_volume Success 96.00 test_volumes.py test_01_create_volume Success 520.84 test_volumes.py test_03_delete_vm_snapshots Success 275.22 test_vm_snapshots.py test_02_revert_vm_snapshots Success 232.23 test_vm_snapshots.py test_01_create_vm_snapshots Success 161.79 test_vm_snapshots.py test_deploy_vm_multiple Success 273.00 test_vm_life_cycle.py test_deploy_vm Success 0.03 test_vm_life_cycle.py test_advZoneVirtualRouter Success 0.03 test_vm_life_cycle.py test_10_attachAndDetach_iso Success 26.88 test_vm_life_cycle.py test_09_expunge_vm Success 125.26 test_vm_life_cycle.py test_08_migrate_vm Success 81.18 test_vm_life_cycle.py test_07_restore_vm Success 0.25 test_vm_life_cycle.py test_06_destroy_vm Success 10.18 test_vm_life_cycle.py test_03_reboot_vm Success 5.18 test_vm_life_cycle.py test_02_start_vm Success 25.32 test_vm_life_cycle.py test_01_stop_vm Success 10.18 test_vm_life_cycle.py test_CreateTemplateWithDuplicateName Success 302.16 test_templates.py test_08_list_system_templates Success 0.03 test_templates.py test_07_list_public_templates Success 0.04 test_templates.py test_05_template_permissions Success 0.07 test_templates.py test_04_extract_template Success 15.23 test_templates.py test_03_delete_template Success 5.11 test_templates.py test_02_edit_template Success 90.13 test_templates.py test_01_create_template Success 141.13 test_templates.py test_10_destroy_cpvm Success 267.05 test_ssvm.py test_09_destroy_ssvm Success 269.14 test_ssvm.py test_08_reboot_cpvm Success 156.62 test_ssvm.py test_07_reboot_ssvm Success 158.69 test_ssvm.py test_06_stop_cpvm Success 177.04 test_ssvm.py test_05_stop_ssvm Success 203.94 test_ssvm.py test_03_ssvm_internals Success 3.15 test_ssvm.py test_02_list_cpvm_vm Success 0.15 test_ssvm.py test_01_list_sec_storage_vm Success 0.17 test_ssvm.py test_01_snapshot_root_disk Success 66.55 test_snapshots.py test_04_change_offering_small Success 92.11 test_service_offerings.py test_03_delete_service_offering Success 0.06 test_service_offerings.py test_02_edit_service_offering Success 0.12 test_service_offerings.py test_01_create_service_offering Success 0.12 test_service_offerings.py test_02_sys_template_ready Success 0.16 test_secondary_storage.py test_01_sys_vm_start Success 0.24 test_secondary_storage.py test_09_reboot_router Success 156.01 test_routers.py test_08_start_router Success 135.91 test_routers.py test_07_stop_router Success 25.27 test_routers.py test_06_router_advanced Success 0.06 test_routers.py test_05_router_basic Success 0.04 test_routers.py test_04_restart_network_wo_cleanup Success 5.67 test_routers.py test_03_restart_network_cleanup Success 191.42 test_routers.py test_02_router_internal_adv Success 1.18 test_routers.py test_01_router_internal_basic Success 0.58 test_routers.py test_router_dns_guestipquery Success 76.85 test_router_dns.py test_router_dns_externalipquery Success 0.07 test_router_dns.py test_router_dhcphosts Success 121.81 test_router_dhcphosts.py test_router_dhcp_opts Success 21.76 test_router_dhcphosts.py test_01_updatevolumedetail Success 0.12 test_resource_detail.py test_01_reset_vm_on_reboot Success 65.62 test_reset_vm_on_reboot.py test_createRegion Success 0.04 test_regions.py test_create_pvlan_network Success 5.27 test_pvlan.py test_dedicatePublicIpRange Success 0.54 test_public_ip_range.py test_03_vpc_privategw_restart_vpc_cleanup Success 1135.21 test_privategw_acl.py test_02_vpc_privategw_static_routes Success 722.96 test_privategw_acl.py test_01_primary_storage_nfs Success 37.52 test_primary_storage.py test_createPortablePublicIPRange Success 15.25 test_portable_publicip.py test_createPortablePublicIPAcquire Success 15.57 test_portable_publicip.py test_isolate_network_password_server Success 64.85 test_password_server.py test_UpdateStorageOverProvisioningFactor Success 0.15 test_over_provisioning.py test_oobm_zchange_password Success 30.85 test_outofbandmanagement.py test_oobm_multiple_mgmt_server_ownership Success 16.47 test_outofbandmanagement.py test_oobm_issue_power_status Success 10.30 test_outofbandmanagement.py test_oobm_issue_power_soft Success 15.42 test_outofbandmanagement.py test_oobm_issue_power_reset Success 15.52 test_outofbandmanagement.py test_oobm_issue_power_on Success 15.35 test_outofbandmanagement.py test_oobm_issue_power_off Success 15.38 test_outofbandmanagement.py test_oobm_issue_power_cycle Success 15.55 test_outofbandmanagement.py test_oobm_enabledisable_across_clusterzones Success 82.98 test_outofbandmanagement.py test_oobm_enable_feature_valid Success 5.19 test_outofbandmanagement.py test_oobm_enable_feature_invalid Success 0.12 test_outofbandmanagement.py test_oobm_disable_feature_valid Success 5.20 test_outofbandmanagement.py test_oobm_disable_feature_invalid Success 0.13 test_outofbandmanagement.py test_oobm_configure_invalid_driver Success 0.12 test_outofbandmanagement.py test_oobm_configure_default_driver Success 0.10 test_outofbandmanagement.py test_oobm_background_powerstate_sync Success 23.53 test_outofbandmanagement.py test_extendPhysicalNetworkVlan Success 15.39 test_non_contigiousvlan.py test_01_nic Success 490.48 test_nic.py test_releaseIP Success 338.83 test_network.py test_reboot_router Success 625.22 test_network.py test_public_ip_user_account Success 10.29 test_network.py test_public_ip_admin_account Success 40.32 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Success 77.04 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Success 61.91 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Success 125.33 test_network.py test_delete_account Success 363.77 test_network.py test_02_port_fwd_on_non_src_nat Success 55.83 test_network.py test_01_port_fwd_on_src_nat Success 111.89 test_network.py test_nic_secondaryip_add_remove Success 218.15 test_multipleips_per_nic.py test_list_zones_metrics Success 0.25 test_metrics_api.py test_list_volumes_metrics Success 5.53 test_metrics_api.py test_list_vms_metrics Success 252.40 test_metrics_api.py test_list_pstorage_metrics Success 0.29 test_metrics_api.py test_list_infrastructure_metrics Success 0.45 test_metrics_api.py test_list_hosts_metrics Success 0.48 test_metrics_api.py test_list_clusters_metrics Success 0.41 test_metrics_api.py login_test_saml_user Success 19.41 test_login.py test_assign_and_removal_lb Success 148.48 test_loadbalance.py test_02_create_lb_rule_non_nat Success 207.42 test_loadbalance.py test_01_create_lb_rule_src_nat Success 207.97 test_loadbalance.py test_03_list_snapshots Success 0.09 test_list_ids_parameter.py test_02_list_templates Success 0.05 test_list_ids_parameter.py test_01_list_volumes Success 0.03 test_list_ids_parameter.py test_07_list_default_iso Success 0.07 test_iso.py test_05_iso_permissions Success 0.07 test_iso.py test_04_extract_Iso Success 5.14 test_iso.py test_03_delete_iso Success 95.18 test_iso.py test_02_edit_iso Success 0.06 test_iso.py test_01_create_iso Success 21.01 test_iso.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Success 586.29 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Success 460.80 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Success 1073.94 test_internal_lb.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Success 802.49 test_internal_lb.py test_dedicateGuestVlanRange Success 10.31 test_guest_vlan_range.py test_UpdateConfigParamWithScope Success 0.14 test_global_settings.py test_rolepermission_lifecycle_update Success 6.36 test_dynamicroles.py test_rolepermission_lifecycle_list Success 6.09 test_dynamicroles.py test_rolepermission_lifecycle_delete Success 5.99 test_dynamicroles.py test_rolepermission_lifecycle_create Success 5.92 test_dynamicroles.py test_rolepermission_lifecycle_concurrent_updates Success 6.19 test_dynamicroles.py test_role_lifecycle_update_role_inuse Success 6.00 test_dynamicroles.py test_role_lifecycle_update Success 6.05 test_dynamicroles.py test_role_lifecycle_list Success 5.96 test_dynamicroles.py test_role_lifecycle_delete Success 5.98 test_dynamicroles.py test_role_lifecycle_create Success 5.95 test_dynamicroles.py test_role_inuse_deletion Success 5.92 test_dynamicroles.py test_role_account_acls_multiple_mgmt_servers Success 8.46 test_dynamicroles.py test_role_account_acls Success 8.78 test_dynamicroles.py test_default_role_deletion Success 6.04 test_dynamicroles.py test_04_create_fat_type_disk_offering Success 0.07 test_disk_offerings.py test_03_delete_disk_offering Success 0.04 test_disk_offerings.py test_02_edit_disk_offering Success 0.05 test_disk_offerings.py test_02_create_sparse_type_disk_offering Success 0.08 test_disk_offerings.py test_01_create_disk_offering Success 0.11 test_disk_offerings.py test_deployvm_userdispersing Success 65.98 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userconcentrated Success 106.19 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_firstfit Success 201.69 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userdata_post Success 30.54 test_deploy_vm_with_userdata.py test_deployvm_userdata Success 156.46 test_deploy_vm_with_userdata.py test_02_deploy_vm_root_resize Success 5.96 test_deploy_vm_root_resize.py test_01_deploy_vm_root_resize Success 5.99 test_deploy_vm_root_resize.py test_00_deploy_vm_root_resize Success 6.39 test_deploy_vm_root_resize.py test_deploy_vm_from_iso Success 207.58 test_deploy_vm_iso.py test_DeployVmAntiAffinityGroup Success 282.58 test_affinity_groups.py test_08_resize_volume Skipped 5.12 test_volumes.py test_07_resize_fail Skipped 10.31 test_volumes.py test_06_copy_template Skipped 0.00 test_templates.py test_static_role_account_acls Skipped 0.04 test_staticroles.py test_01_scale_vm Skipped 66.35 test_scale_vm.py test_01_primary_storage_iscsi Skipped 0.04 test_primary_storage.py test_06_copy_iso Skipped 0.00 test_iso.py test_deploy_vgpu_enabled_vm Skipped 0.01 test_deploy_vgpu_enabled_vm.py @blueorangutan test @rhtyd a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests There seems to be some new (intermittent) errors that don't seem to be related to this PR, however I would like to investigate and possibly fix them (in case they are caused some other recent changes to 4.9/master). I'll also kick tests for xenserver: @blueorangutan test centos7 xenserver-65sp1 @rhtyd a Trillian-Jenkins test job (centos7 mgmt + xenserver-65sp1) has been kicked to run smoke tests Trillian test result (tid-1104) Environment: xenserver-65sp1 (x2), Advanced Networking with Mgmt server 7 Total time taken: 49407 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1104-xenserver-65sp1.zip Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py Intermitten failure detected: /marvin/tests/smoke/test_templates.py Intermitten failure detected: /marvin/tests/smoke/test_volumes.py Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py Test completed. 45 look ok, 4 have error(s) Test Result Time (s) Test File test_05_rvpc_multi_tiers Failure 493.10 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Failure 1427.86 test_vpc_redundant.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Failure 585.31 test_vpc_redundant.py test_04_rvpc_privategw_static_routes Failure 846.93 test_privategw_acl.py test_01_create_template Error 80.55 test_templates.py test_02_list_snapshots_with_removed_data_store Error 98.86 test_snapshots.py test_02_list_snapshots_with_removed_data_store Error 103.92 test_snapshots.py test_01_vpc_site2site_vpn Success 395.67 test_vpc_vpn.py test_01_vpc_remote_access_vpn Success 206.46 test_vpc_vpn.py test_01_redundant_vpc_site2site_vpn Success 566.05 test_vpc_vpn.py test_02_VPC_default_routes Success 570.10 test_vpc_router_nics.py test_01_VPC_nics_after_destroy Success 838.70 test_vpc_router_nics.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Success 927.60 test_vpc_redundant.py test_02_redundant_VPC_default_routes Success 1135.49 test_vpc_redundant.py test_09_delete_detached_volume Success 10.45 test_volumes.py test_08_resize_volume Success 115.84 test_volumes.py test_07_resize_fail Success 125.92 test_volumes.py test_06_download_detached_volume Success 20.28 test_volumes.py test_05_detach_volume Success 110.27 test_volumes.py test_04_delete_attached_volume Success 15.17 test_volumes.py test_03_download_attached_volume Success 15.22 test_volumes.py test_02_attach_volume Success 15.87 test_volumes.py test_01_create_volume Success 393.08 test_volumes.py test_03_delete_vm_snapshots Success 280.20 test_vm_snapshots.py test_02_revert_vm_snapshots Success 232.46 test_vm_snapshots.py test_01_create_vm_snapshots Success 130.94 test_vm_snapshots.py test_deploy_vm_multiple Success 317.28 test_vm_life_cycle.py test_deploy_vm Success 0.02 test_vm_life_cycle.py test_advZoneVirtualRouter Success 0.02 test_vm_life_cycle.py test_10_attachAndDetach_iso Success 118.16 test_vm_life_cycle.py test_09_expunge_vm Success 125.14 test_vm_life_cycle.py test_08_migrate_vm Success 90.95 test_vm_life_cycle.py test_07_restore_vm Success 0.07 test_vm_life_cycle.py test_06_destroy_vm Success 15.17 test_vm_life_cycle.py test_03_reboot_vm Success 20.18 test_vm_life_cycle.py test_02_start_vm Success 25.18 test_vm_life_cycle.py test_01_stop_vm Success 35.24 test_vm_life_cycle.py test_CreateTemplateWithDuplicateName Success 196.08 test_templates.py test_08_list_system_templates Success 0.02 test_templates.py test_07_list_public_templates Success 0.03 test_templates.py test_05_template_permissions Success 0.04 test_templates.py test_04_extract_template Success 5.24 test_templates.py test_03_delete_template Success 5.08 test_templates.py test_02_edit_template Success 90.05 test_templates.py test_10_destroy_cpvm Success 196.40 test_ssvm.py test_09_destroy_ssvm Success 234.40 test_ssvm.py test_08_reboot_cpvm Success 156.51 test_ssvm.py test_07_reboot_ssvm Success 304.67 test_ssvm.py test_06_stop_cpvm Success 181.77 test_ssvm.py test_05_stop_ssvm Success 204.90 test_ssvm.py test_04_cpvm_internals Success 1.32 test_ssvm.py test_03_ssvm_internals Success 3.69 test_ssvm.py test_02_list_cpvm_vm Success 0.10 test_ssvm.py test_01_list_sec_storage_vm Success 0.10 test_ssvm.py test_01_snapshot_root_disk Success 36.05 test_snapshots.py test_04_change_offering_small Success 121.29 test_service_offerings.py test_03_delete_service_offering Success 0.03 test_service_offerings.py test_02_edit_service_offering Success 0.04 test_service_offerings.py test_01_create_service_offering Success 0.06 test_service_offerings.py test_02_sys_template_ready Success 0.12 test_secondary_storage.py test_01_sys_vm_start Success 0.14 test_secondary_storage.py test_01_scale_vm Success 5.13 test_scale_vm.py test_09_reboot_router Success 105.55 test_routers.py test_08_start_router Success 70.39 test_routers.py test_07_stop_router Success 20.16 test_routers.py test_06_router_advanced Success 0.04 test_routers.py test_05_router_basic Success 0.03 test_routers.py test_04_restart_network_wo_cleanup Success 5.56 test_routers.py test_03_restart_network_cleanup Success 160.81 test_routers.py test_02_router_internal_adv Success 0.70 test_routers.py test_01_router_internal_basic Success 0.37 test_routers.py test_router_dns_guestipquery Success 48.84 test_router_dns.py test_router_dns_externalipquery Success 0.06 test_router_dns.py test_router_dhcphosts Success 116.97 test_router_dhcphosts.py test_router_dhcp_opts Success 31.23 test_router_dhcphosts.py test_01_updatevolumedetail Success 5.08 test_resource_detail.py test_01_reset_vm_on_reboot Success 201.09 test_reset_vm_on_reboot.py test_createRegion Success 0.03 test_regions.py test_create_pvlan_network Success 5.16 test_pvlan.py test_dedicatePublicIpRange Success 0.27 test_public_ip_range.py test_03_vpc_privategw_restart_vpc_cleanup Success 764.48 test_privategw_acl.py test_02_vpc_privategw_static_routes Success 785.96 test_privategw_acl.py test_01_vpc_privategw_acl Success 252.27 test_privategw_acl.py test_01_primary_storage_nfs Success 38.45 test_primary_storage.py test_01_primary_storage_iscsi Success 77.64 test_primary_storage.py test_createPortablePublicIPRange Success 15.15 test_portable_publicip.py test_createPortablePublicIPAcquire Success 10.29 test_portable_publicip.py test_isolate_network_password_server Success 65.95 test_password_server.py test_UpdateStorageOverProvisioningFactor Success 0.09 test_over_provisioning.py test_oobm_zchange_password Success 25.56 test_outofbandmanagement.py test_oobm_multiple_mgmt_server_ownership Success 16.28 test_outofbandmanagement.py test_oobm_issue_power_status Success 10.19 test_outofbandmanagement.py test_oobm_issue_power_soft Success 10.24 test_outofbandmanagement.py test_oobm_issue_power_reset Success 15.26 test_outofbandmanagement.py test_oobm_issue_power_on Success 15.28 test_outofbandmanagement.py test_oobm_issue_power_off Success 15.28 test_outofbandmanagement.py test_oobm_issue_power_cycle Success 15.29 test_outofbandmanagement.py test_oobm_enabledisable_across_clusterzones Success 102.34 test_outofbandmanagement.py test_oobm_enable_feature_valid Success 0.10 test_outofbandmanagement.py test_oobm_enable_feature_invalid Success 0.07 test_outofbandmanagement.py test_oobm_disable_feature_valid Success 5.13 test_outofbandmanagement.py test_oobm_disable_feature_invalid Success 0.08 test_outofbandmanagement.py test_oobm_configure_invalid_driver Success 0.06 test_outofbandmanagement.py test_oobm_configure_default_driver Success 0.06 test_outofbandmanagement.py test_oobm_background_powerstate_sync Success 29.31 test_outofbandmanagement.py test_extendPhysicalNetworkVlan Success 15.24 test_non_contigiousvlan.py test_01_nic Success 574.15 test_nic.py test_releaseIP Success 297.41 test_network.py test_reboot_router Success 624.00 test_network.py test_public_ip_user_account Success 10.20 test_network.py test_public_ip_admin_account Success 40.21 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Success 86.50 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Success 76.22 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Success 102.65 test_network.py test_delete_account Success 398.07 test_network.py test_02_port_fwd_on_non_src_nat Success 70.68 test_network.py test_01_port_fwd_on_src_nat Success 83.76 test_network.py test_nic_secondaryip_add_remove Success 126.90 test_multipleips_per_nic.py test_list_zones_metrics Success 0.21 test_metrics_api.py test_list_volumes_metrics Success 0.35 test_metrics_api.py test_list_vms_metrics Success 222.91 test_metrics_api.py test_list_pstorage_metrics Success 0.53 test_metrics_api.py test_list_infrastructure_metrics Success 0.55 test_metrics_api.py test_list_hosts_metrics Success 0.41 test_metrics_api.py test_list_clusters_metrics Success 0.62 test_metrics_api.py login_test_saml_user Success 18.09 test_login.py test_assign_and_removal_lb Success 148.48 test_loadbalance.py test_02_create_lb_rule_non_nat Success 207.38 test_loadbalance.py test_01_create_lb_rule_src_nat Success 207.63 test_loadbalance.py test_03_list_snapshots Success 0.06 test_list_ids_parameter.py test_02_list_templates Success 0.03 test_list_ids_parameter.py test_01_list_volumes Success 0.03 test_list_ids_parameter.py test_07_list_default_iso Success 0.05 test_iso.py test_05_iso_permissions Success 0.05 test_iso.py test_04_extract_Iso Success 5.21 test_iso.py test_03_delete_iso Success 95.14 test_iso.py test_02_edit_iso Success 0.04 test_iso.py test_01_create_iso Success 20.74 test_iso.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Success 464.00 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Success 317.96 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Success 1051.89 test_internal_lb.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Success 699.84 test_internal_lb.py test_dedicateGuestVlanRange Success 10.20 test_guest_vlan_range.py test_UpdateConfigParamWithScope Success 0.10 test_global_settings.py test_rolepermission_lifecycle_update Success 5.83 test_dynamicroles.py test_rolepermission_lifecycle_list Success 5.72 test_dynamicroles.py test_rolepermission_lifecycle_delete Success 5.58 test_dynamicroles.py test_rolepermission_lifecycle_create Success 5.59 test_dynamicroles.py test_rolepermission_lifecycle_concurrent_updates Success 5.68 test_dynamicroles.py test_role_lifecycle_update_role_inuse Success 5.67 test_dynamicroles.py test_role_lifecycle_update Success 5.64 test_dynamicroles.py test_role_lifecycle_list Success 5.59 test_dynamicroles.py test_role_lifecycle_delete Success 5.66 test_dynamicroles.py test_role_lifecycle_create Success 5.64 test_dynamicroles.py test_role_inuse_deletion Success 5.58 test_dynamicroles.py test_role_account_acls_multiple_mgmt_servers Success 6.93 test_dynamicroles.py test_role_account_acls Success 7.17 test_dynamicroles.py test_default_role_deletion Success 5.69 test_dynamicroles.py test_04_create_fat_type_disk_offering Success 0.05 test_disk_offerings.py test_03_delete_disk_offering Success 0.08 test_disk_offerings.py test_02_edit_disk_offering Success 0.06 test_disk_offerings.py test_02_create_sparse_type_disk_offering Success 0.07 test_disk_offerings.py test_01_create_disk_offering Success 0.07 test_disk_offerings.py test_deployvm_userdispersing Success 30.49 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userconcentrated Success 80.82 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_firstfit Success 140.91 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userdata_post Success 25.37 test_deploy_vm_with_userdata.py test_deployvm_userdata Success 171.13 test_deploy_vm_with_userdata.py test_02_deploy_vm_root_resize Success 5.56 test_deploy_vm_root_resize.py test_01_deploy_vm_root_resize Success 5.59 test_deploy_vm_root_resize.py test_00_deploy_vm_root_resize Success 5.68 test_deploy_vm_root_resize.py test_deploy_vm_from_iso Success 196.84 test_deploy_vm_iso.py test_DeployVmAntiAffinityGroup Success 151.11 test_affinity_groups.py test_06_copy_template Skipped 0.00 test_templates.py test_static_role_account_acls Skipped 0.02 test_staticroles.py test_06_copy_iso Skipped 0.00 test_iso.py test_deploy_vgpu_enabled_vm Skipped 0.01 test_deploy_vgpu_enabled_vm.py Trillian test result (tid-1105) Environment: vmware-55u3 (x2), Advanced Networking with Mgmt server 7 Total time taken: 53273 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1105-vmware-55u3.zip Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py Test completed. 47 look ok, 2 have error(s) Test Result Time (s) Test File test_04_rvpc_privategw_static_routes Failure 1087.47 test_privategw_acl.py test_02_list_snapshots_with_removed_data_store Error 131.60 test_snapshots.py test_02_list_snapshots_with_removed_data_store Error 136.75 test_snapshots.py test_01_vpc_site2site_vpn Success 578.92 test_vpc_vpn.py test_01_vpc_remote_access_vpn Success 272.61 test_vpc_vpn.py test_01_redundant_vpc_site2site_vpn Success 885.64 test_vpc_vpn.py test_02_VPC_default_routes Success 471.84 test_vpc_router_nics.py test_01_VPC_nics_after_destroy Success 947.31 test_vpc_router_nics.py test_05_rvpc_multi_tiers Success 896.47 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Success 1691.65 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Success 893.04 test_vpc_redundant.py test_02_redundant_VPC_default_routes Success 909.00 test_vpc_redundant.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Success 1605.68 test_vpc_redundant.py test_09_delete_detached_volume Success 36.03 test_volumes.py test_06_download_detached_volume Success 106.02 test_volumes.py test_05_detach_volume Success 105.34 test_volumes.py test_04_delete_attached_volume Success 15.24 test_volumes.py test_03_download_attached_volume Success 20.37 test_volumes.py test_02_attach_volume Success 63.82 test_volumes.py test_01_create_volume Success 525.89 test_volumes.py test_03_delete_vm_snapshots Success 280.29 test_vm_snapshots.py test_02_revert_vm_snapshots Success 237.50 test_vm_snapshots.py test_01_create_vm_snapshots Success 167.11 test_vm_snapshots.py test_deploy_vm_multiple Success 278.17 test_vm_life_cycle.py test_deploy_vm Success 0.04 test_vm_life_cycle.py test_advZoneVirtualRouter Success 0.03 test_vm_life_cycle.py test_10_attachAndDetach_iso Success 26.85 test_vm_life_cycle.py test_09_expunge_vm Success 125.28 test_vm_life_cycle.py test_08_migrate_vm Success 91.31 test_vm_life_cycle.py test_07_restore_vm Success 0.11 test_vm_life_cycle.py test_06_destroy_vm Success 10.18 test_vm_life_cycle.py test_03_reboot_vm Success 5.15 test_vm_life_cycle.py test_02_start_vm Success 25.27 test_vm_life_cycle.py test_01_stop_vm Success 5.14 test_vm_life_cycle.py test_CreateTemplateWithDuplicateName Success 498.59 test_templates.py test_08_list_system_templates Success 0.03 test_templates.py test_07_list_public_templates Success 0.04 test_templates.py test_05_template_permissions Success 0.06 test_templates.py test_04_extract_template Success 15.24 test_templates.py test_03_delete_template Success 5.12 test_templates.py test_02_edit_template Success 90.29 test_templates.py test_01_create_template Success 247.00 test_templates.py test_10_destroy_cpvm Success 327.12 test_ssvm.py test_09_destroy_ssvm Success 268.97 test_ssvm.py test_08_reboot_cpvm Success 186.76 test_ssvm.py test_07_reboot_ssvm Success 158.61 test_ssvm.py test_06_stop_cpvm Success 207.03 test_ssvm.py test_05_stop_ssvm Success 208.88 test_ssvm.py test_04_cpvm_internals Success 1.25 test_ssvm.py test_03_ssvm_internals Success 3.62 test_ssvm.py test_02_list_cpvm_vm Success 0.15 test_ssvm.py test_01_list_sec_storage_vm Success 0.15 test_ssvm.py test_01_snapshot_root_disk Success 61.83 test_snapshots.py test_04_change_offering_small Success 91.92 test_service_offerings.py test_03_delete_service_offering Success 0.04 test_service_offerings.py test_02_edit_service_offering Success 0.09 test_service_offerings.py test_01_create_service_offering Success 0.12 test_service_offerings.py test_02_sys_template_ready Success 0.15 test_secondary_storage.py test_01_sys_vm_start Success 0.19 test_secondary_storage.py test_09_reboot_router Success 176.28 test_routers.py test_08_start_router Success 156.31 test_routers.py test_07_stop_router Success 25.32 test_routers.py test_06_router_advanced Success 0.07 test_routers.py test_05_router_basic Success 0.04 test_routers.py test_04_restart_network_wo_cleanup Success 5.75 test_routers.py test_03_restart_network_cleanup Success 151.27 test_routers.py test_02_router_internal_adv Success 1.06 test_routers.py test_01_router_internal_basic Success 0.62 test_routers.py test_router_dns_guestipquery Success 77.24 test_router_dns.py test_router_dns_externalipquery Success 0.09 test_router_dns.py test_router_dhcphosts Success 191.38 test_router_dhcphosts.py test_router_dhcp_opts Success 27.16 test_router_dhcphosts.py test_01_updatevolumedetail Success 0.13 test_resource_detail.py test_01_reset_vm_on_reboot Success 45.62 test_reset_vm_on_reboot.py test_createRegion Success 0.05 test_regions.py test_create_pvlan_network Success 5.24 test_pvlan.py test_dedicatePublicIpRange Success 0.55 test_public_ip_range.py test_03_vpc_privategw_restart_vpc_cleanup Success 1301.80 test_privategw_acl.py test_02_vpc_privategw_static_routes Success 799.16 test_privategw_acl.py test_01_vpc_privategw_acl Success 198.59 test_privategw_acl.py test_01_primary_storage_nfs Success 39.15 test_primary_storage.py test_createPortablePublicIPRange Success 15.26 test_portable_publicip.py test_createPortablePublicIPAcquire Success 15.52 test_portable_publicip.py test_isolate_network_password_server Success 99.69 test_password_server.py test_UpdateStorageOverProvisioningFactor Success 0.13 test_over_provisioning.py test_oobm_zchange_password Success 25.93 test_outofbandmanagement.py test_oobm_multiple_mgmt_server_ownership Success 16.37 test_outofbandmanagement.py test_oobm_issue_power_status Success 10.36 test_outofbandmanagement.py test_oobm_issue_power_soft Success 15.39 test_outofbandmanagement.py test_oobm_issue_power_reset Success 15.36 test_outofbandmanagement.py test_oobm_issue_power_on Success 15.37 test_outofbandmanagement.py test_oobm_issue_power_off Success 15.37 test_outofbandmanagement.py test_oobm_issue_power_cycle Success 10.34 test_outofbandmanagement.py test_oobm_enabledisable_across_clusterzones Success 92.82 test_outofbandmanagement.py test_oobm_enable_feature_valid Success 0.14 test_outofbandmanagement.py test_oobm_enable_feature_invalid Success 0.10 test_outofbandmanagement.py test_oobm_disable_feature_valid Success 5.20 test_outofbandmanagement.py test_oobm_disable_feature_invalid Success 0.11 test_outofbandmanagement.py test_oobm_configure_invalid_driver Success 0.09 test_outofbandmanagement.py test_oobm_configure_default_driver Success 0.11 test_outofbandmanagement.py test_oobm_background_powerstate_sync Success 23.43 test_outofbandmanagement.py test_extendPhysicalNetworkVlan Success 15.35 test_non_contigiousvlan.py test_01_nic Success 622.63 test_nic.py test_releaseIP Success 400.02 test_network.py test_reboot_router Success 710.96 test_network.py test_public_ip_user_account Success 10.26 test_network.py test_public_ip_admin_account Success 40.31 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Success 77.48 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Success 61.84 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Success 125.79 test_network.py test_delete_account Success 313.39 test_network.py test_02_port_fwd_on_non_src_nat Success 61.14 test_network.py test_01_port_fwd_on_src_nat Success 111.89 test_network.py test_nic_secondaryip_add_remove Success 227.95 test_multipleips_per_nic.py test_list_zones_metrics Success 0.36 test_metrics_api.py test_list_volumes_metrics Success 5.56 test_metrics_api.py test_list_vms_metrics Success 257.37 test_metrics_api.py test_list_pstorage_metrics Success 0.28 test_metrics_api.py test_list_infrastructure_metrics Success 0.42 test_metrics_api.py test_list_hosts_metrics Success 0.52 test_metrics_api.py test_list_clusters_metrics Success 0.39 test_metrics_api.py login_test_saml_user Success 19.42 test_login.py test_assign_and_removal_lb Success 150.29 test_loadbalance.py test_02_create_lb_rule_non_nat Success 207.97 test_loadbalance.py test_01_create_lb_rule_src_nat Success 208.85 test_loadbalance.py test_03_list_snapshots Success 0.10 test_list_ids_parameter.py test_02_list_templates Success 0.05 test_list_ids_parameter.py test_01_list_volumes Success 0.03 test_list_ids_parameter.py test_07_list_default_iso Success 0.06 test_iso.py test_05_iso_permissions Success 0.13 test_iso.py test_04_extract_Iso Success 5.18 test_iso.py test_03_delete_iso Success 95.14 test_iso.py test_02_edit_iso Success 0.06 test_iso.py test_01_create_iso Success 21.01 test_iso.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Success 702.24 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Success 460.02 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Success 1119.03 test_internal_lb.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Success 844.21 test_internal_lb.py test_dedicateGuestVlanRange Success 10.29 test_guest_vlan_range.py test_UpdateConfigParamWithScope Success 0.15 test_global_settings.py test_rolepermission_lifecycle_update Success 6.22 test_dynamicroles.py test_rolepermission_lifecycle_list Success 5.99 test_dynamicroles.py test_rolepermission_lifecycle_delete Success 5.88 test_dynamicroles.py test_rolepermission_lifecycle_create Success 5.90 test_dynamicroles.py test_rolepermission_lifecycle_concurrent_updates Success 6.02 test_dynamicroles.py test_role_lifecycle_update_role_inuse Success 5.91 test_dynamicroles.py test_role_lifecycle_update Success 11.00 test_dynamicroles.py test_role_lifecycle_list Success 6.09 test_dynamicroles.py test_role_lifecycle_delete Success 5.93 test_dynamicroles.py test_role_lifecycle_create Success 5.96 test_dynamicroles.py test_role_inuse_deletion Success 5.87 test_dynamicroles.py test_role_account_acls_multiple_mgmt_servers Success 8.14 test_dynamicroles.py test_role_account_acls Success 8.39 test_dynamicroles.py test_default_role_deletion Success 5.99 test_dynamicroles.py test_04_create_fat_type_disk_offering Success 0.07 test_disk_offerings.py test_03_delete_disk_offering Success 0.06 test_disk_offerings.py test_02_edit_disk_offering Success 0.05 test_disk_offerings.py test_02_create_sparse_type_disk_offering Success 0.08 test_disk_offerings.py test_01_create_disk_offering Success 0.10 test_disk_offerings.py test_deployvm_userdispersing Success 65.96 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userconcentrated Success 136.56 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_firstfit Success 221.86 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userdata_post Success 50.73 test_deploy_vm_with_userdata.py test_deployvm_userdata Success 186.59 test_deploy_vm_with_userdata.py test_02_deploy_vm_root_resize Success 6.14 test_deploy_vm_root_resize.py test_01_deploy_vm_root_resize Success 5.84 test_deploy_vm_root_resize.py test_00_deploy_vm_root_resize Success 6.00 test_deploy_vm_root_resize.py test_deploy_vm_from_iso Success 252.86 test_deploy_vm_iso.py test_DeployVmAntiAffinityGroup Success 307.74 test_affinity_groups.py test_08_resize_volume Skipped 5.13 test_volumes.py test_07_resize_fail Skipped 10.30 test_volumes.py test_06_copy_template Skipped 0.00 test_templates.py test_static_role_account_acls Skipped 0.02 test_staticroles.py test_01_scale_vm Skipped 66.35 test_scale_vm.py test_01_primary_storage_iscsi Skipped 0.04 test_primary_storage.py test_06_copy_iso Skipped 0.00 test_iso.py test_deploy_vgpu_enabled_vm Skipped 0.01 test_deploy_vgpu_enabled_vm.py VMware and KVM look okay. Some new errors in xensever, will re-kick to see if we reproduce them again; @blueorangutan test centos7 xenserver-65sp1 @rhtyd a Trillian-Jenkins test job (centos7 mgmt + xenserver-65sp1) has been kicked to run smoke tests Trillian test result (tid-1114) Environment: xenserver-65sp1 (x2), Advanced Networking with Mgmt server 7 Total time taken: 46141 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1114-xenserver-65sp1.zip Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py Test completed. 46 look ok, 3 have error(s) Test Result Time (s) Test File test_05_rvpc_multi_tiers Failure 579.18 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Failure 1385.93 test_vpc_redundant.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Failure 585.29 test_vpc_redundant.py test_04_rvpc_privategw_static_routes Failure 833.77 test_privategw_acl.py test_02_list_snapshots_with_removed_data_store Error 89.31 test_snapshots.py test_02_list_snapshots_with_removed_data_store Error 94.40 test_snapshots.py test_01_vpc_site2site_vpn Success 341.00 test_vpc_vpn.py test_01_vpc_remote_access_vpn Success 166.68 test_vpc_vpn.py test_01_redundant_vpc_site2site_vpn Success 567.94 test_vpc_vpn.py test_02_VPC_default_routes Success 394.93 test_vpc_router_nics.py test_01_VPC_nics_after_destroy Success 703.29 test_vpc_router_nics.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Success 953.46 test_vpc_redundant.py test_02_redundant_VPC_default_routes Success 1101.33 test_vpc_redundant.py test_09_delete_detached_volume Success 15.64 test_volumes.py test_08_resize_volume Success 85.79 test_volumes.py test_07_resize_fail Success 100.88 test_volumes.py test_06_download_detached_volume Success 20.30 test_volumes.py test_05_detach_volume Success 100.26 test_volumes.py test_04_delete_attached_volume Success 10.23 test_volumes.py test_03_download_attached_volume Success 15.26 test_volumes.py test_02_attach_volume Success 10.77 test_volumes.py test_01_create_volume Success 392.58 test_volumes.py test_03_delete_vm_snapshots Success 280.25 test_vm_snapshots.py test_02_revert_vm_snapshots Success 181.32 test_vm_snapshots.py test_01_create_vm_snapshots Success 105.84 test_vm_snapshots.py test_deploy_vm_multiple Success 227.38 test_vm_life_cycle.py test_deploy_vm Success 0.03 test_vm_life_cycle.py test_advZoneVirtualRouter Success 0.02 test_vm_life_cycle.py test_10_attachAndDetach_iso Success 41.88 test_vm_life_cycle.py test_09_expunge_vm Success 125.30 test_vm_life_cycle.py test_08_migrate_vm Success 81.20 test_vm_life_cycle.py test_07_restore_vm Success 0.10 test_vm_life_cycle.py test_06_destroy_vm Success 15.18 test_vm_life_cycle.py test_03_reboot_vm Success 20.21 test_vm_life_cycle.py test_02_start_vm Success 25.24 test_vm_life_cycle.py test_01_stop_vm Success 30.24 test_vm_life_cycle.py test_CreateTemplateWithDuplicateName Success 171.15 test_templates.py test_08_list_system_templates Success 0.03 test_templates.py test_07_list_public_templates Success 0.04 test_templates.py test_05_template_permissions Success 0.05 test_templates.py test_04_extract_template Success 5.13 test_templates.py test_03_delete_template Success 5.11 test_templates.py test_02_edit_template Success 90.13 test_templates.py test_01_create_template Success 95.67 test_templates.py test_10_destroy_cpvm Success 196.53 test_ssvm.py test_09_destroy_ssvm Success 198.78 test_ssvm.py test_08_reboot_cpvm Success 141.41 test_ssvm.py test_07_reboot_ssvm Success 143.74 test_ssvm.py test_06_stop_cpvm Success 191.61 test_ssvm.py test_05_stop_ssvm Success 163.93 test_ssvm.py test_04_cpvm_internals Success 0.91 test_ssvm.py test_03_ssvm_internals Success 3.55 test_ssvm.py test_02_list_cpvm_vm Success 0.12 test_ssvm.py test_01_list_sec_storage_vm Success 0.16 test_ssvm.py test_01_snapshot_root_disk Success 21.17 test_snapshots.py test_04_change_offering_small Success 119.12 test_service_offerings.py test_03_delete_service_offering Success 0.04 test_service_offerings.py test_02_edit_service_offering Success 0.05 test_service_offerings.py test_01_create_service_offering Success 0.11 test_service_offerings.py test_02_sys_template_ready Success 0.12 test_secondary_storage.py test_01_sys_vm_start Success 0.18 test_secondary_storage.py test_01_scale_vm Success 5.23 test_scale_vm.py test_09_reboot_router Success 65.48 test_routers.py test_08_start_router Success 45.36 test_routers.py test_07_stop_router Success 15.18 test_routers.py test_06_router_advanced Success 0.05 test_routers.py test_05_router_basic Success 0.04 test_routers.py test_04_restart_network_wo_cleanup Success 5.52 test_routers.py test_03_restart_network_cleanup Success 115.79 test_routers.py test_02_router_internal_adv Success 0.66 test_routers.py test_01_router_internal_basic Success 0.41 test_routers.py test_router_dns_guestipquery Success 48.86 test_router_dns.py test_router_dns_externalipquery Success 0.08 test_router_dns.py test_router_dhcphosts Success 107.49 test_router_dhcphosts.py test_router_dhcp_opts Success 31.22 test_router_dhcphosts.py test_01_updatevolumedetail Success 5.11 test_resource_detail.py test_01_reset_vm_on_reboot Success 45.42 test_reset_vm_on_reboot.py test_createRegion Success 0.04 test_regions.py test_create_pvlan_network Success 5.21 test_pvlan.py test_dedicatePublicIpRange Success 0.41 test_public_ip_range.py test_03_vpc_privategw_restart_vpc_cleanup Success 998.03 test_privategw_acl.py test_02_vpc_privategw_static_routes Success 732.25 test_privategw_acl.py test_01_vpc_privategw_acl Success 187.76 test_privategw_acl.py test_01_primary_storage_nfs Success 39.71 test_primary_storage.py test_01_primary_storage_iscsi Success 79.23 test_primary_storage.py test_createPortablePublicIPRange Success 15.20 test_portable_publicip.py test_createPortablePublicIPAcquire Success 15.46 test_portable_publicip.py test_isolate_network_password_server Success 36.32 test_password_server.py test_UpdateStorageOverProvisioningFactor Success 0.12 test_over_provisioning.py test_oobm_zchange_password Success 30.71 test_outofbandmanagement.py test_oobm_multiple_mgmt_server_ownership Success 11.29 test_outofbandmanagement.py test_oobm_issue_power_status Success 10.25 test_outofbandmanagement.py test_oobm_issue_power_soft Success 15.39 test_outofbandmanagement.py test_oobm_issue_power_reset Success 15.37 test_outofbandmanagement.py test_oobm_issue_power_on Success 15.34 test_outofbandmanagement.py test_oobm_issue_power_off Success 15.32 test_outofbandmanagement.py test_oobm_issue_power_cycle Success 15.56 test_outofbandmanagement.py test_oobm_enabledisable_across_clusterzones Success 92.69 test_outofbandmanagement.py test_oobm_enable_feature_valid Success 5.14 test_outofbandmanagement.py test_oobm_enable_feature_invalid Success 0.09 test_outofbandmanagement.py test_oobm_disable_feature_valid Success 5.17 test_outofbandmanagement.py test_oobm_disable_feature_invalid Success 5.17 test_outofbandmanagement.py test_oobm_configure_invalid_driver Success 0.08 test_outofbandmanagement.py test_oobm_configure_default_driver Success 0.07 test_outofbandmanagement.py test_oobm_background_powerstate_sync Success 23.41 test_outofbandmanagement.py test_extendPhysicalNetworkVlan Success 15.28 test_non_contigiousvlan.py test_01_nic Success 650.19 test_nic.py test_releaseIP Success 318.29 test_network.py test_reboot_router Success 589.53 test_network.py test_public_ip_user_account Success 10.29 test_network.py test_public_ip_admin_account Success 40.32 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Success 86.49 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Success 76.56 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Success 100.54 test_network.py test_delete_account Success 262.65 test_network.py test_02_port_fwd_on_non_src_nat Success 70.78 test_network.py test_01_port_fwd_on_src_nat Success 83.82 test_network.py test_nic_secondaryip_add_remove Success 252.73 test_multipleips_per_nic.py test_list_zones_metrics Success 0.32 test_metrics_api.py test_list_volumes_metrics Success 5.50 test_metrics_api.py test_list_vms_metrics Success 126.40 test_metrics_api.py test_list_pstorage_metrics Success 0.38 test_metrics_api.py test_list_infrastructure_metrics Success 0.51 test_metrics_api.py test_list_hosts_metrics Success 0.40 test_metrics_api.py test_list_clusters_metrics Success 0.36 test_metrics_api.py login_test_saml_user Success 19.15 test_login.py test_assign_and_removal_lb Success 148.83 test_loadbalance.py test_02_create_lb_rule_non_nat Success 207.47 test_loadbalance.py test_01_create_lb_rule_src_nat Success 208.09 test_loadbalance.py test_03_list_snapshots Success 0.05 test_list_ids_parameter.py test_02_list_templates Success 0.04 test_list_ids_parameter.py test_01_list_volumes Success 0.03 test_list_ids_parameter.py test_07_list_default_iso Success 0.08 test_iso.py test_05_iso_permissions Success 0.09 test_iso.py test_04_extract_Iso Success 5.14 test_iso.py test_03_delete_iso Success 95.21 test_iso.py test_02_edit_iso Success 0.07 test_iso.py test_01_create_iso Success 21.00 test_iso.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Success 494.53 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Success 324.00 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Success 1073.40 test_internal_lb.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Success 722.96 test_internal_lb.py test_dedicateGuestVlanRange Success 10.29 test_guest_vlan_range.py test_UpdateConfigParamWithScope Success 0.13 test_global_settings.py test_rolepermission_lifecycle_update Success 6.24 test_dynamicroles.py test_rolepermission_lifecycle_list Success 6.03 test_dynamicroles.py test_rolepermission_lifecycle_delete Success 5.93 test_dynamicroles.py test_rolepermission_lifecycle_create Success 6.14 test_dynamicroles.py test_rolepermission_lifecycle_concurrent_updates Success 6.08 test_dynamicroles.py test_role_lifecycle_update_role_inuse Success 5.97 test_dynamicroles.py test_role_lifecycle_update Success 10.97 test_dynamicroles.py test_role_lifecycle_list Success 5.88 test_dynamicroles.py test_role_lifecycle_delete Success 10.94 test_dynamicroles.py test_role_lifecycle_create Success 5.94 test_dynamicroles.py test_role_inuse_deletion Success 5.89 test_dynamicroles.py test_role_account_acls_multiple_mgmt_servers Success 8.30 test_dynamicroles.py test_role_account_acls Success 8.34 test_dynamicroles.py test_default_role_deletion Success 6.01 test_dynamicroles.py test_04_create_fat_type_disk_offering Success 0.07 test_disk_offerings.py test_03_delete_disk_offering Success 0.04 test_disk_offerings.py test_02_edit_disk_offering Success 0.05 test_disk_offerings.py test_02_create_sparse_type_disk_offering Success 0.06 test_disk_offerings.py test_01_create_disk_offering Success 0.08 test_disk_offerings.py test_deployvm_userdispersing Success 121.17 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userconcentrated Success 50.81 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_firstfit Success 221.68 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userdata_post Success 20.48 test_deploy_vm_with_userdata.py test_deployvm_userdata Success 217.05 test_deploy_vm_with_userdata.py test_02_deploy_vm_root_resize Success 5.84 test_deploy_vm_root_resize.py test_01_deploy_vm_root_resize Success 5.83 test_deploy_vm_root_resize.py test_00_deploy_vm_root_resize Success 5.97 test_deploy_vm_root_resize.py test_deploy_vm_from_iso Success 333.13 test_deploy_vm_iso.py test_DeployVmAntiAffinityGroup Success 141.59 test_affinity_groups.py test_06_copy_template Skipped 0.00 test_templates.py test_static_role_account_acls Skipped 0.02 test_staticroles.py test_06_copy_iso Skipped 0.00 test_iso.py test_deploy_vgpu_enabled_vm Skipped 0.00 test_deploy_vgpu_enabled_vm.py Test results LGTM, I don't see any new regressions. Some of the failing tests are known intermittent failures around rvpcs, snapshot/disk (snapshot failures/tests are fixed on master, backported to 4.9 here: #2052 ) etc. Pinging - @DaanHoogland @abhinandanprateek @ustcweizhou - please review and see if this is okay? /cc @karuturi @rhtyd Please add some integration tests for hard shutdown scenario. Also consider the case of restoring a VM after hard shutdown. The CS documentation should be updated to point to the relevant HV specific hard shutdown pros and cons. PS: A similar thing was done in Cloudplatform but for the Destroy VM API. See if you would like to include that scenario as well in this PR or want it to be done separately. @koushik-das I've added a stop_vm_forced test case in the vm lifecycle tests however it will difficult to validate a hard-shutdown scenarios. It's an operational improvement and I'm not sure what we will validate/verify when testing a case of starting a hard shutdown VM? CloudStack api docs will automatically be generated and updated during the release/website-update. If needed, we can mention this change in the release notes under 'what's new/changed' etc. I see that the destroy VM codebase already uses a global setting to force shutdown the VM on destroy -- however, in ACS the global setting vm.destroy.forcestop is not honoured to force stop VMs while destroying them, if it is set to true. This changes fixes that. @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos6 ✔centos7 ✔debian. JID-743 @blueorangutan test centos7 vmware-55u3 @rhtyd a Trillian-Jenkins test job (centos7 mgmt + vmware-55u3) has been kicked to run smoke tests Trillian test result (tid-1128) Environment: vmware-55u3 (x2), Advanced Networking with Mgmt server 7 Total time taken: 45568 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1128-vmware-55u3.zip Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py Test completed. 47 look ok, 2 have error(s) Test Result Time (s) Test File test_04_rvpc_privategw_static_routes Failure 1029.96 test_privategw_acl.py test_02_list_snapshots_with_removed_data_store Error 131.16 test_snapshots.py test_02_list_snapshots_with_removed_data_store Error 136.24 test_snapshots.py test_01_vpc_site2site_vpn Success 461.84 test_vpc_vpn.py test_01_vpc_remote_access_vpn Success 186.84 test_vpc_vpn.py test_01_redundant_vpc_site2site_vpn Success 673.22 test_vpc_vpn.py test_02_VPC_default_routes Success 571.00 test_vpc_router_nics.py test_01_VPC_nics_after_destroy Success 867.93 test_vpc_router_nics.py test_05_rvpc_multi_tiers Success 794.01 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Success 1665.02 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Success 867.67 test_vpc_redundant.py test_02_redundant_VPC_default_routes Success 785.84 test_vpc_redundant.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Success 1451.45 test_vpc_redundant.py test_09_delete_detached_volume Success 35.93 test_volumes.py test_06_download_detached_volume Success 95.77 test_volumes.py test_05_detach_volume Success 110.32 test_volumes.py test_04_delete_attached_volume Success 20.26 test_volumes.py test_03_download_attached_volume Success 25.36 test_volumes.py test_02_attach_volume Success 58.87 test_volumes.py test_01_create_volume Success 520.70 test_volumes.py test_03_delete_vm_snapshots Success 280.31 test_vm_snapshots.py test_02_revert_vm_snapshots Success 205.01 test_vm_snapshots.py test_01_create_vm_snapshots Success 161.77 test_vm_snapshots.py test_deploy_vm_multiple Success 302.90 test_vm_life_cycle.py test_deploy_vm Success 0.03 test_vm_life_cycle.py test_advZoneVirtualRouter Success 0.02 test_vm_life_cycle.py test_10_attachAndDetach_iso Success 26.85 test_vm_life_cycle.py test_09_expunge_vm Success 125.23 test_vm_life_cycle.py test_08_migrate_vm Success 81.21 test_vm_life_cycle.py test_07_restore_vm Success 0.10 test_vm_life_cycle.py test_06_destroy_vm Success 10.15 test_vm_life_cycle.py test_03_reboot_vm Success 5.17 test_vm_life_cycle.py test_02_start_vm Success 25.26 test_vm_life_cycle.py test_01_stop_vm_forced Success 5.14 test_vm_life_cycle.py test_01_stop_vm Success 10.15 test_vm_life_cycle.py test_CreateTemplateWithDuplicateName Success 377.36 test_templates.py test_08_list_system_templates Success 0.03 test_templates.py test_07_list_public_templates Success 0.04 test_templates.py test_05_template_permissions Success 0.05 test_templates.py test_04_extract_template Success 25.38 test_templates.py test_03_delete_template Success 5.11 test_templates.py test_02_edit_template Success 90.14 test_templates.py test_01_create_template Success 196.28 test_templates.py test_10_destroy_cpvm Success 241.81 test_ssvm.py test_09_destroy_ssvm Success 268.86 test_ssvm.py test_08_reboot_cpvm Success 156.54 test_ssvm.py test_07_reboot_ssvm Success 158.51 test_ssvm.py test_06_stop_cpvm Success 206.94 test_ssvm.py test_05_stop_ssvm Success 178.76 test_ssvm.py test_04_cpvm_internals Success 1.17 test_ssvm.py test_03_ssvm_internals Success 3.91 test_ssvm.py test_02_list_cpvm_vm Success 0.11 test_ssvm.py test_01_list_sec_storage_vm Success 0.12 test_ssvm.py test_01_snapshot_root_disk Success 31.28 test_snapshots.py test_04_change_offering_small Success 101.96 test_service_offerings.py test_03_delete_service_offering Success 0.04 test_service_offerings.py test_02_edit_service_offering Success 0.08 test_service_offerings.py test_01_create_service_offering Success 0.19 test_service_offerings.py test_02_sys_template_ready Success 0.13 test_secondary_storage.py test_01_sys_vm_start Success 0.17 test_secondary_storage.py test_09_reboot_router Success 166.17 test_routers.py test_08_start_router Success 140.92 test_routers.py test_07_stop_router Success 25.24 test_routers.py test_06_router_advanced Success 0.05 test_routers.py test_05_router_basic Success 0.05 test_routers.py test_04_restart_network_wo_cleanup Success 5.67 test_routers.py test_03_restart_network_cleanup Success 171.16 test_routers.py test_02_router_internal_adv Success 1.02 test_routers.py test_01_router_internal_basic Success 0.57 test_routers.py test_router_dns_guestipquery Success 76.77 test_router_dns.py test_router_dns_externalipquery Success 0.08 test_router_dns.py test_router_dhcphosts Success 149.10 test_router_dhcphosts.py test_router_dhcp_opts Success 21.71 test_router_dhcphosts.py test_01_updatevolumedetail Success 0.08 test_resource_detail.py test_01_reset_vm_on_reboot Success 35.38 test_reset_vm_on_reboot.py test_createRegion Success 0.04 test_regions.py test_create_pvlan_network Success 5.25 test_pvlan.py test_dedicatePublicIpRange Success 0.47 test_public_ip_range.py test_03_vpc_privategw_restart_vpc_cleanup Success 1113.47 test_privategw_acl.py test_02_vpc_privategw_static_routes Success 742.65 test_privategw_acl.py test_01_vpc_privategw_acl Success 208.06 test_privategw_acl.py test_01_primary_storage_nfs Success 38.76 test_primary_storage.py test_createPortablePublicIPRange Success 15.21 test_portable_publicip.py test_createPortablePublicIPAcquire Success 15.44 test_portable_publicip.py test_isolate_network_password_server Success 94.41 test_password_server.py test_UpdateStorageOverProvisioningFactor Success 0.12 test_over_provisioning.py test_oobm_zchange_password Success 30.75 test_outofbandmanagement.py test_oobm_multiple_mgmt_server_ownership Success 16.35 test_outofbandmanagement.py test_oobm_issue_power_status Success 10.28 test_outofbandmanagement.py test_oobm_issue_power_soft Success 15.41 test_outofbandmanagement.py test_oobm_issue_power_reset Success 15.35 test_outofbandmanagement.py test_oobm_issue_power_on Success 15.44 test_outofbandmanagement.py test_oobm_issue_power_off Success 15.33 test_outofbandmanagement.py test_oobm_issue_power_cycle Success 15.36 test_outofbandmanagement.py test_oobm_enabledisable_across_clusterzones Success 87.60 test_outofbandmanagement.py test_oobm_enable_feature_valid Success 5.17 test_outofbandmanagement.py test_oobm_enable_feature_invalid Success 0.11 test_outofbandmanagement.py test_oobm_disable_feature_valid Success 5.18 test_outofbandmanagement.py test_oobm_disable_feature_invalid Success 0.10 test_outofbandmanagement.py test_oobm_configure_invalid_driver Success 0.07 test_outofbandmanagement.py test_oobm_configure_default_driver Success 0.08 test_outofbandmanagement.py test_oobm_background_powerstate_sync Success 23.40 test_outofbandmanagement.py test_extendPhysicalNetworkVlan Success 15.39 test_non_contigiousvlan.py test_01_nic Success 510.38 test_nic.py test_releaseIP Success 353.73 test_network.py test_reboot_router Success 715.72 test_network.py test_public_ip_user_account Success 10.28 test_network.py test_public_ip_admin_account Success 40.31 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Success 76.92 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Success 61.82 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Success 125.61 test_network.py test_delete_account Success 343.41 test_network.py test_02_port_fwd_on_non_src_nat Success 55.77 test_network.py test_01_port_fwd_on_src_nat Success 111.83 test_network.py test_nic_secondaryip_add_remove Success 222.67 test_multipleips_per_nic.py test_list_zones_metrics Success 0.27 test_metrics_api.py test_list_volumes_metrics Success 5.46 test_metrics_api.py test_list_vms_metrics Success 242.35 test_metrics_api.py test_list_pstorage_metrics Success 0.38 test_metrics_api.py test_list_infrastructure_metrics Success 0.51 test_metrics_api.py test_list_hosts_metrics Success 0.49 test_metrics_api.py test_list_clusters_metrics Success 0.34 test_metrics_api.py login_test_saml_user Success 19.42 test_login.py test_assign_and_removal_lb Success 148.57 test_loadbalance.py test_02_create_lb_rule_non_nat Success 207.26 test_loadbalance.py test_01_create_lb_rule_src_nat Success 208.51 test_loadbalance.py test_03_list_snapshots Success 0.09 test_list_ids_parameter.py test_02_list_templates Success 0.07 test_list_ids_parameter.py test_01_list_volumes Success 0.03 test_list_ids_parameter.py test_07_list_default_iso Success 0.06 test_iso.py test_05_iso_permissions Success 0.06 test_iso.py test_04_extract_Iso Success 5.18 test_iso.py test_03_delete_iso Success 95.16 test_iso.py test_02_edit_iso Success 0.06 test_iso.py test_01_create_iso Success 21.12 test_iso.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Success 608.01 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Success 404.84 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Success 1120.61 test_internal_lb.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Success 862.56 test_internal_lb.py test_dedicateGuestVlanRange Success 10.30 test_guest_vlan_range.py test_UpdateConfigParamWithScope Success 0.15 test_global_settings.py test_rolepermission_lifecycle_update Success 6.23 test_dynamicroles.py test_rolepermission_lifecycle_list Success 5.97 test_dynamicroles.py test_rolepermission_lifecycle_delete Success 5.86 test_dynamicroles.py test_rolepermission_lifecycle_create Success 5.91 test_dynamicroles.py test_rolepermission_lifecycle_concurrent_updates Success 6.04 test_dynamicroles.py test_role_lifecycle_update_role_inuse Success 5.94 test_dynamicroles.py test_role_lifecycle_update Success 11.00 test_dynamicroles.py test_role_lifecycle_list Success 5.94 test_dynamicroles.py test_role_lifecycle_delete Success 10.93 test_dynamicroles.py test_role_lifecycle_create Success 5.93 test_dynamicroles.py test_role_inuse_deletion Success 5.91 test_dynamicroles.py test_role_account_acls_multiple_mgmt_servers Success 8.35 test_dynamicroles.py test_role_account_acls Success 8.23 test_dynamicroles.py test_default_role_deletion Success 5.97 test_dynamicroles.py test_04_create_fat_type_disk_offering Success 0.07 test_disk_offerings.py test_03_delete_disk_offering Success 0.04 test_disk_offerings.py test_02_edit_disk_offering Success 0.05 test_disk_offerings.py test_02_create_sparse_type_disk_offering Success 0.08 test_disk_offerings.py test_01_create_disk_offering Success 0.10 test_disk_offerings.py test_deployvm_userdispersing Success 71.11 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userconcentrated Success 126.25 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_firstfit Success 292.28 test_deploy_vms_with_varied_deploymentplanners.py test_deployvm_userdata_post Success 35.56 test_deploy_vm_with_userdata.py test_deployvm_userdata Success 236.82 test_deploy_vm_with_userdata.py test_02_deploy_vm_root_resize Success 5.86 test_deploy_vm_root_resize.py test_01_deploy_vm_root_resize Success 5.87 test_deploy_vm_root_resize.py test_00_deploy_vm_root_resize Success 6.10 test_deploy_vm_root_resize.py test_deploy_vm_from_iso Success 222.52 test_deploy_vm_iso.py test_DeployVmAntiAffinityGroup Success 196.99 test_affinity_groups.py test_08_resize_volume Skipped 10.14 test_volumes.py test_07_resize_fail Skipped 15.32 test_volumes.py test_06_copy_template Skipped 0.00 test_templates.py test_static_role_account_acls Skipped 0.03 test_staticroles.py test_01_scale_vm Skipped 66.34 test_scale_vm.py test_01_primary_storage_iscsi Skipped 0.03 test_primary_storage.py test_06_copy_iso Skipped 0.00 test_iso.py test_deploy_vgpu_enabled_vm Skipped 0.00 test_deploy_vgpu_enabled_vm.py Test LGTM. This is ready for merge, has enough test results, a new marvin test, reviews, LGTMs. /cc @karuturi Thanks @rhtyd. Code changes LGTM @karuturi can you consider reviewing/merging this? this has several test results and lgtms @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos6 ✔centos7 ✔debian. JID-757 @karuturi thanks for merging, however as @borisstoyanov has pointed out, there seems to be regression caused due to some refactoring changes around stopVM code for KVM. I'll send a separate FR swiftly. @karuturi here's the reboot/regression fix: https://github.com/apache/cloudstack/pull/2135/files @borisstoyanov please review/verify ^^
gharchive/pull-request
2017-05-16T09:30:05
2025-04-01T04:33:27.493428
{ "authors": [ "blueorangutan", "koushik-das", "remibergsma", "resmo", "rhtyd", "ustcweizhou", "wido" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/2108", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
548935717
Display capability info in listNetwork response Description Display some properties (eg RedundantRouter) in listNetworks response which are initializd in VirtualRouterElement.setCapabilities() Types of changes [ ] Breaking change (fix or feature that would cause existing functionality to change) [ ] New feature (non-breaking change which adds functionality) [ ] Bug fix (non-breaking change which fixes an issue) [X] Enhancement (improves an existing feature and functionality) [ ] Cleanup (Code refactoring and cleanup, that may add test cases) Screenshots (if appropriate): How Has This Been Tested? listNetworks response for StaticNat service before the change { "name": "StaticNat" }, listNetworks response for StaticNat after the change { "capability": [ { "canchooseservicecapability": false, "name": "ElasticIp", "value": "false" }, { "canchooseservicecapability": false, "name": "AssociatePublicIP", "value": "true" } ], "name": "StaticNat", "provider": [ { "name": "VirtualRouter" } ] }, @ravening Moving to 4.16 based on base branch. Please update if needed @ravening Moving to 4.16 based on base branch. Please update if needed @ravening Moving to 4.16 based on base branch. Please update if needed made it 4.15 (didn't try 4.14, making assumptions) @ravening Moving to 4.16 based on base branch. Please update if needed made it 4.15 (didn't try 4.14, making assumptions) @blueorangutan package @shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2781 @blueorangutan package @shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2787 Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2797 Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2800 Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2809 Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 29 @blueorangutan package @Pearl1594 a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 239 @Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests Trillian test result (tid-252) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 53122 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t252-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py Smoke tests completed. 76 look OK, 10 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_DeployVmAntiAffinityGroup_in_project Error 58.94 test_affinity_groups_projects.py test_DeployVmAntiAffinityGroup Error 36.39 test_affinity_groups.py test_04_rvpc_privategw_static_routes Error 0.44 test_privategw_acl.py test_04_deploy_and_scale_kubernetes_cluster Failure 23.64 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 34.23 test_kubernetes_clusters.py ContextSuite context=TestRAMCPUResourceAccounting>:setup Error 0.00 test_resource_accounting.py ContextSuite context=TestRouterDHCPHosts>:setup Error 0.00 test_router_dhcphosts.py ContextSuite context=TestRouterDHCPOpts>:setup Error 0.00 test_router_dhcphosts.py ContextSuite context=TestIsolatedNetworks>:setup Error 0.00 test_routers_network_ops.py ContextSuite context=TestRedundantIsolateNetworks>:setup Error 0.00 test_routers_network_ops.py ContextSuite context=TestCpuCapServiceOfferings>:setup Error 0.00 test_service_offerings.py ContextSuite context=TestServiceOfferings>:setup Error 0.10 test_service_offerings.py ContextSuite context=TestSnapshotRootDisk>:setup Error 0.00 test_snapshots.py test_hostha_kvm_host_degraded Failure 677.99 test_hostha_kvm.py test_hostha_kvm_host_fencing Failure 644.32 test_hostha_kvm.py test_hostha_kvm_host_recovering Failure 648.34 test_hostha_kvm.py @blueorangutan test @Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests Trillian test result (tid-258) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 40933 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t258-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_virtio_scsi_vm.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py Smoke tests completed. 82 look OK, 4 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_DeployVmAntiAffinityGroup_in_project Error 59.90 test_affinity_groups_projects.py test_DeployVmAntiAffinityGroup Error 35.36 test_affinity_groups.py test_04_deploy_and_scale_kubernetes_cluster Failure 24.69 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 35.16 test_kubernetes_clusters.py test_hostha_kvm_host_degraded Failure 678.33 test_hostha_kvm.py test_hostha_kvm_host_fencing Failure 643.83 test_hostha_kvm.py test_hostha_kvm_host_recovering Failure 643.73 test_hostha_kvm.py @blueorangutan package @Pearl1594 a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 272 @blueorangutan test @Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests Code LGTM. verified the output of listnetworks. Waiting for the test results. Trillian test result (tid-280) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 46771 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t280-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_usage.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Smoke tests completed. 81 look OK, 5 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 85.98 test_kubernetes_clusters.py test_01_isolate_network_FW_PF_default_routes_egress_true Failure 65.59 test_routers_network_ops.py test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false Failure 121.82 test_routers_network_ops.py ContextSuite context=TestISOUsage>:setup Error 0.00 test_usage.py test_01_migrate_VM_and_root_volume Error 70.32 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 50.05 test_vm_life_cycle.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Failure 494.01 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Failure 481.57 test_vpc_redundant.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 321 @blueorangutan test @Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests Trillian test result (tid-347) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36633 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t347-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 69.95 test_kubernetes_clusters.py test_01_migrate_VM_and_root_volume Error 69.26 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 50.04 test_vm_life_cycle.py @ravening @weizhouapache cc @Pearl1594 should this information be restricted to just admins or to all users? Re-ping @weizhouapache @ravening Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 101 Trillian test result (tid-766) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32216 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t766-kvm-centos7.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ping @ravening - can you address outstanding remarks and advise if this is ready for merging? I see last rounds of tests have passeed, on the design side - do we need those capabilities as part of the list API response (even for non-admin users?) re-kick travis @ravening @weizhouapache cc @Pearl1594 should this information be restricted to just admins or to all users? @rhtyd this can be displayed to both users and admins @blueorangutan package @DaanHoogland a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: el7 :heavy_check_mark: el8 :heavy_check_mark: debian. SL-JID 947 @blueorangutan test @DaanHoogland a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests Trillian test result (tid-1735) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32476 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t1735-kvm-centos7.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Ping @Pearl1594 @DaanHoogland can you review/test this, would it cause any leaks or regression (for UI?) Moving to 4.16, as this introduces changes in API response - we'll have more time to review/test this against UI/clients. @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_multiplication_x: el7 :heavy_check_mark: el8 :heavy_multiplication_x: debian :heavy_check_mark: suse15. SL-JID 1161 @blueorangutan package @sureshanaparti a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_multiplication_x: el7 :heavy_check_mark: el8 :heavy_check_mark: debian :heavy_check_mark: suse15. SL-JID 1173 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: el7 :heavy_check_mark: el8 :heavy_check_mark: debian :heavy_check_mark: suse15. SL-JID 1238 @blueorangutan test @sureshanaparti a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests @blueorangutan test @rhtyd a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests @blueorangutan test matrix @sureshanaparti a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-2122) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36862 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t2122-kvm-centos7.zip Smoke tests completed. 89 look OK, 0 have errors Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-2121) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 37333 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t2121-xenserver-71.zip Smoke tests completed. 89 look OK, 0 have errors Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-2123) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 38750 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t2123-vmware-65u2.zip Smoke tests completed. 89 look OK, 0 have errors Only failed tests results shown below: Test Result Time (s) Test File Hi @Pearl1594 can you test this PR for any regressions in the UI given the change on the networks response? I'll have a look at it @nvazquez LGTM
gharchive/pull-request
2020-01-13T13:37:48
2025-04-01T04:33:27.572532
{ "authors": [ "DaanHoogland", "Pearl1594", "blueorangutan", "nvazquez", "ravening", "rhtyd", "shwstppr", "sureshanaparti" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/3804", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
811830319
[HEALTH][4.15] Health Check Run Dummy PR for tracking health on 4.15 branch @blueorangutan package @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos7 ✔centos8 ✔debian. JID-2814 @blueorangutan test matrix @DaanHoogland a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-3606) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 31484 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t3606-kvm-centos7.zip Smoke tests completed. 86 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-3605) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32355 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t3605-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 85 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_scale_vm Failure 9.28 test_scale_vm.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos7 ✔centos8 ✔debian. JID-2820 @blueorangutan package @blueorangutan package @shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2845 @blueorangutan package @shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✔centos7 ✔centos8 ✔debian. JID-2856 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2883 [S] Trillian test result (tid-75) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 34945 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t75-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 85 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_migrate_VM_and_root_volume Error 77.47 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 50.19 test_vm_life_cycle.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S] Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 86 [S] Trillian test result (tid-74) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 319816 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t74-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py Intermittent failure detected: /marvin/tests/smoke/test_async_job.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_iso.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_root_resize.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vms_with_varied_deploymentplanners.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_with_userdata.py Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_domain_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_domain_vpc_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py Intermittent failure detected: /marvin/tests/smoke/test_metrics_api.py Intermittent failure detected: /marvin/tests/smoke/test_multipleips_per_nic.py Intermittent failure detected: /marvin/tests/smoke/test_network_acl.py Intermittent failure detected: /marvin/tests/smoke/test_nic.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_portforwardingrules.py Intermittent failure detected: /marvin/tests/smoke/test_primary_storage.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_projects.py Intermittent failure detected: /marvin/tests/smoke/test_public_ip_range.py Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_routers.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_secondary_storage.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py Intermittent failure detected: /marvin/tests/smoke/test_storage_policy.py Intermittent failure detected: /marvin/tests/smoke/test_templates.py Intermittent failure detected: /marvin/tests/smoke/test_usage.py Intermittent failure detected: /marvin/tests/smoke/test_vm_deployment_planner.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 38 look OK, 48 have error(s) Only failed tests results shown below: Test Result Time (s) Test File @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S] Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 104 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S] Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_multiplication_x: debian. SL-JID 106 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S] @blueorangutan package @davidjumani a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S] @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S] Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 194 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests [S] [S] Trillian test result (tid-190) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 90748 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t190-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_deploy_vgpu_enabled_vm.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py Intermittent failure detected: /marvin/tests/smoke/test_human_readable_logs.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py Intermittent failure detected: /marvin/tests/smoke/test_network.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_routers.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Smoke tests completed. 60 look OK, 26 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_create_lb_rule_src_nat Failure 309.69 test_loadbalance.py test_02_create_lb_rule_non_nat Failure 36.55 test_loadbalance.py test_assign_and_removal_lb Failure 36.54 test_loadbalance.py ContextSuite context=TestDeployvGPUenabledVM>:setup Error 0.00 test_deploy_vgpu_enabled_vm.py test_07_deploy_vm_with_extraconfig_xenserver Error 659.01 test_deploy_vm_extra_config_data.py test_01_port_fwd_on_src_nat Failure 607.31 test_network.py test_02_port_fwd_on_non_src_nat Failure 609.07 test_network.py test_reboot_router Failure 386.76 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Error 609.38 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Error 610.16 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Error 613.22 test_network.py test_isolate_network_password_server Failure 161.16 test_password_server.py test_02_vpc_privategw_static_routes Failure 817.02 test_privategw_acl.py test_03_vpc_privategw_restart_vpc_cleanup Failure 815.13 test_privategw_acl.py test_04_rvpc_privategw_static_routes Failure 919.01 test_privategw_acl.py test_01_disableHumanReadableLogs Error 602.72 test_human_readable_logs.py test_02_enableHumanReadableLogs Error 602.73 test_human_readable_logs.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Failure 307.91 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Failure 386.90 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Error 168.94 test_internal_lb.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Error 220.07 test_internal_lb.py test_01_deploy_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 0.01 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_06_deploy_invalid_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_01_add_delete_kubernetes_supported_version Error 0.01 test_kubernetes_supported_versions.py test_router_dhcphosts Failure 168.68 test_router_dhcphosts.py ContextSuite context=TestRouterDHCPHosts>:teardown Error 182.03 test_router_dhcphosts.py test_router_dhcp_opts Error 613.47 test_router_dhcphosts.py test_router_dns_guestipquery Failure 456.87 test_router_dns.py test_router_dns_guestipquery Failure 456.63 test_router_dnsservice.py test_02_routervm_iptables_policies Error 689.05 test_routers_iptables_default_policy.py test_01_single_VPC_iptables_policies Error 751.75 test_routers_iptables_default_policy.py test_01_isolate_network_FW_PF_default_routes_egress_true Failure 238.18 test_routers_network_ops.py test_02_isolate_network_FW_PF_default_routes_egress_false Failure 248.52 test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failure 277.54 test_routers_network_ops.py test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false Failure 276.85 test_routers_network_ops.py test_03_RVR_Network_check_router_state Error 720.81 test_routers_network_ops.py test_01_router_internal_basic Error 602.86 test_routers.py test_02_router_internal_adv Error 602.78 test_routers.py test_04_restart_network_wo_cleanup Error 604.98 test_routers.py test_01_scale_vm Failure 10.24 test_scale_vm.py test_04_change_offering_small Failure 711.70 test_service_offerings.py test_01_snapshot_root_disk Error 611.99 test_snapshots.py test_03_ssvm_internals Error 602.84 test_ssvm.py test_04_cpvm_internals Error 602.79 test_ssvm.py test_05_stop_ssvm Error 676.77 test_ssvm.py test_06_stop_cpvm Error 657.54 test_ssvm.py test_07_reboot_ssvm Error 727.29 test_ssvm.py test_08_reboot_cpvm Error 634.19 test_ssvm.py test_09_destroy_ssvm Error 683.58 test_ssvm.py test_10_destroy_cpvm Error 657.51 test_ssvm.py test_10_attachAndDetach_iso Failure 609.62 test_vm_life_cycle.py test_change_service_offering_for_vm_with_snapshots Failure 695.43 test_vm_snapshots.py test_01_create_vm_snapshots Failure 602.39 test_vm_snapshots.py test_02_revert_vm_snapshots Failure 601.87 test_vm_snapshots.py test_03_delete_vm_snapshots Failure 0.02 test_vm_snapshots.py test_01_create_volume Failure 608.39 test_volumes.py test_02_attach_volume Failure 618.85 test_volumes.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Error 942.30 test_vpc_redundant.py test_02_redundant_VPC_default_routes Error 937.70 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Error 793.40 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Error 739.49 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 847.81 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 885.91 test_vpc_redundant.py test_01_VPC_nics_after_destroy Failure 783.44 test_vpc_router_nics.py test_02_VPC_default_routes Failure 770.89 test_vpc_router_nics.py test_01_redundant_vpc_site2site_vpn Failure 431.44 test_vpc_vpn.py test_01_vpc_site2site_vpn_multiple_options Error 282.70 test_vpc_vpn.py test_01_vpc_site2site_vpn Error 305.28 test_vpc_vpn.py [S] Trillian test result (tid-192) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 96345 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t192-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_human_readable_logs.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py Intermittent failure detected: /marvin/tests/smoke/test_nested_virtualization.py Intermittent failure detected: /marvin/tests/smoke/test_network.py Intermittent failure detected: /marvin/tests/smoke/test_nic.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_routers.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 59 look OK, 27 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_create_lb_rule_src_nat Failure 308.54 test_loadbalance.py test_02_create_lb_rule_non_nat Failure 34.65 test_loadbalance.py test_assign_and_removal_lb Failure 34.66 test_loadbalance.py test_01_port_fwd_on_src_nat Failure 605.88 test_network.py test_02_port_fwd_on_non_src_nat Failure 607.81 test_network.py test_reboot_router Failure 400.82 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Error 607.58 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Error 610.18 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Error 612.17 test_network.py test_05_deploy_vm_with_extraconfig_vmware Error 687.45 test_deploy_vm_extra_config_data.py test_01_nic Error 744.28 test_nic.py test_isolate_network_password_server Failure 157.36 test_password_server.py test_02_vpc_privategw_static_routes Failure 818.75 test_privategw_acl.py test_03_vpc_privategw_restart_vpc_cleanup Failure 816.90 test_privategw_acl.py test_04_rvpc_privategw_static_routes Failure 941.68 test_privategw_acl.py test_01_disableHumanReadableLogs Error 602.73 test_human_readable_logs.py test_02_enableHumanReadableLogs Error 602.72 test_human_readable_logs.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Failure 325.26 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Failure 385.99 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Error 173.62 test_internal_lb.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Error 229.63 test_internal_lb.py test_01_deploy_kubernetes_cluster Failure 0.01 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 0.01 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 0.01 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_06_deploy_invalid_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_01_add_delete_kubernetes_supported_version Error 0.01 test_kubernetes_supported_versions.py test_nested_virtualization_vmware Error 756.30 test_nested_virtualization.py test_router_dhcphosts Failure 157.19 test_router_dhcphosts.py ContextSuite context=TestRouterDHCPHosts>:teardown Error 169.51 test_router_dhcphosts.py test_router_dhcp_opts Error 609.25 test_router_dhcphosts.py test_router_dns_guestipquery Failure 456.44 test_router_dns.py test_router_dns_guestipquery Failure 455.30 test_router_dnsservice.py test_02_routervm_iptables_policies Error 685.08 test_routers_iptables_default_policy.py test_01_single_VPC_iptables_policies Error 761.96 test_routers_iptables_default_policy.py test_01_isolate_network_FW_PF_default_routes_egress_true Failure 234.08 test_routers_network_ops.py test_02_isolate_network_FW_PF_default_routes_egress_false Failure 229.31 test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failure 285.89 test_routers_network_ops.py test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false Failure 278.31 test_routers_network_ops.py test_03_RVR_Network_check_router_state Error 734.63 test_routers_network_ops.py test_01_router_internal_basic Error 602.85 test_routers.py test_02_router_internal_adv Error 602.86 test_routers.py test_04_restart_network_wo_cleanup Error 605.01 test_routers.py test_01_scale_vm Error 604.53 test_scale_vm.py test_04_change_offering_small Failure 719.49 test_service_offerings.py test_01_snapshot_root_disk Error 650.81 test_snapshots.py test_03_ssvm_internals Error 602.85 test_ssvm.py test_04_cpvm_internals Error 602.91 test_ssvm.py test_05_stop_ssvm Error 675.52 test_ssvm.py test_06_stop_cpvm Error 673.38 test_ssvm.py test_07_reboot_ssvm Error 703.26 test_ssvm.py test_08_reboot_cpvm Error 611.04 test_ssvm.py test_09_destroy_ssvm Error 705.68 test_ssvm.py test_10_destroy_cpvm Error 666.71 test_ssvm.py ContextSuite context=TestVAppsVM>:setup Error 41.18 test_vm_life_cycle.py test_10_attachAndDetach_iso Failure 609.21 test_vm_life_cycle.py test_change_service_offering_for_vm_with_snapshots Failure 698.23 test_vm_snapshots.py test_01_create_vm_snapshots Failure 604.54 test_vm_snapshots.py test_02_revert_vm_snapshots Failure 601.39 test_vm_snapshots.py test_03_delete_vm_snapshots Failure 0.01 test_vm_snapshots.py test_01_create_volume Failure 609.12 test_volumes.py test_02_attach_volume Failure 606.96 test_volumes.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Error 905.83 test_vpc_redundant.py test_02_redundant_VPC_default_routes Error 905.08 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Error 791.44 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Error 751.93 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 869.92 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 905.03 test_vpc_redundant.py test_01_VPC_nics_after_destroy Failure 769.04 test_vpc_router_nics.py test_02_VPC_default_routes Failure 766.56 test_vpc_router_nics.py test_01_redundant_vpc_site2site_vpn Failure 437.55 test_vpc_vpn.py test_01_vpc_site2site_vpn_multiple_options Error 298.69 test_vpc_vpn.py test_01_vpc_site2site_vpn Error 311.23 test_vpc_vpn.py [S] Trillian test result (tid-191) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 101140 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t191-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_deploy_virtio_scsi_vm.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py Intermittent failure detected: /marvin/tests/smoke/test_human_readable_logs.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py Intermittent failure detected: /marvin/tests/smoke/test_network.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_routers.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py Smoke tests completed. 60 look OK, 26 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_create_lb_rule_src_nat Failure 307.78 test_loadbalance.py test_02_create_lb_rule_non_nat Failure 35.59 test_loadbalance.py test_assign_and_removal_lb Failure 34.60 test_loadbalance.py test_01_verify_libvirt Error 602.64 test_deploy_virtio_scsi_vm.py test_02_verify_libvirt_after_restart Error 609.98 test_deploy_virtio_scsi_vm.py test_03_verify_libvirt_attach_disk Error 605.85 test_deploy_virtio_scsi_vm.py test_04_verify_guest_lspci Error 602.08 test_deploy_virtio_scsi_vm.py test_05_change_vm_ostype_restart Error 610.07 test_deploy_virtio_scsi_vm.py test_06_verify_guest_lspci_again Error 602.06 test_deploy_virtio_scsi_vm.py ContextSuite context=TestAddConfigtoDeployVM>:setup Error 0.00 test_deploy_vm_extra_config_data.py test_01_port_fwd_on_src_nat Failure 605.08 test_network.py test_02_port_fwd_on_non_src_nat Failure 607.91 test_network.py ContextSuite context=TestPrivateVlansL2Networks>:setup Error 1230.16 test_network.py test_reboot_router Failure 358.50 test_network.py test_network_rules_acquired_public_ip_1_static_nat_rule Error 607.44 test_network.py test_network_rules_acquired_public_ip_2_nat_rule Error 609.15 test_network.py test_network_rules_acquired_public_ip_3_Load_Balancer_Rule Error 613.23 test_network.py test_isolate_network_password_server Failure 157.18 test_password_server.py test_02_vpc_privategw_static_routes Failure 828.76 test_privategw_acl.py test_03_vpc_privategw_restart_vpc_cleanup Failure 754.69 test_privategw_acl.py test_04_rvpc_privategw_static_routes Failure 851.45 test_privategw_acl.py test_01_disableHumanReadableLogs Error 602.87 test_human_readable_logs.py test_02_enableHumanReadableLogs Error 602.61 test_human_readable_logs.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Failure 268.37 test_internal_lb.py test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80 Failure 317.03 test_internal_lb.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Error 143.52 test_internal_lb.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Error 172.93 test_internal_lb.py test_01_deploy_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_06_deploy_invalid_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_01_add_delete_kubernetes_supported_version Error 0.01 test_kubernetes_supported_versions.py test_router_dhcphosts Failure 157.09 test_router_dhcphosts.py ContextSuite context=TestRouterDHCPHosts>:teardown Error 168.40 test_router_dhcphosts.py test_router_dhcp_opts Error 609.10 test_router_dhcphosts.py test_router_dns_guestipquery Failure 454.70 test_router_dns.py test_router_dns_guestipquery Failure 454.62 test_router_dnsservice.py test_02_routervm_iptables_policies Error 660.60 test_routers_iptables_default_policy.py test_01_single_VPC_iptables_policies Error 720.34 test_routers_iptables_default_policy.py test_01_isolate_network_FW_PF_default_routes_egress_true Failure 209.90 test_routers_network_ops.py test_02_isolate_network_FW_PF_default_routes_egress_false Failure 210.91 test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failure 243.33 test_routers_network_ops.py test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false Failure 243.11 test_routers_network_ops.py test_03_RVR_Network_check_router_state Error 689.27 test_routers_network_ops.py test_01_router_internal_basic Error 602.80 test_routers.py test_02_router_internal_adv Error 602.75 test_routers.py test_04_restart_network_wo_cleanup Error 604.99 test_routers.py test_01_service_offering_cpu_limit_use Error 100.50 test_service_offerings.py test_04_change_offering_small Failure 715.21 test_service_offerings.py test_01_snapshot_root_disk Error 606.04 test_snapshots.py test_03_ssvm_internals Error 602.83 test_ssvm.py test_04_cpvm_internals Error 602.81 test_ssvm.py test_05_stop_ssvm Error 676.59 test_ssvm.py test_06_stop_cpvm Error 652.34 test_ssvm.py test_07_reboot_ssvm Error 720.59 test_ssvm.py test_08_reboot_cpvm Error 630.16 test_ssvm.py test_09_destroy_ssvm Error 672.43 test_ssvm.py test_10_destroy_cpvm Error 653.40 test_ssvm.py test_01_migrate_VM_and_root_volume Error 72.69 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 54.24 test_vm_life_cycle.py test_01_secure_vm_migration Error 703.25 test_vm_life_cycle.py test_02_unsecure_vm_migration Error 637.16 test_vm_life_cycle.py test_03_secured_to_nonsecured_vm_migration Error 637.97 test_vm_life_cycle.py test_04_nonsecured_to_secured_vm_migration Error 636.14 test_vm_life_cycle.py test_10_attachAndDetach_iso Failure 608.16 test_vm_life_cycle.py test_01_create_vm_snapshots Failure 604.51 test_vm_snapshots.py test_02_revert_vm_snapshots Failure 601.21 test_vm_snapshots.py test_03_delete_vm_snapshots Failure 0.01 test_vm_snapshots.py test_01_create_volume Failure 606.75 test_volumes.py test_02_attach_volume Failure 605.85 test_volumes.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Error 825.80 test_vpc_redundant.py test_02_redundant_VPC_default_routes Error 824.98 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Error 743.03 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Error 709.98 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 801.08 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 827.08 test_vpc_redundant.py test_01_VPC_nics_after_destroy Failure 752.13 test_vpc_router_nics.py test_02_VPC_default_routes Failure 724.23 test_vpc_router_nics.py test_01_redundant_vpc_site2site_vpn Failure 345.91 test_vpc_vpn.py test_01_vpc_site2site_vpn_multiple_options Error 249.38 test_vpc_vpn.py test_01_vpc_site2site_vpn Error 286.15 test_vpc_vpn.py test_hostha_enable_ha_when_host_disabled Error 0.35 test_hostha_kvm.py test_hostha_enable_ha_when_host_disconected Error 605.03 test_hostha_kvm.py test_hostha_enable_ha_when_host_in_maintenance Error 303.11 test_hostha_kvm.py test_hostha_kvm_host_fencing Error 7.53 test_hostha_kvm.py Trillian test result (tid-202) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 66060 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t202-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vms_with_varied_deploymentplanners.py Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_with_userdata.py Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_iso.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py Intermittent failure detected: /marvin/tests/smoke/test_metrics_api.py Intermittent failure detected: /marvin/tests/smoke/test_multipleips_per_nic.py Intermittent failure detected: /marvin/tests/smoke/test_nic_adapter_type.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_portforwardingrules.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_projects.py Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py Smoke tests completed. 71 look OK, 15 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_DeployVmAntiAffinityGroup_in_project Error 58.76 test_affinity_groups_projects.py test_DeployVmAntiAffinityGroup Error 34.34 test_affinity_groups.py test_list_clusters_metrics Error 1527.61 test_metrics_api.py test_list_vms_metrics Error 0.26 test_metrics_api.py ContextSuite context=TestDeployVmWithVariedPlanners>:setup Error 0.00 test_deploy_vms_with_varied_deploymentplanners.py ContextSuite context=TestDeployVmWithUserData>:setup Error 0.00 test_deploy_vm_with_userdata.py ContextSuite context=TestRemoteDiagnostics>:setup Error 0.00 test_diagnostics.py ContextSuite context=TestAdapterTypeForNic>:setup Error 0.00 test_nic_adapter_type.py test_03_vpc_internallb_haproxy_stats_on_all_interfaces Error 197.81 test_internal_lb.py test_04_rvpc_internallb_haproxy_stats_on_all_interfaces Error 240.88 test_internal_lb.py test_01_create_iso_with_checksum_sha1 Error 65.33 test_iso.py test_02_create_iso_with_checksum_sha256 Error 65.37 test_iso.py test_03_create_iso_with_checksum_md5 Error 65.30 test_iso.py test_04_create_iso_with_no_checksum Error 65.37 test_iso.py test_01_create_iso Failure 1512.00 test_iso.py ContextSuite context=TestISO>:setup Error 3025.08 test_iso.py test_01_deploy_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.00 test_kubernetes_clusters.py test_06_deploy_invalid_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.00 test_kubernetes_clusters.py test_01_add_delete_kubernetes_supported_version Error 60.58 test_kubernetes_supported_versions.py ContextSuite context=TestListIdsParams>:setup Error 0.00 test_list_ids_parameter.py ContextSuite context=TestPortForwardingRules>:setup Error 0.00 test_portforwardingrules.py ContextSuite context=TestPrivateGwACL>:setup Error 0.00 test_privategw_acl.py test_hostha_kvm_host_degraded Failure 690.01 test_hostha_kvm.py test_hostha_kvm_host_fencing Failure 644.69 test_hostha_kvm.py test_hostha_kvm_host_recovering Failure 648.20 test_hostha_kvm.py @blueorangutan test matrix @davidjumani a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian Build Failed (tid-213) Trillian Build Failed (tid-215) Trillian Build Failed (tid-214) @blueorangutan test matrix @davidjumani a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian Build Failed (tid-216) Trillian Build Failed (tid-219) Trillian test result (tid-217) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 23343 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t217-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_secondary_storage.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py Smoke tests completed. 76 look OK, 10 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_isolate_network_FW_PF_default_routes_egress_true Failure 131.08 test_routers_network_ops.py test_02_isolate_network_FW_PF_default_routes_egress_false Failure 109.27 test_routers_network_ops.py test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true Failure 204.51 test_routers_network_ops.py test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false Failure 211.38 test_routers_network_ops.py test_01_sys_vm_start Failure 0.04 test_secondary_storage.py ContextSuite context=TestDeployVM>:setup Error 0.00 test_vm_life_cycle.py test_01_migrate_VM_and_root_volume Error 1.19 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 1.13 test_vm_life_cycle.py test_01_secure_vm_migration Error 0.06 test_vm_life_cycle.py test_02_unsecure_vm_migration Error 1.07 test_vm_life_cycle.py test_03_secured_to_nonsecured_vm_migration Error 0.06 test_vm_life_cycle.py test_04_nonsecured_to_secured_vm_migration Error 1.08 test_vm_life_cycle.py ContextSuite context=TestVMLifeCycle>:setup Error 1.63 test_vm_life_cycle.py ContextSuite context=TestVmSnapshot>:setup Error 1.60 test_vm_snapshots.py ContextSuite context=TestCreateVolume>:setup Error 0.00 test_volumes.py ContextSuite context=TestVolumes>:setup Error 0.00 test_volumes.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Error 6.65 test_vpc_redundant.py test_02_redundant_VPC_default_routes Error 8.69 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Error 7.66 test_vpc_redundant.py test_04_rvpc_network_garbage_collector_nics Error 7.78 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 7.73 test_vpc_redundant.py test_01_VPC_nics_after_destroy Error 4.65 test_vpc_router_nics.py test_02_VPC_default_routes Error 5.66 test_vpc_router_nics.py test_01_redundant_vpc_site2site_vpn Failure 7.31 test_vpc_vpn.py test_01_vpc_site2site_vpn_multiple_options Failure 5.26 test_vpc_vpn.py test_01_vpc_remote_access_vpn Failure 3.15 test_vpc_vpn.py test_01_vpc_site2site_vpn Failure 5.27 test_vpc_vpn.py test_01_cancel_host_maintenace_with_no_migration_jobs Error 0.04 test_host_maintenance.py test_02_cancel_host_maintenace_with_migration_jobs Error 0.04 test_host_maintenance.py test_03_cancel_host_maintenace_with_migration_jobs_failure Error 0.04 test_host_maintenance.py test_01_cancel_host_maintenance_ssh_enabled_agent_connected Error 0.01 test_host_maintenance.py test_03_cancel_host_maintenance_ssh_disabled_agent_connected Error 0.01 test_host_maintenance.py test_04_cancel_host_maintenance_ssh_disabled_agent_disconnected Error 0.01 test_host_maintenance.py test_hostha_enable_ha_when_host_in_maintenance Error 303.52 test_hostha_kvm.py Trillian test result (tid-220) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36258 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t220-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 85 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_scale_vm Failure 11.27 test_scale_vm.py Trillian test result (tid-218) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 51077 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t218-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 82 look OK, 4 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_10_traceroute_in_vr Failure 61.13 test_diagnostics.py test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80 Failure 406.03 test_internal_lb.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 800.50 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 3615.34 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 87.96 test_kubernetes_clusters.py ContextSuite context=TestVAppsVM>:setup Error 46.39 test_vm_life_cycle.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_multiplication_x: debian. SL-JID 223 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_multiplication_x: debian. SL-JID 229 @blueorangutan package @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 288 Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 284 @blueorangutan test matrix cc @davidjumani @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-310) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 37814 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t310-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 80.36 test_kubernetes_clusters.py test_01_migrate_VM_and_root_volume Error 75.35 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 50.13 test_vm_life_cycle.py Trillian test result (tid-311) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 38159 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t311-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 83 look OK, 3 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_10_traceroute_in_vr Failure 61.19 test_diagnostics.py test_01_deploy_kubernetes_cluster Failure 118.00 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 108.44 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 65.53 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 77.75 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 84.81 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 66.41 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 82.90 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 71.83 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 112.23 test_kubernetes_clusters.py ContextSuite context=TestVAppsVM>:setup Error 45.53 test_vm_life_cycle.py Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 310 Trillian test result (tid-332) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 35328 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t332-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 70.07 test_kubernetes_clusters.py test_01_migrate_VM_and_root_volume Error 68.19 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 51.08 test_vm_life_cycle.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 316 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-339) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36700 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t339-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Smoke tests completed. 83 look OK, 3 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 89.11 test_kubernetes_clusters.py test_01_scale_vm Failure 10.29 test_scale_vm.py test_01_cancel_host_maintenace_with_no_migration_jobs Failure 306.06 test_host_maintenance.py test_02_cancel_host_maintenace_with_migration_jobs Failure 306.08 test_host_maintenance.py test_03_cancel_host_maintenace_with_migration_jobs_failure Failure 0.19 test_host_maintenance.py Trillian test result (tid-341) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 37255 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t341-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_deploy_kubernetes_cluster Failure 89.25 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 95.24 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 58.45 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 68.63 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 62.51 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 72.65 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 95.15 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 56.43 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 99.09 test_kubernetes_clusters.py ContextSuite context=TestVAppsVM>:setup Error 58.95 test_vm_life_cycle.py Trillian test result (tid-340) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 38569 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t340-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 78.23 test_kubernetes_clusters.py test_01_migrate_VM_and_root_volume Error 67.26 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 49.09 test_vm_life_cycle.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 327 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-355) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 35150 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t355-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_deploy_kubernetes_cluster Failure 106.86 test_kubernetes_clusters.py test_02_invalid_upgrade_kubernetes_cluster Failure 101.29 test_kubernetes_clusters.py test_03_deploy_and_upgrade_kubernetes_cluster Failure 59.63 test_kubernetes_clusters.py test_04_deploy_and_scale_kubernetes_cluster Failure 54.75 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 52.15 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 59.38 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 55.18 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 51.22 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 90.47 test_kubernetes_clusters.py ContextSuite context=TestVAppsVM>:setup Error 44.56 test_vm_life_cycle.py Trillian test result (tid-354) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36025 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t354-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_password_server.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 69.78 test_kubernetes_clusters.py test_01_migrate_VM_and_root_volume Error 69.29 test_vm_life_cycle.py test_02_migrate_VM_with_two_data_disks Error 51.05 test_vm_life_cycle.py test_10_attachAndDetach_iso Error 12.42 test_vm_life_cycle.py Trillian test result (tid-353) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 42021 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t353-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 84 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestKubernetesCluster>:teardown Error 86.04 test_kubernetes_clusters.py test_01_scale_vm Failure 12.30 test_scale_vm.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 380 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-402) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32668 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t402-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_scale_vm Failure 10.27 test_scale_vm.py Trillian test result (tid-403) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 34375 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t403-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-404) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36173 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t404-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Smoke tests completed. 85 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_10_traceroute_in_vr Failure 61.13 test_diagnostics.py ContextSuite context=TestVAppsVM>:setup Error 79.67 test_vm_life_cycle.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 388 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 390 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 401 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-422) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 36092 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t422-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_scale_vm Failure 9.24 test_scale_vm.py Trillian test result (tid-423) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 35809 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t423-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-424) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 35981 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t424-vmware-65u2.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 420 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-459) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32999 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t459-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_scale_vm Failure 11.26 test_scale_vm.py Trillian test result (tid-461) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 35111 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t461-vmware-65u2.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-460) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 48079 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t460-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_iso.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_03_deploy_and_scale_kubernetes_cluster Failure 3641.88 test_kubernetes_clusters.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 442 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-496) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 33717 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t496-kvm-centos7.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-495) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 42563 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t495-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_multipleips_per_nic.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Smoke tests completed. 85 look OK, 2 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_invalid_upgrade_kubernetes_cluster Failure 3603.00 test_kubernetes_clusters.py test_02_deploy_and_upgrade_kubernetes_cluster Failure 3616.86 test_kubernetes_clusters.py test_03_deploy_and_scale_kubernetes_cluster Failure 0.07 test_kubernetes_clusters.py test_04_basic_lifecycle_kubernetes_cluster Failure 0.04 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.04 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 112.63 test_kubernetes_clusters.py test_01_scale_vm Failure 9.23 test_scale_vm.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 457 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-522) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 41013 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t522-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_scale_vm Failure 10.25 test_scale_vm.py Trillian test result (tid-524) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 41907 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t524-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_07_deploy_kubernetes_ha_cluster Failure 3613.91 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.06 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 51.22 test_kubernetes_clusters.py Trillian test result (tid-523) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 44200 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t523-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_07_deploy_kubernetes_ha_cluster Failure 3612.37 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 80.94 test_kubernetes_clusters.py @blueorangutan ui @davidjumani a Jenkins job has been kicked to build UI QA env. I'll keep you posted as I make progress. UI build: :heavy_check_mark: Live QA URL: http://qa.cloudstack.cloud:8080/client/pr/4710 (SL-JID-109) Trillian test result (tid-582) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32934 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t582-xenserver-71.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-583) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 32915 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t583-kvm-centos7.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 516 @blueorangutan package @shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 518 @blueorangutan test matrix @rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests Trillian test result (tid-605) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 33567 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t605-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-607) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 35381 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t607-vmware-65u2.zip Smoke tests completed. 87 look OK, 0 have error(s) Only failed tests results shown below: Test Result Time (s) Test File Trillian test result (tid-606) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 54867 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t606-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_storage_policy.py Intermittent failure detected: /marvin/tests/smoke/test_templates.py Intermittent failure detected: /marvin/tests/smoke/test_usage.py Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py Intermittent failure detected: /marvin/tests/smoke/test_volumes.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py Smoke tests completed. 78 look OK, 9 have error(s) Only failed tests results shown below: Test Result Time (s) Test File ContextSuite context=TestVMWareStoragePolicies>:setup Error 0.00 test_storage_policy.py test_02_create_template_with_checksum_sha1 Error 65.37 test_templates.py test_03_create_template_with_checksum_sha256 Error 65.36 test_templates.py test_04_create_template_with_checksum_md5 Error 65.39 test_templates.py test_05_create_template_with_no_checksum Error 65.37 test_templates.py test_04_extract_template Failure 128.33 test_templates.py ContextSuite context=TestISOUsage>:setup Error 0.00 test_usage.py test_01_volume_usage Failure 787.32 test_usage.py test_10_attachAndDetach_iso Failure 1511.07 test_vm_life_cycle.py test_06_download_detached_volume Failure 312.60 test_volumes.py ContextSuite context=TestVPCRedundancy>:setup Error 0.00 test_vpc_redundant.py ContextSuite context=TestVPCNics>:setup Error 0.00 test_vpc_router_nics.py ContextSuite context=TestRVPCSite2SiteVpn>:setup Error 0.00 test_vpc_vpn.py ContextSuite context=TestVPCSite2SiteVPNMultipleOptions>:setup Error 0.00 test_vpc_vpn.py ContextSuite context=TestVpcRemoteAccessVpn>:setup Error 0.00 test_vpc_vpn.py ContextSuite context=TestVpcSite2SiteVpn>:setup Error 0.00 test_vpc_vpn.py test_disable_oobm_ha_state_ineligible Error 1511.93 test_hostha_kvm.py @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 524 @blueorangutan package @rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 531 Trillian test result (tid-666) Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7 Total time taken: 51950 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t666-xenserver-71.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_usage.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Smoke tests completed. 86 look OK, 1 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_01_invalid_upgrade_kubernetes_cluster Failure 3526.47 test_kubernetes_clusters.py test_02_deploy_and_upgrade_kubernetes_cluster Failure 3612.57 test_kubernetes_clusters.py test_03_deploy_and_scale_kubernetes_cluster Failure 0.05 test_kubernetes_clusters.py test_04_basic_lifecycle_kubernetes_cluster Failure 0.05 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.04 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 38.06 test_kubernetes_clusters.py Trillian test result (tid-667) Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7 Total time taken: 83767 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t667-kvm-centos7.zip Intermittent failure detected: /marvin/tests/smoke/test_deploy_virtio_scsi_vm.py Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_network.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_routers.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py Smoke tests completed. 75 look OK, 12 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_04_verify_guest_lspci Error 810.60 test_deploy_virtio_scsi_vm.py test_06_verify_guest_lspci_again Error 810.57 test_deploy_virtio_scsi_vm.py test_01_invalid_upgrade_kubernetes_cluster Failure 3610.37 test_kubernetes_clusters.py test_03_deploy_and_scale_kubernetes_cluster Failure 36.75 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 43.86 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.04 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 84.64 test_kubernetes_clusters.py ContextSuite context=TestResetVmOnReboot>:setup Error 0.00 test_reset_vm_on_reboot.py ContextSuite context=TestRAMCPUResourceAccounting>:setup Error 0.00 test_resource_accounting.py ContextSuite context=TestRouterDHCPHosts>:setup Error 0.00 test_router_dhcphosts.py ContextSuite context=TestRouterDHCPOpts>:setup Error 0.00 test_router_dhcphosts.py ContextSuite context=TestRouterDns>:setup Error 0.00 test_router_dns.py ContextSuite context=TestRouterDnsService>:setup Error 0.00 test_router_dnsservice.py ContextSuite context=TestRouterIpTablesPolicies>:setup Error 0.00 test_routers_iptables_default_policy.py ContextSuite context=TestVPCIpTablesPolicies>:setup Error 0.00 test_routers_iptables_default_policy.py ContextSuite context=TestIsolatedNetworks>:setup Error 0.00 test_routers_network_ops.py ContextSuite context=TestRedundantIsolateNetworks>:setup Error 0.00 test_routers_network_ops.py ContextSuite context=TestRouterServices>:setup Error 0.00 test_routers.py ContextSuite context=TestCpuCapServiceOfferings>:setup Error 0.00 test_service_offerings.py ContextSuite context=TestServiceOfferings>:setup Error 0.16 test_service_offerings.py ContextSuite context=TestSnapshotRootDisk>:setup Error 0.00 test_snapshots.py Trillian test result (tid-668) Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7 Total time taken: 85457 seconds Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t668-vmware-65u2.zip Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py Intermittent failure detected: /marvin/tests/smoke/test_routers.py Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py Smoke tests completed. 74 look OK, 13 have error(s) Only failed tests results shown below: Test Result Time (s) Test File test_03_vpc_privategw_restart_vpc_cleanup Failure 1046.65 test_privategw_acl.py test_01_invalid_upgrade_kubernetes_cluster Failure 3608.66 test_kubernetes_clusters.py test_02_deploy_and_upgrade_kubernetes_cluster Failure 3609.05 test_kubernetes_clusters.py test_03_deploy_and_scale_kubernetes_cluster Failure 0.05 test_kubernetes_clusters.py test_04_basic_lifecycle_kubernetes_cluster Failure 0.05 test_kubernetes_clusters.py test_05_delete_kubernetes_cluster Failure 0.05 test_kubernetes_clusters.py test_07_deploy_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_08_deploy_and_upgrade_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py test_09_delete_kubernetes_ha_cluster Failure 0.05 test_kubernetes_clusters.py ContextSuite context=TestKubernetesCluster>:teardown Error 51.06 test_kubernetes_clusters.py ContextSuite context=TestResetVmOnReboot>:setup Error 0.00 test_reset_vm_on_reboot.py ContextSuite context=TestRAMCPUResourceAccounting>:setup Error 0.00 test_resource_accounting.py ContextSuite context=TestRouterDHCPHosts>:setup Error 0.00 test_router_dhcphosts.py ContextSuite context=TestRouterDHCPOpts>:setup Error 0.00 test_router_dhcphosts.py ContextSuite context=TestRouterDns>:setup Error 0.00 test_router_dns.py ContextSuite context=TestRouterDnsService>:setup Error 0.00 test_router_dnsservice.py ContextSuite context=TestRouterIpTablesPolicies>:setup Error 0.00 test_routers_iptables_default_policy.py ContextSuite context=TestVPCIpTablesPolicies>:setup Error 0.00 test_routers_iptables_default_policy.py ContextSuite context=TestIsolatedNetworks>:setup Error 0.00 test_routers_network_ops.py ContextSuite context=TestRedundantIsolateNetworks>:setup Error 0.00 test_routers_network_ops.py ContextSuite context=TestRouterServices>:setup Error 0.00 test_routers.py ContextSuite context=TestServiceOfferings>:setup Error 1511.59 test_service_offerings.py ContextSuite context=TestSnapshotRootDisk>:setup Error 0.00 test_snapshots.py test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL Failure 581.90 test_vpc_redundant.py test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers Failure 650.51 test_vpc_redundant.py test_05_rvpc_multi_tiers Failure 526.25 test_vpc_redundant.py test_05_rvpc_multi_tiers Error 562.43 test_vpc_redundant.py
gharchive/pull-request
2021-02-19T08:51:57
2025-04-01T04:33:28.045494
{ "authors": [ "DaanHoogland", "blueorangutan", "davidjumani", "rhtyd", "shwstppr" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/4710", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
107130098
Bug-ID: CLOUDSTACK-8870: Skip external device usage collection if no external devices exist external network device usage monitor thread that runs every 5mins by default (based on global config external.network.stats.interval) and runs coalesce query to acquire a lock. When there are no external devices exist, there is no need to run usage collection. Added test case to verify that usage collection task is not run when there are no External LB or External FW Ping @remibergsma @borisroman @kishankavala added a unit test to cover his change. I went through the code and haven't found any issue. The if block if(_hostDao.listByType(Host.Type.ExternalFirewall).isEmpty() && _hostDao.listByType(Host.Type.ExternalLoadBalancer).isEmpty()){ added to ExternalDeviceUsageManagerImpl make sense and are covered. Please give it a go with the integration tests we have. Cheers, Wilder I have gone through the code. patch LGTM. I need one more LGTM. I will add this to my CI queue. Thx... Although I had pointed a typo in a comment line (nothing serious), the code LGTM. Thank you @GabrielBrascher. @kishankavala would you mind fixing that typo? I will try to get this run though CI soon... @kishankavala please rebase against latest master and push -f, update on status of your PR I've left a NPE check, we should merge it once you fix that. Thanks. tag:easypr cc @swill @kishankavala please rebase this PR as we have merge conflicts. Thanks... @swill rebased with latest master @GabrielBrascher fixed the typo LGTM CI RESULTS Tests Run: 85 Skipped: 0 Failed: 2 Errors: 1 Duration: 10h 41m 55s Summary of the problem(s): ERROR: Test to verify access to loadbalancer haproxy admin stats page ---------------------------------------------------------------------- Traceback (most recent call last): File "/data/git/cs1/cloudstack/test/integration/smoke/test_internal_lb.py", line 854, in tearDown raise Exception("Cleanup failed with %s" % e) Exception: Cleanup failed with Job failed: {jobprocstatus : 0, created : u'2016-05-24T12:24:25+0200', jobresult : {errorcode : 530, errortext : u'Failed to delete network'}, cmd : u'org.apache.cloudstack.api.command.user.network.DeleteNetworkCmd', userid : u'a5538db6-2168-11e6-932f-5254001daa61', jobstatus : 2, jobid : u'8e7ebb93-fb36-4842-8e2d-5cefea2ff80a', jobresultcode : 530, jobresulttype : u'object', jobinstancetype : u'Network', accountid : u'a5537331-2168-11e6-932f-5254001daa61'} ---------------------------------------------------------------------- Additional details in: /tmp/MarvinLogs/test_network_F00C21/results.txt FAIL: test_02_vpc_privategw_static_routes (integration.smoke.test_privategw_acl.TestPrivateGwACL) ---------------------------------------------------------------------- Traceback (most recent call last): File "/data/git/cs1/cloudstack/test/integration/smoke/test_privategw_acl.py", line 262, in test_02_vpc_privategw_static_routes self.performVPCTests(vpc_off) File "/data/git/cs1/cloudstack/test/integration/smoke/test_privategw_acl.py", line 325, in performVPCTests privateGw_1 = self.createPvtGw(vpc_1, "10.0.3.100", "10.0.3.101", acl1.id, vlan_1) File "/data/git/cs1/cloudstack/test/integration/smoke/test_privategw_acl.py", line 595, in createPvtGw self.fail("Failed to create Private Gateway ==> %s" % e) AssertionError: Failed to create Private Gateway ==> Execute cmd: createprivategateway failed, due to: errorCode: 431, errorText:Network with vlan vlan://100 already exists in zone 1 ---------------------------------------------------------------------- Additional details in: /tmp/MarvinLogs/test_network_F00C21/results.txt FAIL: Test destroy(expunge) Virtual Machine ---------------------------------------------------------------------- Traceback (most recent call last): File "/data/git/cs1/cloudstack/test/integration/smoke/test_vm_life_cycle.py", line 646, in test_09_expunge_vm self.assertEqual(list_vm_response,None,"Check Expunged virtual machine is in listVirtualMachines response") AssertionError: Check Expunged virtual machine is in listVirtualMachines response ---------------------------------------------------------------------- Additional details in: /tmp/MarvinLogs/test_vpc_routers_J2J0PP/results.txt Associated Uploads /tmp/MarvinLogs/DeployDataCenter__May_24_2016_06_36_18_ATGDYK: dc_entries.obj failed_plus_exceptions.txt runinfo.txt /tmp/MarvinLogs/test_network_F00C21: failed_plus_exceptions.txt results.txt runinfo.txt /tmp/MarvinLogs/test_vpc_routers_J2J0PP: failed_plus_exceptions.txt results.txt runinfo.txt Uploads will be available until 2016-07-25 02:00:00 +0200 CEST Comment created by upr comment. CI RESULTS Tests Run: 82 Skipped: 0 Failed: 0 Errors: 3 Duration: 8h 33m 42s Summary of the problem(s): ERROR: test suite for <class 'integration.smoke.test_vpc_vpn.TestRVPCSite2SiteVpn'> ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 209, in run self.setUp() File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 292, in setUp self.setupContext(ancestor) File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 315, in setupContext try_run(context, names) File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/util.py", line 471, in try_run return func() File "/data/git/cs1/cloudstack/test/integration/smoke/test_vpc_vpn.py", line 835, in setUpClass cls.template.download(cls.apiclient) File "/usr/lib/python2.7/site-packages/marvin/lib/base.py", line 1350, in download elif 'Downloaded' in template.status: TypeError: argument of type 'NoneType' is not iterable ---------------------------------------------------------------------- Additional details in: /tmp/MarvinLogs/test_network_VKMDK0/results.txt ERROR: test suite for <class 'integration.smoke.test_vpc_vpn.TestVpcRemoteAccessVpn'> ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 209, in run self.setUp() File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 292, in setUp self.setupContext(ancestor) File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 315, in setupContext try_run(context, names) File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/util.py", line 471, in try_run return func() File "/data/git/cs1/cloudstack/test/integration/smoke/test_vpc_vpn.py", line 293, in setUpClass cls.template.download(cls.apiclient) File "/usr/lib/python2.7/site-packages/marvin/lib/base.py", line 1350, in download elif 'Downloaded' in template.status: TypeError: argument of type 'NoneType' is not iterable ---------------------------------------------------------------------- Additional details in: /tmp/MarvinLogs/test_network_VKMDK0/results.txt ERROR: test suite for <class 'integration.smoke.test_vpc_vpn.TestVpcSite2SiteVpn'> ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 209, in run self.setUp() File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 292, in setUp self.setupContext(ancestor) File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 315, in setupContext try_run(context, names) File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/util.py", line 471, in try_run return func() File "/data/git/cs1/cloudstack/test/integration/smoke/test_vpc_vpn.py", line 472, in setUpClass cls.template.download(cls.apiclient) File "/usr/lib/python2.7/site-packages/marvin/lib/base.py", line 1350, in download elif 'Downloaded' in template.status: TypeError: argument of type 'NoneType' is not iterable ---------------------------------------------------------------------- Additional details in: /tmp/MarvinLogs/test_network_VKMDK0/results.txt Associated Uploads /tmp/MarvinLogs/DeployDataCenter__May_25_2016_19_10_57_8DI3FD: dc_entries.obj failed_plus_exceptions.txt runinfo.txt /tmp/MarvinLogs/test_network_VKMDK0: failed_plus_exceptions.txt results.txt runinfo.txt /tmp/MarvinLogs/test_vpc_routers_PXO5U1: failed_plus_exceptions.txt results.txt runinfo.txt Uploads will be available until 2016-07-26 02:00:00 +0200 CEST Comment created by upr comment. This is a cleaner run because this issue is a known issue. I think this one is ready now... code/patch LGTM.
gharchive/pull-request
2015-09-18T04:14:17
2025-04-01T04:33:28.069207
{ "authors": [ "GabrielBrascher", "kishankavala", "rajesh-battala", "rhtyd", "swill", "wilderrodrigues" ], "repo": "apache/cloudstack", "url": "https://github.com/apache/cloudstack/pull/846", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
416567771
TEXT-152 - new tests and fix for infinite loop Hi, I have added a test that show shows the infinite loop, and the code that fixes it. In order to make the test pass, and not only make the infinite loop stop happening, more changes were needed because the code for TEXT-111 did not fix the replacement of 0 length matches when we have multiple matches. Coverage decreased (-0.02%) to 97.885% when pulling ad07efa10b4091ef3b9d1a9ee9ec0190726e8c55 on CAPS50:TEXT-152 into d880f686338f63b8e0860ad37da190fda792eecf on apache:master. @CAPS50, May you please update changes.xml? Thank you, Gary @garydgregory I think the contributor goes under due-to. But sometimes new contributors add it as dev. And that way, it's tricky to know which value to put in dev. So it might be easier for a committer to add it when merging? At least that was my workflow so far Thank you. @kinow shall I change the changes.xml ? and if so should be the value of dev and due-to? @kinow shall I change the changes.xml ? and if so should be the value of dev and due-to? IMO, you can leave dev empty and put your name in due-to (or you GitHub ID, Jira ID, whatever you prefer). thanks, I did that. Thanks @CAPS50 ! Merging now
gharchive/pull-request
2019-03-03T21:33:49
2025-04-01T04:33:28.075322
{ "authors": [ "CAPS50", "coveralls", "garydgregory", "kinow" ], "repo": "apache/commons-text", "url": "https://github.com/apache/commons-text/pull/107", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
543673347
Update IndexedDB platform support General This proposal supersedes and closes #1011 (PR #1011) /cc @jgaehring with thanks for the original proposal in #1011 Platforms affected Android, browser, Electron, iOS, "osx" (macOS) Motivation and Context From a second look at PR #1011 I discovered that the documentation of IndexedDB support is outdated for multiple platforms. I would like to get the documentation updated for all affected platforms at once. Description Update the description to make it clear that this functionality is supported by the underlying WebView on all platforms, with possible limitations added for the (web) browser platform The "4.4 and above" condition for Android now gone, since we have already dropped support for Android pre-4.4 "Not supported on iOS" item now completely and unconditionally gone from disadvantages section due to this comment: https://github.com/apache/cordova-docs/pull/1011#issuecomment-567250310 Testing Visual inspection Checklist [ ] [TODO] I've run the tests to see all new and existing tests pass I added automated test coverage as appropriate for this change Commit is prefixed with (platform) if this change only applies to one platform (e.g. (android)) [x] If this Pull Request resolves an issue another PR, I linked to the issue PR in the text above (and used the correct keyword to close issues using keywords) [x] I've updated the documentation if as necessary FOLLOWUP TODO: [ ] If merged, we may need to do a manual rebuild (see #1035) Is 5 MB limitation correct for indexedDB at iOS, Android and browser? I'm new to this topic, but I thought indexedDB was invented to store more. Is it some Cordova plugin's additional limitation? Good catch, no 5MB is not accurate anymore. For recent iOS/Safari version AFAIK there are no real sources and Android's limitation are here: https://stackoverflow.com/a/55483549/1902598 (the most up to date source, I've found). Chrome doesn't really give a limit in a traditional sense, but it tries to not consume all of the disk space. It does this by determining how much remaining space is available and will start implementing quotas if remaining space gets low. Chrome will either try to keep at least 10% of the total storage or 2GB, whichever is lower, of the disk free, and will start implementing quotas once it reaches this threshold. The table below comes directly from the chromium sources, and is accurately as of Chrome 76. // The amount of the device's storage the browser attempts to --   | // keep free. If there is less than this amount of storage free   | // on the device, Chrome will grant 0 quota to origins.   | //   | // Prior to M66, this was 10% of total storage instead of a fixed value on   | // all devices. Now the minimum of a fixed value (2GB) and 10% is used to   | // limit the reserve on devices with plenty of storage, but scale down for   | // devices with extremely limited storage.   | // * 1TB storage -- min(100GB,2GB) = 2GB   | // * 500GB storage -- min(50GB,2GB) = 2GB   | // * 64GB storage -- min(6GB,2GB) = 2GB   | // * 16GB storage -- min(1.6GB,2GB) = 1.6GB   | // * 8GB storage -- min(800MB,2GB) = 800MB https://chromium.googlesource.com/chromium/src.git/+/refs/tags/76.0.3775.5/storage/browser/quota/quota_settings.cc Not sure about iOS/Safari. @dfahlander - sorry for the mention, but do you happen to know recent iOS and/or Safari quotas? It's worth noting that the behaviour in web browsers can also differ from the behaviour in web views. @timbru31 I'm sorry but I don't know the quotas for iOS/Safari. @dfahlander @brodybits @timbru31 did you have more infos about indexedDb allowed memory size on iOS, and also on Android ? Thanks a lot ! @aarbel On android using the chrome browser, you can refer to this table https://github.com/apache/cordova-docs/pull/1042#issuecomment-611551394 It isn't clear on the quotas for wkwebview for ios. It isn't documented and afaik wkwebview isn't open sourced. No, Safari is, at least to me, a blackbox. Last person to ask/tag is @beidson from Apple: Do you know any current size limits of the IndexedDB in iOS? After PR #1043 is merged in, this PR should be rebased. The only other minor feedback I have is that we do not need to break up the lines. E.g. IndexedDB is supported by the underlying WebView on all platforms, with known limitations on the following platforms: This could be written on one line. IndexedDB is supported by the underlying WebView on all platforms, with known limitations on the following platforms: IMO, I look at markdown as a text document. When I write a technical specification document in Word, I usual do not do something like this. But this is just my opinion and not a blocker. Brody, could you rebase this PR so we can merge this in? Travis did not like the rebase and rapid commits. Once https://travis-ci.org/github/apache/cordova-docs/builds/741877609 finishes, I'll merge this in. Thanks Tim for taking on the edit and merging!
gharchive/pull-request
2019-12-30T01:56:43
2025-04-01T04:33:28.107940
{ "authors": [ "Aarbel", "breautek", "brodybits", "dfahlander", "dpogue", "erisu", "timbru31", "vdshb" ], "repo": "apache/cordova-docs", "url": "https://github.com/apache/cordova-docs/pull/1042", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
579199457
How to take a photo or select a file? How do I make sure that when I activate the camera, I can select whether I want to take a photo or select the image in the gallery? document.addEventListener("deviceready", onDeviceReady, false); function onDeviceReady() { $(document).on("click", ".tirarFoto", tirarFoto); function tirarFoto(e) { navigator.camera.getPicture(onSuccess, onFail,{ quality: 35, saveToPhotoAlbum: true, destinationType: Camera.DestinationType.DATA_URL, targetWidth: 800, targetHeight: 0 }); function onSuccess(imageData) { img_.src = "data:image/jpeg;base64," + imageData; input_.value = img_.src; $(input_).change(); } function onFail(message){ alert('A camera falhou: ' + message); } } } you have to create a custom UI for asking (or use dialogs plugin) and based on the user answer configure the camera plugin, this is not supported by the camera plugin
gharchive/issue
2020-03-11T11:36:30
2025-04-01T04:33:28.109848
{ "authors": [ "hostcia", "jcesarmobile" ], "repo": "apache/cordova-plugin-camera", "url": "https://github.com/apache/cordova-plugin-camera/issues/570", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
348335998
Support renaming a database @davisp: Distributed systems makes renaming things hard since the dbs db is eventually consistent. For a stop gap, a tool to do the rename offline (ie, stop all nodes in the cluster) could probably be done easily enough. @janl: good call, I thought that was going to a be a thing. First reaction was maybe _dbs docs allow for an alias that all nodes can converge towards, dunno if that actually works tho @davisp: Also, some of the obvious changes we could make here to do this in an eventually consistent fashion: Replace the dbname -> shard file name mapping, ie, when we create a database document it stores a UUID that we reference on disk. Similar to UUID but the "rename" or "alias" doc just lists the original filenames on disk The downsides to these approaches is that there's currently a large amount of operator knowledge based on the fact that database names map to file system locations. If we break that mapping then we'll have to work on tooling for anyone operating a CouchDB cluster. Also, with alias docs we'd end up having a single database available under two different names without some other sort of voodoo. So that might be something to think about (though it seems fine on the face of it, just maybe surprising to customers in corner cases). @wohali: There actually was a request for DB name aliases 5 years ago in the 2012 summit - so that is very on topic @davisp. Also, this would help with things like adjusting a db's q (replicate to a new db then change the alias when it's ready). It occurs to me that this would help with compaction, too? Noting that this one becomes much easier in 4.0. We are already aliasing the DB name to a short key prefix, so renaming a database only requires a (transactional) update of a single KV. Also gotta bump the db version key so we invalidate all of the cached db handles.
gharchive/issue
2018-08-07T14:03:41
2025-04-01T04:33:28.118651
{ "authors": [ "davisp", "kocolosk", "wohali" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/issues/1502", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1274816395
Replicator giving issue , with jobs getting into crashed type Hi team, i have a 3 node couch cluster deployed on kubernetes, replication is set in a way like A-B, B-A, B-C, C-B, A-C, C-A using replicator, with continuous replication enabled. From UI, i can see no issues with replication and everything is replicated properly but when i see the logs, i am seeing following logs, Replicator GET ****** failed due to error sel_conn_closed Can someone please help, with this issue? can this logs be ignored? Also, one more thing would like to share is with scheduler/docs api, i do not see anything in the error count, its 0 It means the _changes feed connection is closed unexpectedly. If the error count is 0, it happens rarely enough that the error count is reset. So it's probably more rare than once a minute or so. A few things to try: Check if you're using a restrictive replication filter. Increase the replicator connection timeout configuration to say 90000(msec). Check if your load balancer (if you have one), has an idle connection timer then bump that too and then, check that it's not set to buffer connection data. In general replicator will retry the replications on failure and continue where it left off. It's pretty resilient like that. So, if you don't want to track down the connection issues, and it seems your data replicates as intended you probably don't have to worry about it. Hi @nickva Thank you for the quick reply, i would need a further help as i am pretty new to couch Where do i check this? restrictive replication filter. Also can i put this connection timeout under local.d directory in any ini file? Thanks Pallavi C If your replication document doesn't have a filter specified then it's not a concern. You can put the timeout in any local..d or local.ini file https://docs.couchdb.org/en/stable/config/replicator.html#replicator/connection_timeout Closing as there has been no activity in a while. Hopefully the suggested reply helped.
gharchive/issue
2022-06-17T09:43:12
2025-04-01T04:33:28.124397
{ "authors": [ "cpallavi", "nickva" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/issues/4066", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
328604971
[5/5] Clustered Purge Implementation Overview This PR implements clustered purge. Its big. There are roughly five large chunks for clustered purge: Single shard APIs Updates to couch_mrview Internal replication Read-repair Clustered API At the shard level the major changes are to add two new indexes for purge requests. These indexes store the history of purge requests to a shard (rather than the old method of just storing the most recent purge). This allows for secondary indices and internal replication to manage their eventual consistency. This is the bulk of the PR in that its adding both the implementation and a whole bunch of tests for clustered purge. One note on naming internally and in commit notes is that a "purge request" is what's sent in a single HTTP request to the purge end point. A "purge info" is a single {DocId, Revs} entry. The new indexes in the shard deal with purge infos as they lose any notion of bundling from which request they came in after they're stored. The changes to couch_mrview are fairly straightforward after the single node API changes. The only thing that will be interesting in this commit is how we use a _local doc to track where the secondary index has processed in the purge sequence. This is necessary so that compaction knows when it can discard purge infos. Internal replication is fairly straightforward. This update just ensures that we synchronize our purge infos between each shard so that we don't inadvertently undo a purge request when synchronizing shards. Read-repair is somewhat tricky. The thing to note here is that when we open a document we track which revisions came from which nodes. Then if read-repair is required we pass that information along with the update request. This way a node can filter out requests from any node that is either a) not up to date with its synchronization or b) we have explicitly purged a revision that has not been sent to the node from whence the revision came back. That last bit of logic may sound kind of odd but the thing to remember is that we could completely remove a revision from a cluster and then re-introduce it later on. This means that we have to reject re-application of a revision for a closed period of time (rather than rejecting it forever). The clustered API is rather straightforward for anyone familiar any of the existing fabric coordinator and HTTPD handler code. Testing recommendations make check Clustered purge comes with some fairly extensive testing though I'm very much open for suggestions on new tests. Related Issues or Pull Requests This PR depends on: #1366 #1367 #1368 #1369 Checklist [x] Code is written and works correctly (hopefully ;) [x] Changes are covered by tests; [ ] Documentation reflects the changes; For docs this doesn't change the existing purge APIs but I will need to open a PR against the docs to write up some of the caveats and footgun warnings that we want to add for this. Also of note, I highly recommend moving through this PR a commit at a time as its quite big and the commits are specifically written to try and separate out all of the major changes to ease review. Thanks @kocolosk for your review. I tried to address most of your comments on commits 01, 02 and 03 in new commit 7e0dabb. Will work on other comments. @davisp hey Paul, feel free to add your response :-) Very nice work all. Historically one could think about external consumers of _changes and internal consumers of the sequence index as being on equal footing, and there's something appealing about that. The latter group has some efficiency advantages but fundamentally an external process could go and do everything that the internal consumers do using the node-local HTTP interface. The lack of HTTP access to the purge index breaks that. An external indexer, for example, has no way of knowing revisions have been purged from the main database short of checking the individual shard's purge sequence and rebuilding the entire index if it changes, whereas the internal consumers can be smarter. All that said, I wouldn't change anything about this PR, as it contains everything needed to surface an external API to the purge structures if we should conclude there's a need for it at a later date. I'm +1 pending resolution of my (minor) review comments. Thanks again for your review @kocolosk Hey @kocolosk @davisp, I tried to address comments related to code change in commit 9038557 and other previous commits. Any additional comments are highly welcome. Very nice work. 💯 for a thorough test suite. I just added commit https://github.com/apache/couchdb/pull/1370/commits/a5c1d1330c8051a9912871049ead60869745c4db to address issue of exposing verify_module and verify_function for unexpected execution. Currently couch_epi:any/5 is used to listen whether there is client reporting a valid index. In addition, it looks that there is problem with couch_db:get_design_docs/1 and I stripped the suffix (the part after dot) before calling fabric:design_docs/1. Hey @iilyak I used 1dae6af to adjust returned value from get_purge_seq/1 and get_oldest_purge_seq/1, and use 8a62f68 to add information about module and line. Also added explantation why we use proplists:get_value/2 in fabric_rpc. I'm +1 pending resolution of my review comments. However I am concerned with deployment aspect of the PR. It is scary that we don't have a rollback. Could we at least have a documentation on what we would do if we would need to rollback to a version which doesn't have additional trees in the header? Ideally I would prefer to have a PR (which we would not be merging) with an implementation of a header downgrade. thanks again @iilyak I added one response https://github.com/apache/couchdb/pull/1370#discussion_r201238042. Hopefully there is no pending comment from your side except for downgrade database. I am working on downgrade case, but this will be in different PR. @davisp @jiangphcn With 2.2.0 down to just 4 open issues, and already with PSE in there, let's release that first. Please hold merging these purge PRs until the 2.2.0 release is cut. I anticipate that in the next 2 weeks or less. okay @wohali thanks for letting us know. Hi Paul, @davisp When using git log --oneline, the order is correct. localhost:src jiangph$ git log --oneline bf809a8 [10/10] Clustered Purge: Clustered HTTP API 110dcdc [09/10] Clustered Purge: Fabric API 4dcfb50 [08/10] Clustered Purge: Update read-repair 973a599 [07/10] Clustered Purge: Internal replication a947b5c [06/10] Clustered Purge: Update mrview indexes 3279c38 [05/10] Clustered Purge: Add upgrade tests a192aef [04/10] Clustered Purge: Update couch_pse_tests ac8d0f6 [03/10] Clustered Purge: Update couch_bt_engine 4d695f1 [02/10] Clustered Purge: Update single node APIs 0254305 [01/10] Clustered Purge: Define new purge API 830240b Create new on_compact trigger 4c87d78 Enhance PSE tests with setup/teardown functions c998aa4 Update to use new couch_pse_tests app 0d85abe Update PSE test definitions for new util module 3c94bc4 Rename PSE test modules db28f4c Move PSE tests to their own app f45c036 Rewrite the PSE test suite to use couch_server da43788 Simplify logic in mem3_rep aea2833 Update fabric_doc_open eunit tests bb27d5e Fix race on couch_db:reopen/1 151b9b0 Fix default security object handling da2e22a Fix bug during purge 3852a00 Fix typos in couch_db_engine.erl Confirmed locally through not GitHub. Must be ordered on dates or something. +1 to merge
gharchive/pull-request
2018-06-01T17:44:02
2025-04-01T04:33:28.139077
{ "authors": [ "davisp", "iilyak", "jiangphcn", "kocolosk", "wohali" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/pull/1370", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
233716428
Don't wrap real error in badmatch exception Overview A change in apache/couchdb#576 made any unexpected return from functions called in get_view_index_state/4 to be presented as badmatch and hide the real error. This unwraps the badmatch and re-throw it as an original response. Testing recommendations A test suite make javascript suites=view_errorsjsshould pass. Related Pull Requests Bug introduced in apache/couchdb#576 Checklist [x] Code is written and works correctly; [ ] Changes are covered by tests; [ ] Documentation reflects the changes; You fixed it! +1
gharchive/pull-request
2017-06-05T21:29:30
2025-04-01T04:33:28.142682
{ "authors": [ "eiri", "wohali" ], "repo": "apache/couchdb", "url": "https://github.com/apache/couchdb/pull/577", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1021012664
Restructure Scala code to support JDK 8 Follow up to https://github.com/jw3/example-daffodil-vscode/issues/94: @arosien said Daffodil required JDK 8, so we could try to also support it We currently need JDK 11 to support use of Path.of, which could perhaps be replaced with something available in older JVMs. Nested object in Scala isn't supported until JDK 9 or greater. Original issue https://github.com/jw3/example-daffodil-vscode/issues/136 I'm going to reopen this issue. We've solved the compile on Java 8 issue, but not the run on Java 8. Running is much lower priority, but it might want to be fixed some day. Is Java 8 still something we need to support or are we requiring Java 11 or greater? Since Daffodil supports Java 8, and there are systems that a still on Java 8, I think there would be value in being able to debug on Java 8 to ensure environments are the same. This doesn't feel high priority to me though.
gharchive/issue
2021-10-08T11:54:00
2025-04-01T04:33:28.145682
{ "authors": [ "jw3", "scholarsmate", "stevedlawrence" ], "repo": "apache/daffodil-vscode", "url": "https://github.com/apache/daffodil-vscode/issues/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
824548602
Update jackson-core to 2.11.4 Updates com.fasterxml.jackson.core:jackson-core from 2.11.3 to 2.11.4. I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! Files still referring to the old version number The following files still refer to the old version number (2.11.3). You might want to review and update them manually. daffodil-cli/bin.NOTICE Ignore future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "com.fasterxml.jackson.core", artifactId = "jackson-core" } ] labels: library-update, semver-patch, old-version-remains I would like us to have a wiki page on our confluence discussing the review criteria for these automated updates. I.e., a checklist of what we should be ckecking on to convince ourselves that one of these upgrades is acceptable. Agreed, I'll create a page. I agree we need a checklist. I just realized I overlooked that daffodil-cli/bin.NOTICE still mentions jackson-core's old version number. The review criteria should include: A determination whether the version update is a patch, minor, or major update. Major updates need more stringent review criteria than the other updates, at least a manual reading of the library's release notes / changelog to check for API changes. How to update any files like daffodil-cli/bin.NOTICE. Should we push a commit to each and every pull request updating that page AND squash the commits together before merging each pull request? Or update the daffodil-cli/bin.NOTICE file once in one last manually made pull request? I've created this page as a starting point for documentation: https://cwiki.apache.org/confluence/display/DAFFODIL/Scala+Steward Feel free to add anything if I missed anything. I also added a way I suggest we make changes, which is exactly the same as our normal work flow except any changes are pushed to the appropriate branch on the scala-steward/daffodil fork rather than our own fork. I perfer it done this way (one for every fork) so that the version bump and any related changes (e.g. API updates, license changes) are in the same commit. This might be a bit of a pain for this first go around, but hopefully things haven't changed their licenses. And once we get through this big set, it should be easier to keep up with it. Another thought, we might want to update our LICENSE/NOTICE files to say something like <VERSION.jar or *.jar so that we don't need to update the LICENSE/NOTICE file for every version bump. That's a bit of a pain, and I'm not sure we gain much extra by including specific jar versions in these files. Yes, let's replace the explicit version numbers in LICENSE/NOTICE files with foo-.jar to reduce the number of times we need to update our LICENSE/NOTICE files manually to only whenever the library changes its own LICENSE. I found a problem with the first git command in the Confluence page https://cwiki.apache.org/confluence/display/DAFFODIL/Scala+Steward>. I logged into Confluence (I have an account with the username "interran" and name "John Interrante") but I don't have the ability to edit the page or add a comment to it so I will have to suggest the correction here. The first git command adds the wrong repo URL as the remote URL (right now it adds Scala Stewart's own github repo as the url, not Scala Stewart's github fork of the Daffodil repo). Please change it to: git remote add scala-steward git@github.com:scala-steward/daffodil.git Note I'm not sure if we will have permission to push a commit to the Scala Stewart fork, but let's find out. @tuxji I've given you full permissions to the wiki. Also, GitHub has a feature that lets owners of a repo to push to fork branches that have been submitted as pull requests as long as the pull request creator allows it (which is the default). Scala steward allows this, so we should be able to push to their daffodil fork. We don't HAVE to mention the branch name in BOTH the second and third git commands in the Confluence page. We could simply fetch from all our upstream repos and only then say which branch we want to check out: interran@GH3WPL13E:~/apache/daffodil-asf$ git fetch --all Fetching origin Fetching scala-steward remote: Enumerating objects: 53, done. remote: Counting objects: 100% (53/53), done. remote: Compressing objects: 100% (13/13), done. remote: Total 55 (delta 40), reused 53 (delta 40), pack-reused 2 Unpacking objects: 100% (55/55), 12.81 KiB | 184.00 KiB/s, done. From github.com:scala-steward/daffodil * [new branch] master -> scala-steward/master * [new branch] runtime2-2202 -> scala-steward/runtime2-2202 * [new branch] update/Saxon-HE-9.9.1-8 -> scala-steward/update/Saxon-HE-9.9.1-8 * [new branch] update/icu4j-68.2 -> scala-steward/update/icu4j-68.2 * [new branch] update/jackson-core-2.11.4 -> scala-steward/update/jackson-core-2.11.4 * [new branch] update/jansi-1.18 -> scala-steward/update/jansi-1.18 * [new branch] update/junit-4.13.2 -> scala-steward/update/junit-4.13.2 * [new branch] update/sbt-1.4.8 -> scala-steward/update/sbt-1.4.8 * [new branch] update/sbt-native-packager-1.7.6 -> scala-steward/update/sbt-native-packager-1.7.6 * [new branch] update/scala-library-2.12.13 -> scala-steward/update/scala-library-2.12.13 * [new branch] update/scalacheck-1.15.3 -> scala-steward/update/scalacheck-1.15.3 * [new branch] update/scallop-4.0.2 -> scala-steward/update/scallop-4.0.2 * [new branch] update/typesafe-1.4.1 -> scala-steward/update/typesafe-1.4.1 * [new branch] update/woodstox-core-6.2.4 -> scala-steward/update/woodstox-core-6.2.4 * [new branch] update/xercesImpl-2.12.1 -> scala-steward/update/xercesImpl-2.12.1 interran@GH3WPL13E:~/apache/daffodil-asf$ git checkout update/jackson-core-2.11.4 Branch 'update/jackson-core-2.11.4' set up to track remote branch 'update/jackson-core-2.11.4' from 'scala-steward'. Switched to a new branch 'update/jackson-core-2.11.4' interran@GH3WPL13E:~/apache/daffodil-asf$ I figured it would be cleaner to just do it on an individual branch basis so it doesn't pollute everything. Maybe steward will delete these branches once they are merged though, so maybe it's not too bad to just fetch everything. I'll update. I've downloaded jackson-core-jackson-core-2.11.3.tar.gz, jackson-core-jackson-core-2.11.4.tar.gz, and jackson-core-jackson-core-2.12.2.tar.gz, diff'ed them, found their NOTICE files, and compared their NOTICE file with our bin.NOTICE. I believe we can safely update to 2.11.4 and then 2.12.2 after that since I didn't see any API changes that should affect us. I've incorporated jackson-core's updated NOTICE in our bin.NOTICE and replaced all version numbers in our bin.LICENSE and bin.NOTICE with the literal string <VERSION> to avoid needing to change bin.LICENSE and bin.NOTICE unless dependencies actually change their LICENSE or NOTICE files. Once I get someone else's +1, I will squash, merge, and move on to the next dependency PR. +1 looks good to me!
gharchive/pull-request
2021-03-08T13:17:10
2025-04-01T04:33:28.158813
{ "authors": [ "mbeckerle", "scala-steward", "stevedlawrence", "tuxji" ], "repo": "apache/daffodil", "url": "https://github.com/apache/daffodil/pull/496", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2581975144
Remove UI As part of effort outlined in https://github.com/apache/datafusion-ballista/pull/1066 and https://github.com/apache/datafusion-ballista/issues/1067 this PR removes UI package from core ballista. Relates to: https://github.com/apache/datafusion-ballista/pull/1066 & https://github.com/apache/datafusion-ballista/issues/1067 Which issue does this PR close? Closes #. Rationale for this change What changes are included in this PR? Are there any user-facing changes? should be removed now, plus few extra lines i found done Looks like there is still a reference in the docker scripts: + cd ballista/scheduler/ui /home/builder/builder-entrypoint.sh: line 27: cd: ballista/scheduler/ui: No such file or directory
gharchive/pull-request
2024-10-11T18:40:50
2025-04-01T04:33:28.163115
{ "authors": [ "andygrove", "milenkovicm" ], "repo": "apache/datafusion-ballista", "url": "https://github.com/apache/datafusion-ballista/pull/1072", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2607572167
[Minor] Update info/list of TPC-DS queries Which issue does this PR close? Keep failing queries up to date, refer to new issue. Rationale for this change What changes are included in this PR? Are these changes tested? Are there any user-facing changes? thanks for the review @comphead
gharchive/pull-request
2024-10-23T07:30:10
2025-04-01T04:33:28.165448
{ "authors": [ "Dandandan" ], "repo": "apache/datafusion", "url": "https://github.com/apache/datafusion/pull/13075", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1040858523
[Bug] [Master] Failed to operate workflow node separately Search before asking [X] I had searched in the issues and found no similar issues. What happened Failed to operate workflow node alone. No new task instance is generated. version: dev There are no problems in the version of 2.0-alpha-release. What you expected to happen Operate workflow node separately and generate new task instance. How to reproduce Operate workflow node separately. Anything else [ERROR] 2021-11-01 15:30:35.452 org.apache.dolphinscheduler.dao.utils.DagHelper:[112] - start node name [a] is not in task node list [[TaskNode{id='null', code=851434614521856, version=1, name='a', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"a\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[]', preTaskNodeList=null, extras='null', depList=[], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}]] [ERROR] 2021-11-01 15:30:35.452 org.apache.dolphinscheduler.dao.utils.DagHelper:[112] - start node name [a] is not in task node list [[TaskNode{id='null', code=851434614521856, version=1, name='a', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"a\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[]', preTaskNodeList=null, extras='null', depList=[], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}]] [ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[512] - processDag is null [ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.dao.utils.DagHelper:[112] - start node name [job_b1] is not in task node list [[TaskNode{id='null', code=851435467538432, version=1, name='job_b1', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"b1\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[851437597130752]', preTaskNodeList=null, extras='null', depList=[851437597130752], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}, TaskNode{id='null', code=851437597130752, version=2, name='depend_a', desc='', type='DEPENDENT', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"waitStartTimeout":{"strategy":"FAILED","interval":null,"checkInterval":null,"enable":false},"switchResult":{}}', preTasks='[]', preTaskNodeList=null, extras='null', depList=[], dependence='{"relation":"AND","dependTaskList":[{"relation":"AND","dependItemList":[{"projectCode":851275541217280,"definitionCode":851435217125376,"depTasks":"a","cycle":"day","dateValue":"today"}]}]}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}, TaskNode{id='null', code=851436591874048, version=1, name='job_b2', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"b2\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[851437597130752]', preTaskNodeList=null, extras='null', depList=[851437597130752], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}]] [ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[512] - processDag is null [ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[512] - processDag is null [ERROR] 2021-11-01 15:30:35.454 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[195] - handler error: java.lang.NullPointerException: null at org.apache.dolphinscheduler.dao.utils.DagHelper.parsePostNodes(DagHelper.java:301) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.submitPostNode(WorkflowExecuteThread.java:797) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.startProcess(WorkflowExecuteThread.java:464) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.run(WorkflowExecuteThread.java:192) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) [ERROR] 2021-11-01 15:30:35.454 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[195] - handler error: java.lang.NullPointerException: null at org.apache.dolphinscheduler.dao.utils.DagHelper.parsePostNodes(DagHelper.java:301) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.submitPostNode(WorkflowExecuteThread.java:797) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.startProcess(WorkflowExecuteThread.java:464) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.run(WorkflowExecuteThread.java:192) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) [ERROR] 2021-11-01 15:30:35.454 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[195] - handler error: java.lang.NullPointerException: null at org.apache.dolphinscheduler.dao.utils.DagHelper.parsePostNodes(DagHelper.java:301) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.submitPostNode(WorkflowExecuteThread.java:797) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.startProcess(WorkflowExecuteThread.java:464) at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.run(WorkflowExecuteThread.java:192) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) Are you willing to submit PR? [ ] Yes I am willing to submit a PR! Code of Conduct [X] I agree to follow this project's Code of Conduct The hi error is the dev code has a judgment logic, I will fix it today @dailidong Can you assign this problem to me hi, this error is the dev code has a judgment logic, I will fix it today @dailidong Can you assign this problem to me done
gharchive/issue
2021-11-01T07:33:49
2025-04-01T04:33:28.175383
{ "authors": [ "boy-xiaozhang", "dailidong", "shangeyao" ], "repo": "apache/dolphinscheduler", "url": "https://github.com/apache/dolphinscheduler/issues/6657", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1423661368
Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.1:compile failed: invalid versionNumber : major.minor Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.1:compile failed: invalid versionNumber : major.minor Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.1:compile failed: invalid versionNumber : major.minor
gharchive/issue
2022-10-26T08:48:47
2025-04-01T04:33:28.177432
{ "authors": [ "Cyb315966632" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/issues/13687", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1560743371
refactor bind slot, function and do type coercion in one rule Proposed changes Issue Number: close #xxx Problem summary Describe your changes. Checklist(Required) Does it affect the original behavior: [ ] Yes [ ] No [ ] I don't know Has unit tests been added: [ ] Yes [ ] No [ ] No Need Has document been added or modified: [ ] Yes [ ] No [ ] No Need Does it need to update dependencies: [ ] Yes [ ] No Are there any changes that cannot be rolled back: [ ] Yes (If Yes, please explain WHY) [ ] No Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... TeamCity pipeline, clickbench performance test result: the sum of best hot time: 34.17 seconds load time: 524 seconds storage size: 17122260141 Bytes https://doris-community-test-1308700295.cos.ap-hongkong.myqcloud.com/tmp/20230128113743_clickbench_pr_85630.html
gharchive/pull-request
2023-01-28T07:53:06
2025-04-01T04:33:28.183597
{ "authors": [ "hello-stephen", "morrySnow" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/16156", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1679846682
Chore add some regression-test case about materialized-view Proposed changes add some regression-test case about materialized-view Problem summary Describe your changes. Checklist(Required) [ ] Does it affect the original behavior [ ] Has unit tests been added [ ] Has document been added or modified [ ] Does it need to update dependencies [ ] Is this PR support rollback (If NO, please explain WHY) Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall
gharchive/pull-request
2023-04-23T04:01:51
2025-04-01T04:33:28.186654
{ "authors": [ "BiteTheDDDDt" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/18946", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1827962547
feature refactor CSV reading process during scanning, and support enclose and escape for stream load Proposed changes Refactor thoughts: close #22383 Descriptions about enclose and escape: #22385 Further comments Cases should be discussed: When an incomplete enclose appears in the beginning of a large scale data, the line delimiter will be unreachable till the EOF, will the buffer become extremely large? What if an infinite line occurs in the case? Essentially, 1. is equivalent to this. Only support stream load as trial in this PR, avoid too many unrelated changes. Docs will be added when enclose and escape is available for all kinds of load. run buildall run buildall run buildall run buildall run buildall run buildall run buildall run buildall run buildall run p0 run p0 run p0 run p0 run click bench run clickbench run p0 Seems weird, p0 has passed as shown. Correctness is verified, now will resolve the performance problem. run buildall run buildall run clickbench run buildall run buildall run buildall run build all run buildall run clickbench run p0
gharchive/pull-request
2023-07-30T14:49:15
2025-04-01T04:33:28.194366
{ "authors": [ "TangSiyang2001", "Yukang-Lian", "dataroaring" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/22382", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1851194115
fixbind slot failed because of "default_cluster" Proposed changes slot bind failed for following querys: select tpch.lineitem.* from lineitem select tpch.lineitem.l_partkey from lineitem the unbound slot is lineitem.l_partkey, but the bounded slot is default_cluster:tpch.lineitem.l_partkey. They are not matched. we need to ignore default_cluster: when compare dbName Issue Number: close #xxx Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall (From new machine)TeamCity pipeline, clickbench performance test result: the sum of best hot time: 46.95 seconds stream load tsv: 510 seconds loaded 74807831229 Bytes, about 139 MB/s stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s stream load parquet: 31 seconds loaded 861443392 Bytes, about 26 MB/s insert into select: 29.2 seconds inserted 10000000 Rows, about 342K ops/s storage size: 17162181158 Bytes run buildall (From new machine)TeamCity pipeline, clickbench performance test result: the sum of best hot time: 46.14 seconds stream load tsv: 509 seconds loaded 74807831229 Bytes, about 140 MB/s stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s stream load parquet: 32 seconds loaded 861443392 Bytes, about 25 MB/s insert into select: 29.2 seconds inserted 10000000 Rows, about 342K ops/s storage size: 17162677169 Bytes run buildall (From new machine)TeamCity pipeline, clickbench performance test result: the sum of best hot time: 45.51 seconds stream load tsv: 512 seconds loaded 74807831229 Bytes, about 139 MB/s stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s stream load parquet: 30 seconds loaded 861443392 Bytes, about 27 MB/s insert into select: 29.2 seconds inserted 10000000 Rows, about 342K ops/s storage size: 17162113308 Bytes
gharchive/pull-request
2023-08-15T10:02:21
2025-04-01T04:33:28.201342
{ "authors": [ "englefly", "hello-stephen" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/23008", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1873279069
Bug fix load db use analyzer to analyze diffrent metaindex Proposed changes fix load db use analyzer to analyze diffrent metaindex Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall (From new machine)TeamCity pipeline, clickbench performance test result: the sum of best hot time: 48.79 seconds stream load tsv: 534 seconds loaded 74807831229 Bytes, about 133 MB/s stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s stream load orc: 64 seconds loaded 1101869774 Bytes, about 16 MB/s stream load parquet: 32 seconds loaded 861443392 Bytes, about 25 MB/s insert into select: 28.9 seconds inserted 10000000 Rows, about 346K ops/s storage size: 17162010827 Bytes
gharchive/pull-request
2023-08-30T09:18:12
2025-04-01T04:33:28.204301
{ "authors": [ "BiteTheDDDDt", "hello-stephen" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/23673", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1931231909
FixQuery the transaction status NPE when the task has not yet started scheduling error log When the task has not yet started scheduling, there is no transaction status at this time. Therefore the query will cause an NPE exception show routine load task where JobName="npe" 2023-10-07 13:52:16,883 WARN (mysql-nio-pool-3|1310) [StmtExecutor.executeByLegacy():807] execute Exception. stmt[1368, fc075cfd62aa4459-a8014ca6b5f574fa] java.lang.NullPointerException: null at org.apache.doris.load.routineload.RoutineLoadJob.lambda$getTasksShowInfo$6(RoutineLoadJob.java:1452) ~[classes/:?] at java.util.ArrayList.forEach(ArrayList.java:1259) ~[?:1.8.0_362] at org.apache.doris.load.routineload.RoutineLoadJob.getTasksShowInfo(RoutineLoadJob.java:1450) ~[classes/:?] at org.apache.doris.qe.ShowExecutor.handleShowRoutineLoadTask(ShowExecutor.java:1605) ~[classes/:?] at org.apache.doris.qe.ShowExecutor.execute(ShowExecutor.java:316) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.handleShow(StmtExecutor.java:2208) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.executeByLegacy(StmtExecutor.java:775) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.execute(StmtExecutor.java:468) ~[classes/:?] at org.apache.doris.qe.StmtExecutor.execute(StmtExecutor.java:438) ~[classes/:?] at org.apache.doris.qe.ConnectProcessor.handleQuery(ConnectProcessor.java:353) ~[classes/:?] at org.apache.doris.qe.ConnectProcessor.dispatch(ConnectProcessor.java:501) ~[classes/:?] at org.apache.doris.qe.ConnectProcessor.processOnce(ConnectProcessor.java:752) ~[classes/:?] at org.apache.doris.mysql.ReadListener.lambda$handleEvent$0(ReadListener.java:52) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) ~[?:1.8.0_362] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ~[?:1.8.0_362] Changes When the task has not yet started scheduling, that is, when txn_id is -1, the transaction status is set to null Test run buildall Run P0 run p0
gharchive/pull-request
2023-10-07T07:15:58
2025-04-01T04:33:28.206967
{ "authors": [ "CalvinKirs" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/25074", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1975480012
FIXFix cast decimalv2 2 decimalv3 Proposed changes Issue Number: close #xxx Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run p0
gharchive/pull-request
2023-11-03T05:25:24
2025-04-01T04:33:28.208919
{ "authors": [ "amorynan" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/26368", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2040110132
chore Add MVN_OPT env variable to enrich building FE with extra options e.g. just export or add it to custom_env.sh export MVN_OPT="-o" will build FE with maven option "-o" (offline), which means maven does not need to download meta from maven repo, it is useful for saving time if the internet is unstable or unusable. run buidlall run buildall run buildall run buildall
gharchive/pull-request
2023-12-13T16:49:29
2025-04-01T04:33:28.211442
{ "authors": [ "gavinchou" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/28375", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2047780550
refactormake NormalizeAggregate rule more clear and readable Proposed changes Issue Number: close #xxx Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run buildall run buildall run buildall
gharchive/pull-request
2023-12-19T01:59:58
2025-04-01T04:33:28.213639
{ "authors": [ "starocean999" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/28607", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2143662213
Fix BE gracefully exit when port conflicts Proposed changes Issue Number: close #xxx almost fixed. still some bugs in rocksdb itself when in Release mode before: coredump in somewhere now log: Doris BE server did not start correctly, exiting Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run buildall
gharchive/pull-request
2024-02-20T06:53:02
2025-04-01T04:33:28.215919
{ "authors": [ "zclllyybb" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/31143", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2168165399
fix json_object can't be odd parameters, need even parameters Proposed changes Issue Number: close #31766 Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... @iwanttobepowerful please add a testcase
gharchive/pull-request
2024-03-05T02:46:58
2025-04-01T04:33:28.217778
{ "authors": [ "iwanttobepowerful", "xiaokang" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/31767", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2224786605
fix exclude cache ttl Proposed changes Issue Number: close #xxx Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall run buildall
gharchive/pull-request
2024-04-04T08:10:09
2025-04-01T04:33:28.219596
{ "authors": [ "dataroaring" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/33253", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2267468380
test add batch size in long time case Proposed changes For even PRs, the batch size is set to 50, and this case runs for 50 minutes. Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... run buildall
gharchive/pull-request
2024-04-28T10:06:38
2025-04-01T04:33:28.221253
{ "authors": [ "Mryange" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/34249", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2363391117
Fix Fix core in orc and parquet reader sometimes after low mem exception. Proposed changes Backport #36574. run buildall
gharchive/pull-request
2024-06-20T02:43:30
2025-04-01T04:33:28.222449
{ "authors": [ "kaka11chen" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/36575", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2387644716
chore Add more clear error message when loading timed out Proposed changes Issue Number: close #xxx run buildall
gharchive/pull-request
2024-07-03T04:29:16
2025-04-01T04:33:28.223476
{ "authors": [ "zclllyybb" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/37198", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2588059377
fixOrToIn and SimplifyRange rules dead loop #41689 (3.0) Proposed changes pick #41689 Issue Number: close #xxx run buildall run buildall run feut
gharchive/pull-request
2024-10-15T08:42:01
2025-04-01T04:33:28.224949
{ "authors": [ "englefly" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/41887", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2631728506
fix Add synchronized to avoid concurrent modification Without synchronized protection, when the user manually executes cancel, the backupHandler may also trigger cancel simultaneously. Modifying the snapshotInfo by both at the same time will cause a concurrent modification exception. The pending state will synchronously wait for the task of creating a replica to end. Therefore, a creating state needs to be added to convert the synchronous wait into asynchronous polling to avoid blocking the user's cancel/list job info requests. run buildall run buildall run buildall
gharchive/pull-request
2024-11-04T03:16:04
2025-04-01T04:33:28.226599
{ "authors": [ "w41ter" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/43172", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2682227059
[DNM] What problem does this PR solve? Issue Number: close #xxx Related PR: #xxx Problem Summary: Release note None Check List (For Author) Test [ ] Regression test [ ] Unit Test [ ] Manual test (add detailed scripts or steps below) [ ] No need to test or manual test. Explain why: [ ] This is a refactor/code format and no logic has been changed. [ ] Previous test can cover this change. [ ] No code files have been changed. [ ] Other reason Behavior changed: [ ] No. [ ] Yes. Does this need documentation? [ ] No. [ ] Yes. Check List (For Reviewer who merge this PR) [ ] Confirm the release note [ ] Confirm test cases [ ] Confirm document [ ] Add branch pick label run buildall run cloud_p0 run compile run cloud_p0
gharchive/pull-request
2024-11-22T07:51:35
2025-04-01T04:33:28.232975
{ "authors": [ "TangSiyang2001" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/44473", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2721901086
branch-2.1: opt Allow to miss auto-increment column and other value columns in partial update #44528 pick https://github.com/apache/doris/pull/44528 run buildall
gharchive/pull-request
2024-12-06T02:33:25
2025-04-01T04:33:28.234298
{ "authors": [ "bobhan1" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/45073", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2757535382
enhanceAdd the default configuration file for ranger under fe/conf What problem does this PR solve? Previously, it was necessary to copy the configuration file from the official website Afterwards, only the necessary configuration items need to be changed Issue Number: close #xxx Related PR: #xxx Problem Summary: Add the default configuration file for ranger under fe/conf Release note None Check List (For Author) Test [ ] Regression test [ ] Unit Test [ ] Manual test (add detailed scripts or steps below) [x] No need to test or manual test. Explain why: [x] This is a refactor/code format and no logic has been changed. [ ] Previous test can cover this change. [ ] No code files have been changed. [ ] Other reason Behavior changed: [x] No. [ ] Yes. Does this need documentation? [ ] No. [x] Yes. Check List (For Reviewer who merge this PR) [ ] Confirm the release note [ ] Confirm test cases [ ] Confirm document [ ] Add branch pick label Thank you for your contribution to Apache Doris. Don't know what should be done next? See How to process your PR. Please clearly describe your PR: What problem was fixed (it's best to include specific error reporting information). How it was fixed. Which behaviors were modified. What was the previous behavior, what is it now, why was it modified, and what possible impacts might there be. What features were added. Why was this function added? Which code was refactored and why was this part of the code refactored? Which functions were optimized and what is the difference before and after the optimization? run buildall run buildall
gharchive/pull-request
2024-12-24T10:05:00
2025-04-01T04:33:28.242911
{ "authors": [ "hello-stephen", "zddr" ], "repo": "apache/doris", "url": "https://github.com/apache/doris/pull/45869", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }