id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1031178743 | 是否有方法更新combo的大小(不是固定大小fixSize)
[ ] I have searched the issues of this repository and believe that this is not a duplicate.
What problem does this feature solve?
combo在折叠后,如果移动了combo位置,再展开后,combo的大小会变成默认的大小,而不是通过combo内节点的位置计算的大小,需要用户手动移动combo中的node,宽高才会更新。
What does the proposed API look like?
能够让开发者手动调用一个函数来实现combo的大小更新呢?
目前已试过combo.refresh()是达不到我想要的效果。
用:
combo.update({size:[width,height]})解决了
我也遇见过这个问题,收起的时候可以控制大小,但是展开的时候有时候不会包裹住内部的节点,不知道楼主有没有遇见过这个问题
| gharchive/issue | 2021-10-20T09:24:38 | 2025-04-01T04:33:26.082706 | {
"authors": [
"user-lihai",
"yangmiao13160"
],
"repo": "antvis/G6",
"url": "https://github.com/antvis/G6/issues/3263",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2059138955 | item.update无法更新type
G6 Version
4.x
Describe the bug
`const model = {
id: 'node',
type: 'rect',
label: 'node',
style: {
fill: 'red',
},
};
item.update(model);
item.refresh();
`
原来的圆形节点未变成矩形
Your Example Website or App
无
Steps to Reproduce the Bug or Issue
无
Expected behavior
无
Screenshots or Videos
No response
Platform
OS: [e.g. macOS, Windows, Linux]
Browser: [e.g. Chrome, Safari, Firefox]
Version: [e.g. 91.1]
Additional context
No response
已解决,4.8.23版本有此问题,将版本后正常
| gharchive/issue | 2023-12-29T07:03:55 | 2025-04-01T04:33:26.086955 | {
"authors": [
"HogoZhang"
],
"repo": "antvis/G6",
"url": "https://github.com/antvis/G6/issues/5308",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1247335577 | 【仕様変更提案】tagsがなくても投稿できるようにしたい
現在、以下のように
if (!tags) {
console.log(
'\n' +
emoji.get('disappointed') +
' 選択した記事にタグが設定されていません.\n記事を投稿するには一つ以上タグが設定されている必要があります.\n'
);
return -1;
}
tags が設定されていないとエラーが出力されて投稿することができません。
記事を投稿する場合にはタグを設定しない場合もあるので、tagsが入っていなければエラーにして落とすのではなく空の配列にして投稿できるようにする形にした方がいいのかなと思いました。
APIの仕様上できそうになかったのでこのissueはclose とします
| gharchive/issue | 2022-05-25T01:53:03 | 2025-04-01T04:33:26.089919 | {
"authors": [
"TakuKobayashi"
],
"repo": "antyuntyuntyun/qiita-cli",
"url": "https://github.com/antyuntyuntyun/qiita-cli/issues/44",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2613717308 | Update the interface of About us section , making it more responsive and good looking.
I will update the interface of about us section.
about us page is ok . but still if u have a better UI design for the page , then raise pr .
check the pages in profile dropdown , they need major improvement . their footer and navbar is random .
u can redesign the pages to match with the theme ( colors ) of the website .
| gharchive/issue | 2024-10-25T10:30:59 | 2025-04-01T04:33:26.127801 | {
"authors": [
"abhi03ruchi",
"prakharpande04"
],
"repo": "anuragverma108/SwapReads",
"url": "https://github.com/anuragverma108/SwapReads/issues/3994",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2353402835 | Added the author filter for book recommendation
Related Issue
[Cite any related issue(s) this pull request addresses. If none, simply state “None”]
Fixes: #1634
Description
[Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.]
Added a new filter of Author for the book recommendation system.
If the user enters only genre then it will recommend books of that genre.
If the user enter genre and author then it will recommend books of that author in that specified genre.
If such book is not present in the system then it will notify that book is not found.
Type of PR
[ ] Bug fix
[X] Feature enhancement
[ ] Documentation update
[ ] Other (specify): _______________
Screenshots / videos (if applicable)
[Attach any relevant screenshots or videos demonstrating the changes. Make sure to attach before & after screenshots in your PR.]
Books recommended on basis of genre ONLY.
Books recommended based on genre and author BOTH.
Even if the entire author name is not added it still suggests appropriate book.
NO BOOK IS FOUND for that genre and author
Checklist:
[X] I have made this change from my own.
[X] I have taken help from some online resources.
[X] My code follows the style guidelines of this project.
[X] I have performed a self-review of my own code.
[X] I have commented my code, particularly in hard-to-understand areas.
[X] I have made corresponding changes to the documentation.
[X] My changes generate no new warnings.
[X] I have tested the changes thoroughly before submitting this pull request.
[X] I have provided relevant issue numbers and screenshots after making the changes.
Conflicting files
assets/html/book_recommend.html
Please see the conflicts highlighted in the image above. Resolve the conflicts in the mentioned files. Ensure you pull all the latest changes before pushing your feature/bug fixes to the main branch. After resolving the conflicts, please open a new PR for review.
Thank you!
| gharchive/pull-request | 2024-06-14T13:34:52 | 2025-04-01T04:33:26.136000 | {
"authors": [
"Archiesachin",
"anuragverma108"
],
"repo": "anuragverma108/SwapReads",
"url": "https://github.com/anuragverma108/SwapReads/pull/1852",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2606135794 | Extended FAQs Section for Improved Buyer Communication
Related Issue
[Cite any related issue(s) this pull request addresses. If none, simply state “None”]
Fixes: #3808
Description
[Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.]
Type of PR
[ ] Bug fix
[x] Feature enhancement
[ ] Documentation update
[ ] Other (specify): _______________
Screenshots / videos (if applicable)
[Attach any relevant screenshots or videos demonstrating the changes. Make sure to attach before & after screenshots in your PR.]
Checklist:
[x] I have made this change from my own.
[x] My code follows the style guidelines of this project.
[x] I have performed a self-review of my own code.
[x] My changes generate no new warnings.
[x] I have tested the changes thoroughly before submitting this pull request.
[x] I have provided relevant issue numbers and screenshots after making the changes.
@atmajaa could you please merge this???
| gharchive/pull-request | 2024-10-22T18:06:47 | 2025-04-01T04:33:26.141321 | {
"authors": [
"sriharsha0x1"
],
"repo": "anuragverma108/SwapReads",
"url": "https://github.com/anuragverma108/SwapReads/pull/3889",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2657027187 | Tony/21_chrome_support
Pull Request Details
What issue does this PR address
Fixed the issue that video cannot be played on Chrome.
Link to associated issue: #21
Checklist
Complete the check-list below to ensure your branch is ready for PR.
Flutter Style Guide: https://survivor.togaware.com/gnulinux/flutter-style.html
[x] Screenshots included in linked issue
[ ] Changes adhere to the team style and coding guideline
[x] No confidential information
[x] No duplicated content
[x] No lint check errors related to your changes (make prep or flutter analyze lib)
[ ] Pre-exisiting lint errors noted: [HERE]
[ ] Tested on at least one device
[ ] Android Phone
[ ] Android Emulator
[ ] Chrome on Android
[x] Chrome
[ ] iOS
[ ] Linux
[x] MacOS
[ ] Windows
[x] Added 2 reviewers (or 1 for private repositories then they add another)
Finalising
Once PR discussion is complete and 2 reviewers have approved:
[ ] Merge dev into the branch
[ ] Resolve any conflicts
[ ] Add one line summary into CHANGELOG.md
[ ] Bump appropriate version number in pubspec.yaml
[ ] Push to git repository and review
[ ] Merge PR into dev
I am getting a black window on Linux. When I click Play the audio is heard but no video.
I am getting a black window on Linux. When I click Play the audio is heard but no video.
Thanks @gjwgit. I'll check it on Linux.
If my black screen on chrome is not seen on other OS then please merge this PR and have the Linux issue as a separate issue for resolution later on.
If my black screen on chrome is not seen on other OS then please merge this PR and have the Linux issue as a separate issue for resolution later on.
Hi @gjwgit, I will fix it in #30. This issue will be merged first.
| gharchive/pull-request | 2024-11-13T22:40:05 | 2025-04-01T04:33:26.152743 | {
"authors": [
"gjwgit",
"tonypioneer"
],
"repo": "anusii/markdown_widgets",
"url": "https://github.com/anusii/markdown_widgets/pull/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2284435674 | SOLID POD: Secure Key request when entered and press ENTER it should be same as OK
User Story
After I have typed the key here, as a user I am inclined to press ENTER. This should be the same a tapping the OK button.
Implemented in pull request https://github.com/anusii/solidpod/pull/164
To test, go to the keypod demo page and
Click "Forget Security Key"
Turn on the "Encrypt Data" switch
Click "Key Value Table Demo"
Add a key-value pair and click "Submit"
In the security key input widget, type your security key and press the enter key.
Known issue:
If you move the input focus off the text input after typing your security key, then press the enter key seems to do nothing (haven't been able to find a fix for this so far).
Tested and works as expected
| gharchive/issue | 2024-05-08T00:18:37 | 2025-04-01T04:33:26.157253 | {
"authors": [
"cdawei",
"gjwgit"
],
"repo": "anusii/solidpod",
"url": "https://github.com/anusii/solidpod/issues/129",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1566244375 | Non numpy vectorization doc
Being paranoidly explicit about appropriate format of the out object expected from the _evaluate method, per discussion in #372
You may want to do the merge with fastforward/squash, which apparently I can't request on my side.
Thanks again!
Thanks for putting the effort into improving the documentation. I used squash before merging indeed.
The paragraph will be available once I release a new version.
You are most welcome, that was a tiny contribution back to into an excellent project which is giving a good help for my research!
| gharchive/pull-request | 2023-02-01T14:45:26 | 2025-04-01T04:33:26.159289 | {
"authors": [
"blankjul",
"davide-q"
],
"repo": "anyoptimization/pymoo",
"url": "https://github.com/anyoptimization/pymoo/pull/375",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2554014912 | [maketmpl] make template builder
it builds the zip file, the ray-app.json metadata file, and various flavors of readme markdown files..
cc @vigneshka
let me know if this PR is too long. I can break it into smaller ones.
| gharchive/pull-request | 2024-09-28T05:11:54 | 2025-04-01T04:33:26.160655 | {
"authors": [
"aslonnie"
],
"repo": "anyscale/templates",
"url": "https://github.com/anyscale/templates/pull/367",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
453770781 | Lab 3
Fixes #295
Review of colleague's PR #
Changes proposed in this PR:
Complete Lab 3
Add letters function
Add sortLetters function
Add Unit Tests for 100% test coverage
Codecov Report
Merging #296 into master will not change coverage.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #296 +/- ##
=====================================
Coverage 100% 100%
=====================================
Files 18 19 +1
Lines 303 322 +19
=====================================
+ Hits 303 322 +19
Impacted Files
Coverage Δ
03_letters/joshcarp/letters.go
100% <100%> (ø)
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 15b200b...dd0e191. Read the comment docs.
The go-course is now closed. Thank you very much for participating.
| gharchive/pull-request | 2019-06-08T08:27:51 | 2025-04-01T04:33:26.168686 | {
"authors": [
"codecov-io",
"joshcarp",
"juliaogris"
],
"repo": "anz-bank/go-course",
"url": "https://github.com/anz-bank/go-course/pull/296",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
204317347 | ViewControllers not loaded if addViewController not at init
Hi sir,
firstly thank you for your great lib :)
I dynamically load the tabs based on server data.
So i call http first to get my tab list.
Then i construct using [self addViewController:];
It works, i can see the tab bar and the titles.
But the view controllers is not showing (viewDidLoad is not called).
only after i drag to slide (even small drag) then the viewDidLoad triggered and then i can see the content.
Any idea how to solve this?
Thank you
Hi sletith,
I have reproduced the bug you mentioned above. there are two problems I found.
I have pushed a new branch [hotfix] trying to fix it.
You can try the new code.
if the update can solve your problem, I will release a new version.
if it can't, I think you'd better show me some code then I can locate the problem more clearly .
Thank you.
thanks sir. it works ^^
| gharchive/issue | 2017-01-31T14:42:21 | 2025-04-01T04:33:26.187373 | {
"authors": [
"aokizen",
"sleith"
],
"repo": "aokizen/SPSlideTabBarController",
"url": "https://github.com/aokizen/SPSlideTabBarController/issues/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
714188901 | Bugfix: Use upstream instead of target branch
I'm pulling from upstream/master to a branch with a different name (ansible-st2-tracking).
That branch is not in the upstream repo, so this gave the error:
New commits being pulled:
fatal: ambiguous argument 'upstream/ansible-st2-tracking': unknown revision or path not in the working tree.
Use '--' to separate paths from revisions, like this:
'git <command> [<revision>...] -- [<file>...]'
Yep, thanks. :)
| gharchive/pull-request | 2020-10-03T22:35:39 | 2025-04-01T04:33:26.204177 | {
"authors": [
"aormsby",
"cognifloyd"
],
"repo": "aormsby/Fork-Sync-With-Upstream-action",
"url": "https://github.com/aormsby/Fork-Sync-With-Upstream-action/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
679199745 | Added sprig support for Go template
Integrated Sprig shall widen the usage of add command. Tested on 1GB+ CSV with adequate complex usage of add command and Sprig function with decent performance.
I believe this integration shall improve the usage of gocsv.
I have also updated the README for better usage.
@aotimme - I mistakenly pushed the gocsv and gocsv.docx, therefore I closed this PR. I have raised another one, https://github.com/aotimme/gocsv/pull/28
| gharchive/pull-request | 2020-08-14T14:35:34 | 2025-04-01T04:33:26.207190 | {
"authors": [
"munim"
],
"repo": "aotimme/gocsv",
"url": "https://github.com/aotimme/gocsv/pull/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
85899975 | aozorahack用のアイコン
aozorahack のアイコンが欲しいです!
slackのものと同一で問題ないでしょうか?
slackで使ってるアイコンの青空写真はパブリックドメインのものなのでご自由にどうぞ!ATNDのトップと同じです。
2015年6月7日日曜日、genneinotifications@github.comさんは書きました:
aozorahack のアイコンが欲しいです!
slackのものと同一で問題ないでしょうか?
—
Reply to this email directly or view it on GitHub
https://github.com/aozorahack/aozorahack/issues/19.
--
iPhoneより送信
:bow:
| gharchive/issue | 2015-06-07T12:59:12 | 2025-04-01T04:33:26.217719 | {
"authors": [
"gennei",
"kskktk"
],
"repo": "aozorahack/aozorahack",
"url": "https://github.com/aozorahack/aozorahack/issues/19",
"license": "unlicense",
"license_type": "permissive",
"license_source": "bigquery"
} |
233640155 | previous doesn't work with :e buffer
This is probably user error, and not a bug, but when I have buffers open, :bprev and :bnext work as expected, but when I do :e to get a file list, the file list becomes a buffer that I can not :prev.
i.e. I can cycle past it using next, but previous doesn't work. Also, the "tab" at the top isn't highlighted when I'm viewing the ":e" buffer.
Any ideas on how I'm supposed to be handling this situation? Thanks.
That’s due to how the Netrw plugin works (which is what Vim uses to handle directory listings), unfortunately. In short, it puts its directory listing in a nameless hidden buffer. This is what ultimately drove me to write the readdir plugin. Trying to do it differently from Netrw then taught me why Netrw is written as it is – both approaches have their problems. I’m not sure what to do here really. I can’t tell everyone to install Readdir just to fix this, nor should they have to, nor would I recommend it to everyone any more. And in the end the UX is suboptimal either way. Kinda crappy situation.
Ok, thanks for the information about the issue. I can live with it, and I'll try readdir to see if that works for me.
| gharchive/issue | 2017-06-05T16:23:03 | 2025-04-01T04:33:26.220267 | {
"authors": [
"ap",
"equant"
],
"repo": "ap/vim-buftabline",
"url": "https://github.com/ap/vim-buftabline/issues/31",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1205978926 | On Singleplayer it says "You can not have a origin"
I've seen people with this issue but with them it happens on servers, mine is happening on singleplayer-- It didn't give me a prompt to pick origin and I cant even use the orb or commands because it comes up with this.
I own an SMP where we use origins and we're updating to 1.18.2 soon and we're stumped as to why this is happening.
The jar file is labelled "Origins-1.18.2-1.4.1" if that helps.
UPDATE: There is an incompatibility with Charm, is there a way for this patched so we can have both?
Thanks for figuring out which mod causes the incompatibility. I'm currently out of town, so can't work on this, but I'll see what I can do when I'm back next week.
I experienced the same bug today and was able to trace it to an incompatibility with DeepslateCutting (v1.3.2).
UPDATE: There is an incompatibility with Charm, is there a way for this patched so we can have both?
Also experience this with Charm, disabled Charm, works perfectly.
Do any of you experiencing this bug have the mod OK Zoomer installed by any chance?
Do any of you experiencing this bug have the mod OK Zoomer installed by any chance?
I have both mods active and have had no problems.
I have both mods active and have had no problems.
By both mods you mean OK Zoomer and DeepslateCutting? If you haven't had any problems, why have you reported experiencing this bug? Or did you mean Charm and OK Zoomer aren't causing problems for you, and they only happen once you had DeepslateCutting?
By both mods you mean OK Zoomer and DeepslateCutting?
No, I meant OK Zoomer and Origins. Sorry about not being clear. For me it is only Deepslate Cutting and Origins that are causing the error if both mods are active.
Even without OK Zoomer active?
My questions are:
Can you run Origins and DeepslateCutting alone?
Can you run Origins and OK Zoomer alone? You answered yes.
Can you run Origins and DeepslateCutting and OK Zoomer together?
Can you run Origins and DeepslateCutting alone?
Can you run Origins and OK Zoomer alone? You answered yes.
Can you run Origins and DeepslateCutting and OK Zoomer together?
Did some testing on a fresh instance with newly downloaded mods. Here are my results of testing:
Origins + DeepslateCutting -> Origins selectable
Origins + OK Zoomer -> Origins selectable
Origins + DeepslateCutting + OK Zoomer -> Origins not selectable
So it seems I did not test the first configuration before and must have thought the problem being with DeepslateCutting alone and not in combination with another mod. Sorry about that.
Mod versions used:
deepslatecutting-1.18.2-1.3.2
fabric-api-0.51.1+1.18.2
okzoomer-5.0.0-beta.6+1.18.2
Origins-1.18.2-1.4.1
Fabric Loader 0.13.3 (OK Zoomer causes crash on startup with Fabric Loader 0.14.0)
Having this problem with 1.20.1 and none of the aforementioned mods
| gharchive/issue | 2022-04-16T01:41:31 | 2025-04-01T04:33:26.230513 | {
"authors": [
"Wahngrok",
"Whocraf",
"apace100",
"skxrlxt",
"wiryfuture"
],
"repo": "apace100/origins-fabric",
"url": "https://github.com/apace100/origins-fabric/issues/523",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
239266879 | Dynamic Allocation: Too many executors than desired?
On running a Spark Pi with dynamic allocation enabled, I see the following somewhat consistently.. When I run the job with these arguments
bin/spark-submit --deploy-mode cluster --class org.apache.spark.examples.SparkPi --master k8s://https://192.168.6.154:6443 --kubernetes-namespace default --conf spark.dynamicAllocation.enabled=true --conf spark.shuffle.service.enabled=true --conf spark.kubernetes.shuffle.namespace=default --conf spark.kubernetes.shuffle.labels=app=spark-shuffle-service,spark-version=2.1.0 --conf spark.app.name=spark-pi --conf spark.kubernetes.driver.docker.image=docker:5000/spark-driver:varun_latest31 --conf spark.kubernetes.executor.docker.image=docker:5000/spark-executor:varun_latest31 --conf spark.kubernetes.initcontainer.docker.image=docker:5000/spark-init-container:varun_latest31 local:///opt/spark/examples/jars/spark-examples_2.11-2.1.0-k8s-0.2.0-SNAPSHOT.jar 61000
The number of pods launched seem to be a much higher number than what is reported in the logs..
..
spark-pi-1498671730530-driver 1/1 Running 0 2m
spark-pi-1498671730530-exec-1 1/1 Running 0 2m
spark-pi-1498671730530-exec-10 1/1 Running 0 1m
spark-pi-1498671730530-exec-11 1/1 Running 0 1m
spark-pi-1498671730530-exec-12 1/1 Running 0 1m
spark-pi-1498671730530-exec-13 1/1 Running 0 1m
spark-pi-1498671730530-exec-14 1/1 Running 0 1m
spark-pi-1498671730530-exec-15 1/1 Running 0 1m
spark-pi-1498671730530-exec-16 1/1 Running 0 1m
spark-pi-1498671730530-exec-17 1/1 Running 0 59s
spark-pi-1498671730530-exec-18 1/1 Running 0 59s
spark-pi-1498671730530-exec-19 1/1 Running 0 59s
spark-pi-1498671730530-exec-2 1/1 Running 0 2m
spark-pi-1498671730530-exec-20 1/1 Running 0 59s
spark-pi-1498671730530-exec-21 1/1 Running 0 59s
spark-pi-1498671730530-exec-22 0/1 Pending 0 38s
spark-pi-1498671730530-exec-23 0/1 Pending 0 38s
spark-pi-1498671730530-exec-24 0/1 Pending 0 38s
spark-pi-1498671730530-exec-25 0/1 Pending 0 38s
spark-pi-1498671730530-exec-26 0/1 Pending 0 38s
spark-pi-1498671730530-exec-3 1/1 Running 0 2m
spark-pi-1498671730530-exec-4 1/1 Running 0 2m
spark-pi-1498671730530-exec-5 1/1 Running 0 2m
spark-pi-1498671730530-exec-6 1/1 Running 0 2m
spark-pi-1498671730530-exec-7 1/1 Running 0 1m
spark-pi-1498671730530-exec-8 1/1 Running 0 1m
spark-pi-1498671730530-exec-9 1/1 Running 0 1m
..
In the driver logs I see the following, I see only upto 5 launched
...
└─[0] <git:(master 29acb7e✱✈) > kubectl logs spark-pi-1498671730530-driver | grep -i backend
2017-06-28 17:42:30 INFO KubernetesClusterSchedulerBackend:54 - SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8
2017-06-28 17:42:35 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 1
2017-06-28 17:42:46 INFO KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint:54 - Registered executor NettyRpcEndpointRef(null) (10.38.0.3:51308) with ID 1
2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 2
2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 3
2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 4
2017-06-28 17:42:47 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 5
...
I tried using the published images from: https://apache-spark-on-k8s.github.io/userdocs/running-on-kubernetes.html
My logs say:
~/g/src k logs spark-pi-1498677689366-driver | grep 'Requesting a new executor, total executors'
2017-06-28 19:21:52 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 1
2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 2
2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 3
2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 4
2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 5
2017-06-28 19:22:02 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 6
2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 7
2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 8
2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 9
2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 10
2017-06-28 19:22:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 11
2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 12
2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 13
2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 14
2017-06-28 19:22:20 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 15
2017-06-28 19:22:21 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 16
2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 17
2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 18
2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 19
2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 20
2017-06-28 19:22:29 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 21
2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 22
2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 23
2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 24
2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 25
2017-06-28 19:22:37 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 26
2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 27
2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 28
2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 29
2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 30
2017-06-28 19:22:45 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 31
2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 32
2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 33
2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 34
2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 35
2017-06-28 19:22:54 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 36
2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 37
2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 38
2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 39
2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 40
2017-06-28 19:23:04 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 41
2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 42
2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 43
2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 44
2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 45
2017-06-28 19:23:13 INFO KubernetesClusterSchedulerBackend:54 - Requesting a new executor, total executors is now 46
Can you try with the published images? Are your images built from HEAD?
These are built from the HEAD...I will check once, if there is some user/operator error on my part and report..
@varunkatta can you try with static allocation? This might be a bug with our images in general - this wouldn't look specific to dynamic allocation, we build the pod's arguments identically.
And also use kubectl describe pod to get the details of the executor pods. I double checked our Dockerfiles on HEAD and they seem sane. Here's the command:
CMD SPARK_CLASSPATH="${SPARK_HOME}/jars/*" && \
if ! [ -z ${SPARK_MOUNTED_CLASSPATH}+x} ]; then SPARK_CLASSPATH="$SPARK_MOUNTED_CLASSPATH:$SPARK_CLASSPATH"; fi && \
if ! [ -z ${SPARK_EXECUTOR_EXTRA_CLASSPATH+x} ]; then SPARK_CLASSPATH="$SPARK_EXECUTOR_EXTRA_CLASSPATH:$SPARK_CLASSPATH"; fi && \
if ! [ -z ${SPARK_MOUNTED_FILES_DIR} ]; then cp -R "$SPARK_MOUNTED_FILES_DIR/." .; fi && \
exec /sbin/tini -- ${JAVA_HOME}/bin/java -Dspark.executor.port=$SPARK_EXECUTOR_PORT -Xms$SPARK_EXECUTOR_MEMORY -Xmx$SPARK_EXECUTOR_MEMORY -cp $SPARK_CLASSPATH org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url $SPARK_DRIVER_URL --executor-id $SPARK_EXECUTOR_ID --cores $SPARK_EXECUTOR_CORES --app-id $SPARK_APPLICATION_ID --hostname $SPARK_EXECUTOR_POD_IP
So we're probably seeing something where SPARK_EXECUTOR_MEMORY isn't being set properly, but looking at KubernetesClusterSchedulerBackend I can't quite see how this would happen. Would be good to inspect the pod to get more insight.
on describe pod of executor, this is what I see. I will try digging more information, if possible.
.....
Volume Mounts:
/tmp from tmp (rw)
/var/run/secrets/kubernetes.io/serviceaccount from default-token-2jnd3 (ro)
Environment Variables:
SPARK_EXECUTOR_PORT: 10000
SPARK_DRIVER_URL: spark://CoarseGrainedScheduler@10.45.128.3:46339
SPARK_EXECUTOR_CORES: 1
SPARK_EXECUTOR_MEMORY: 2g
SPARK_APPLICATION_ID: spark-5fefae5c99a04ca4b40249e078a653a1
SPARK_EXECUTOR_ID: 1
SPARK_MOUNTED_CLASSPATH: /var/spark-data/spark-jars/*
SPARK_EXECUTOR_POD_IP: (v1:status.podIP)
K8s versions:
Client Version: version.Info{Major:"1", Minor:"4", GitVersion:"v1.4.6+e569a27", GitCommit:"e569a27d02001e343cb68086bc06d47804f62af6", GitTreeState:"not a git tree", BuildDate:"2016-11-12T09:29:14Z", GoVersion:"go1.7.3", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"5", GitVersion:"v1.5.4", GitCommit:"7243c69eb523aa4377bce883e7c0dd76b84709a1", GitTreeState:"clean", BuildDate:"2017-03-07T23:34:32Z", GoVersion:"go1.7.4", Compiler:"gc", Platform:"linux/amd64"}
This issue is stale now.
| gharchive/issue | 2017-06-28T19:16:07 | 2025-04-01T04:33:26.242472 | {
"authors": [
"foxish",
"mccheah",
"varunkatta"
],
"repo": "apache-spark-on-k8s/spark",
"url": "https://github.com/apache-spark-on-k8s/spark/issues/360",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
263282421 | Delete Kubernetes resources when the client waits for and sees app completion
What changes were proposed in this pull request?
This PR fixes #519 for the case where the submission client waits for the submitted application to finish. Upon completion of the application, the submission client deletes all Kubernetes resources created for the application to run.
rerun integration tests please
I want to re-iterate on this issue/PR. If we have concern around losing some objects like the ConfigMap for setting up the init-container, as I said above, we could log information stored in it for debugging purpose. This, IMO, is better than making the ConfigMap stick around just for debugging. Thoughts?
@mccheah @foxish
Any thoughts on this ? It would be good to cleanup resources after completion. In a normal scenario, this is filling up a lot of services in completed state for example.
Thanks @felixcheung for jumping on this :)
hey where are we on this?
and how about going upstream?
@felixcheung Yes, I think we should go upstream. I created https://issues.apache.org/jira/browse/SPARK-23571.
Sorry, didn't see this before. Same comment as in https://github.com/apache/spark/pull/20722#discussion_r171968410. Why not do this during driver shutdown - that way, 1) we lose the driver and k8s garbage collection cleans up 2) driver terminates and we clean up executors as well as auxiliary resources like configmaps etc.
I want to re-iterate on this issue/PR. If we have concern around losing some objects like the ConfigMap for setting up the init-container, as I said above, we could log information stored in it for debugging purpose. This, IMO, is better than making the ConfigMap stick around just for debugging. Thoughts?
I agree. We can dump all k8s objects. My hunch is that it's not that useful, given it's a pretty deeply buried implementation detail.
As discussed in https://github.com/apache/spark/pull/20722, we think the right solution is move resource management into the driver pod. This way, cleanup of auxiliary resources upon completion is guaranteed regardless of which deployment mode is used and whether the client waits for application to complete or not.
| gharchive/pull-request | 2017-10-05T22:11:13 | 2025-04-01T04:33:26.248502 | {
"authors": [
"dharmeshkakadia",
"felixcheung",
"foxish",
"liyinan926"
],
"repo": "apache-spark-on-k8s/spark",
"url": "https://github.com/apache-spark-on-k8s/spark/pull/520",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
805347634 | feat(legacy-preset-chart-nvd3): subject NVD3 Bar chart sort by
Add control for sort by "asc" or "desc"
Associated with: https://github.com/apache/superset/pull/13049
AFTER
https://user-images.githubusercontent.com/8277264/107492589-50831880-6b95-11eb-8042-c94c3b21500e.mp4
Does this re-order bars as well? I'm worried there may be confusion if it doesn't.
Does this re-order bars as well? I'm worried there may be confusion if it doesn't.
@ktmud Actually the sort is working with a row limit, for example, you have a limit of 50, and sort descendent can be different values
Yes, I understand how it works, but users might expect it to sort the x-axis.
How about we change "Sort by" to "Sort rows by"?
cc @junlincc
I think we should default to no ordering, and only opt-in to sorting, both ascending and descending. Perhaps have a "sort bars" control and another "sort descending" one?
in this case, I agree we should leave it to "no default" since it ties to row limit.
We also discussed about having a 'switch' for ascending and descending. @mihir174
How about we change "Sort by" to "Sort rows by"?
I see your point @ktmud Introducing a new label to one specific chart might create more design debt.
let's add tooltip "Sort rows by. changing row limit may affect sorting resulting" something like that?
tag it for design-revisit
| gharchive/pull-request | 2021-02-10T09:44:33 | 2025-04-01T04:33:26.255455 | {
"authors": [
"junlincc",
"ktmud",
"maloun96"
],
"repo": "apache-superset/superset-ui",
"url": "https://github.com/apache-superset/superset-ui/pull/947",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
562213768 | ARTEMIS-2617 Improve AMQP Journal loading
I've opened this PR for discussion.
Currently it brings 4 improvements:
enable G1GC string deduplication
uses CoreMessageObjectPools on journal loading to save allocations
allows AMQP lazy message data scanning on journal loading
half-brute force search x-opt-delivery-time to avoid message data scanning
@tabish121 @gemmellr @clebertsuconic @michaelandrepearce
This PR contains many changes and I know there are parts that could be simplified a lot.
The most important bits that can really be a game-changer for AMQP on journal loading are on https://github.com/apache/activemq-artemis/pull/2975/commits/fc77a546ca1ee047218e938f0413c55ed8c838c6.
So please review me especially that part.
An important note:
I've found that AMQPMessagePersisterV2::decode does something weird re AMQPMessage::setAddress: AMQPMessagePersister::decode can AMQPMessage::setAddress, allocating AMQPMessage::extraProperties, but
AMQPMessagePersisterV2::decode can decode a new extraProperties totally overwriting the existing one. Is it a bug or it's suppose to work like that?
I see that this behaviour has been introduced on ARTEMIS-1858 in 1ae2784dc6075875b18780fa8ba40f86cb895f7b with this comment:
/**
* This will set the address on CoreMessage.
*
* Note for AMQPMessages:
* in AMQPMessages this will not really change the address on the message. Instead it will add a property
* on extraProperties which only transverse internally at the broker.
* Whatever you change here it won't affect anything towards the received message.
*
* If you wish to change AMQPMessages address you will have to do it directly at the AMQP Message, however beware
* that AMQPMessages are not supposed to be changed at the broker, so only do it if you know what you are doing.
* @param address
* @return
*/
Message setAddress(SimpleString address);
I've tried to mitigate this by using CoreMessageObjectPools to save SimpleString allocations here and by allowing TypedProperties::decode to append/replace any existing extraProperty, saving an HashMap allocation, but IMO it could be addressed by properly fixing it, wdyt? any idea how?
@michaelandrepearce I suppose you are interested in numbers: in my syntethic tests a 2 GB journal with very small messages has wasn't able to be loaded by a broker with 4GB heap size, always leading to a OOM: now it can load without relevant GC pauses eg
We're not quite yet at the same level of Core messages, but very near: I think that the duplicate address is responsible of the difference here.
The commit that allows to half-brute force search x-opt-delivery-time is what I'm not sure about:
that method should be optimized and performed optimally by proton itself IMO (nudge nudge @tabish121 @gemmellr), but I would be super happy to provide a PR for it
currently it can give rare false positive (causing message scanning when not necessary)
I've yet to verify how it behave with many message annotations and longer messages
If @michaelandrepearce has some real journal to test with and compare with master it would be great, given that this change can affect positively your usage of the broker ;)
In addition, I hope I haven't missed to force message data scanning on getter of other properties eg routingType.
@tabish121 I've tried to improve the fail-fast behaviour for both the search algorithm and the MessageAnnotations section search. The latter by using an IdentityHasMap to save checking the N possible message sections that would terminate the algorithm: let me know if that makes sense according to the AMQP spec :+1:
I need this merged, as this will impact the changes I'm making for large messages.
I see that you still need other work done, such as adding more tests.. on which case can you please add any further changes into a new PR?
thanks a lot.
| gharchive/pull-request | 2020-02-09T17:53:51 | 2025-04-01T04:33:26.264663 | {
"authors": [
"clebertsuconic",
"franz1981"
],
"repo": "apache/activemq-artemis",
"url": "https://github.com/apache/activemq-artemis/pull/2975",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
897389870 | ARTEMIS-3311 - ensure visibility of error state on operation context …
…callback registration, fix and test
did you run a whole test suite.. if you did it LGTM
yep. full tests look good.
| gharchive/pull-request | 2021-05-20T20:13:06 | 2025-04-01T04:33:26.266401 | {
"authors": [
"clebertsuconic",
"gtully"
],
"repo": "apache/activemq-artemis",
"url": "https://github.com/apache/activemq-artemis/pull/3592",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
216221166 | APEXCORE-676 Show description for DefaultProperties in get-app-packag…
…e-info command only when user requests it by providing --with-description flag
Output without --with-description flag
Command : get-app-package-info {apa-file}
"defaultProperties": { "dt.application..operator..attr.JVM_OPTIONS": "-Xmx128M", "dt.application..operator..attr.MEMORY_MB": "256" }
Output with --with-description flag
Command : get-app-package-info {apa-file} --with-description
"defaultProperties": {
"dt.application..operator..attr.JVM_OPTIONS":
{ "value": "-Xmx128M", "description": null }
,
"dt.application..operator..attr.MEMORY_MB":
{ "value": "256", "description": null }
}
Invalid argument
get-app-package-info {apa-file} --with-descripti
Output:
com.datatorrent.stram.cli.ApexCli$CliException: Invalid parameter --with-descripti
at com.datatorrent.stram.cli.ApexCli$GetAppPackageInfoCommand.execute(ApexCli.java:3482)
at com.datatorrent.stram.cli.ApexCli$3.run(ApexCli.java:1519)
@sandeshh @vrozov Kindly review
@vrozov Incorporated your comments
Jenkins build failed. Closing and reopening the request.
| gharchive/pull-request | 2017-03-22T21:58:48 | 2025-04-01T04:33:26.368227 | {
"authors": [
"ajaygit158"
],
"repo": "apache/apex-core",
"url": "https://github.com/apache/apex-core/pull/493",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1527746035 | feat: support disablestatus
Type of change:
[ ] Bugfix
[ ] New feature provided
[ ] Improve performance
[ ] Backport patches
What this PR does / why we need it:
Pre-submission checklist:
[ ] Did you explain what problem does this PR solve? Or what new features have been added?
[ ] Have you added corresponding test cases?
[ ] Have you modified the corresponding document?
[ ] Is this PR backward compatible? If it is not backward compatible, please discuss on the mailing list first
Codecov Report
Merging #1595 (46269c2) into master (486b46a) will increase coverage by 0.26%.
The diff coverage is 45.58%.
:exclamation: Current head 46269c2 differs from pull request most recent head cc876d7. Consider uploading reports for the commit cc876d7 to get more accurate results
@@ Coverage Diff @@
## master #1595 +/- ##
==========================================
+ Coverage 41.44% 41.70% +0.26%
==========================================
Files 87 87
Lines 7420 7457 +37
==========================================
+ Hits 3075 3110 +35
- Misses 3990 3992 +2
Partials 355 355
Impacted Files
Coverage Δ
pkg/config/config.go
63.54% <ø> (ø)
pkg/providers/ingress/ingress.go
4.05% <0.00%> (-0.04%)
:arrow_down:
pkg/providers/utils/ingress_status.go
31.34% <77.77%> (+31.34%)
:arrow_up:
pkg/providers/utils/status.go
30.43% <77.77%> (+30.43%)
:arrow_up:
cmd/ingress/ingress.go
82.90% <100.00%> (+0.14%)
:arrow_up:
pkg/providers/apisix/translation/apisix_route.go
31.21% <100.00%> (+0.81%)
:arrow_up:
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
| gharchive/pull-request | 2023-01-10T17:16:51 | 2025-04-01T04:33:26.436524 | {
"authors": [
"AlinsRan",
"codecov-commenter"
],
"repo": "apache/apisix-ingress-controller",
"url": "https://github.com/apache/apisix-ingress-controller/pull/1595",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2166265414 | fix: rat check error in vscode ide
Which issue does this PR close?
Closes #160 .
Rationale for this change
What changes are included in this PR?
How are these changes tested?
@viirya cc
Thanks @thexiay @snmvaughan
| gharchive/pull-request | 2024-03-04T08:36:44 | 2025-04-01T04:33:26.448357 | {
"authors": [
"thexiay",
"viirya"
],
"repo": "apache/arrow-datafusion-comet",
"url": "https://github.com/apache/arrow-datafusion-comet/pull/161",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1672964330 | Improve RoundRobin RepartitionExec
Describe the bug
RoundRobin repartitioning currently does not distribute the input tasks evenly over the output channels, causing the work to be not distributed evenly.
To Reproduce
When loading the data in memory in the TPC-H benchmark, this can be seen in the number of batches in MemoryExec (which uses RoundRobin partitioning).
MemoryExec: partitions=32, partition_sizes=[32, 32, 32, 32, 32, 32, 32, 32, 26, 26, 26, 25, 25, 25, 25, 25, 25, 25, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16], metrics=[]
It has a bias for the first output partitions/channels.
Expected behavior
Batches should be distributed more evenly over output channels.
Additional context
No response
Batches should be distributed more evenly over output channels.
Seems to be a load balancing issue. If you could count the number of batches already distributed to each channel and not completed then the classical The Power of Two Choices in Randomized Load Balancing algorithm could be evaluated.
@cristian-ilies-vasile yes, instead of round-robin repartitioning an improved scheme could be implemented based on number of buffered batches.
One good article describing this technique can be read here:
Deterministic Aperture: A distributed, load balancing algorithm
https://blog.twitter.com/engineering/en_us/topics/infrastructure/2019/daperture-load-balancer
| gharchive/issue | 2023-04-18T11:57:25 | 2025-04-01T04:33:26.452277 | {
"authors": [
"Dandandan",
"cristian-ilies-vasile"
],
"repo": "apache/arrow-datafusion",
"url": "https://github.com/apache/arrow-datafusion/issues/6043",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1988120175 | Fix: Do not try and preserve order when there is no order to preserve in RepartitionExec
Which issue does this PR close?
Closes https://github.com/apache/arrow-datafusion/issues/8043
Rationale for this change
We encountered problem is our downstream tests that RepartitionExec was trying to preserve an order even when there is no order to preserve.
This was because one codepath (RepartitionExec::new_with_children) it set the preserve_order flag to true when it creates a new RepartitionExec even if the new child had no order to preserve. During execution, this cause the code to try and merge a stream with no sort exprs, which causes an internal error
What changes are included in this PR?
Update RepartitionExec::with_preserve_order to avoid
Improve documentation
add test
Are these changes tested?
Yes, though I am not thrilled with the test (I will comment inline)
Are there any user-facing changes?
Is it possible to move this test to the same file with struct RepartitionExec as a unit test?
yes, that is a good idea -- I did so in ee55ec3d016574851aa2e77d0cf773990bc7aa32
I also see one call with_preserve_order(false) which is unnecessary and confusing in enforce_distribution.rs as well as one such call in replace_with_order_preserving_variants.rs. Let's remove them.
That is an excellent idea, I did so in 0c492e16fc7e0db5b17395559d878e8cd38caeba and I think the logic is much clearer now
Thanks @alamb for this PR. Having this check at initialization is great, and doesn't lead to unexpected behavior.
| gharchive/pull-request | 2023-11-10T18:14:22 | 2025-04-01T04:33:26.457761 | {
"authors": [
"alamb",
"mustafasrepo"
],
"repo": "apache/arrow-datafusion",
"url": "https://github.com/apache/arrow-datafusion/pull/8127",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
867664618 | Added support to length of Binary and List.
Note: migrated from original JIRA: https://issues.apache.org/jira/browse/ARROW-11420
None
I guess we could close this issue now.
| gharchive/issue | 2021-04-26T12:46:15 | 2025-04-01T04:33:26.459258 | {
"authors": [
"HaoYang670",
"alamb"
],
"repo": "apache/arrow-rs",
"url": "https://github.com/apache/arrow-rs/issues/172",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1519442146 | Release Arrow 30.0.1 (maintenance release for 30.0.0)
Is your feature request related to a problem or challenge? Please describe what you are trying to do.
We recently released arrow 30.0.0: https://github.com/apache/arrow-rs/issues/3336 and found two issues:
[ ] The upgrade to multiversion https://github.com/apache/arrow-rs/pull/3396 means now all crates that use arrow-rs require rebuild on every compilation (see https://github.com/calebzulawski/target-features/issues/1)
[x] https://github.com/apache/arrow-rs/issues/3454
Describe the solution you'd like
A new release version 30.0.1 with just the fixes for:
[ ] https://github.com/apache/arrow-rs/pull/3452
[ ] https://github.com/apache/arrow-rs/pull/3451
Describe alternatives you've considered
Possibly release 30.1.0 (need dependency)?
Just wait until the release of 31.0.0 #3415
Additional context
Found while trying to update datafusion. This may block the upgrade and possibly delay the release https://github.com/apache/arrow-datafusion/issues/4776 cc @andygrove
Release thread - https://lists.apache.org/thread/p5dqfopymxw0or7hqh1tl5npcj4338qw
Shall this be closed since the release has already taken place?
| gharchive/issue | 2023-01-04T19:02:34 | 2025-04-01T04:33:26.465677 | {
"authors": [
"alamb",
"iajoiner",
"tustvold"
],
"repo": "apache/arrow-rs",
"url": "https://github.com/apache/arrow-rs/issues/3455",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1688007803 | object_store: delete multiple
Problem
I want to delete multiple objects atomically.
Solution
Add a new method to ObjectStore. It would follow the API similar to S3 DeleteObjects
Alternatives
Multiple calls to delete aren't atomic.
Discussion
Not all implementations of ObjectStore support this. We could have an extension trait, ObjectStoreAtomic: ObjectStore.
Would you consider accepting a PR that added this? Any thoughts on implementation?
I believe this is a duplicate of #2615
FWIW S3 DeleteObjects is not atomic, it can partially fail - https://docs.aws.amazon.com/AmazonS3/latest/API/API_DeleteObjects.html
Agreed. And thanks, good to know that about S3.
| gharchive/issue | 2023-04-28T06:56:24 | 2025-04-01T04:33:26.468979 | {
"authors": [
"Kinrany",
"tustvold"
],
"repo": "apache/arrow-rs",
"url": "https://github.com/apache/arrow-rs/issues/4154",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1040107683 | Fix 876, allow null array to be cast from and to other types
Which issue does this PR close?
Closes #876
Depends on #591
Rationale for this change
What changes are included in this PR?
Are there any user-facing changes?
Looks like the tests are failing with dictionaries:
Test casting Dictionary(Int8, Int32) --> Null
thread 'compute::kernels::cast::tests::test_can_cast_types' panicked at 'not implemented: Take not supported for data type Null', arrow/src/compute/kernels/take.rs:283:14
failures:
compute::kernels::cast::tests::test_can_cast_types
Codecov Report
Merging #884 (5143f63) into master (898924f) will increase coverage by 0.02%.
The diff coverage is 100.00%.
@@ Coverage Diff @@
## master #884 +/- ##
==========================================
+ Coverage 82.45% 82.47% +0.02%
==========================================
Files 168 168
Lines 48231 48232 +1
==========================================
+ Hits 39767 39778 +11
+ Misses 8464 8454 -10
Impacted Files
Coverage Δ
arrow/src/compute/kernels/cast.rs
94.63% <100.00%> (+<0.01%)
:arrow_up:
arrow/src/datatypes/datatype.rs
65.36% <0.00%> (-0.44%)
:arrow_down:
parquet/src/arrow/array_reader.rs
77.72% <0.00%> (-0.09%)
:arrow_down:
arrow/src/ipc/writer.rs
84.28% <0.00%> (+0.18%)
:arrow_up:
arrow/src/array/data.rs
74.93% <0.00%> (+0.81%)
:arrow_up:
arrow/src/array/array.rs
85.71% <0.00%> (+2.33%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 898924f...5143f63. Read the comment docs.
| gharchive/pull-request | 2021-10-30T10:14:04 | 2025-04-01T04:33:26.482838 | {
"authors": [
"Jimexist",
"alamb",
"codecov-commenter"
],
"repo": "apache/arrow-rs",
"url": "https://github.com/apache/arrow-rs/pull/884",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1614833496 | [Python] Make ChunkedArray & Array iterable
Describe the enhancement requested
This would allow for more Pythonic patterns in downstream libraries.
>>> next(pa.chunked_array([[1,2,3]]))
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
TypeError: 'pyarrow.lib.ChunkedArray' object is not an iterator
Component(s)
Python
Can you give an example of in what kind of pattern you would want to use this?
And for a ChunkedArray, do you expect that it gives a chunk or a scalar?
| gharchive/issue | 2023-03-08T08:13:42 | 2025-04-01T04:33:26.486503 | {
"authors": [
"Fokko",
"jorisvandenbossche"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/issues/34495",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1767775375 | [Python] Python to array conversion sometimes ignores "safe=False" or is using unsafe instead of safe cast
Describe the bug, including details regarding any error messages, version, and platform.
There are two cases where the array constructor taking the python conversion path (python_to_arrow.cc) doesn't handle the safe keyword properly.
The safe keyword is set to True by default and is ignored if passed
One example where python list is taking python_to_arrow.cc code path are Decimals (which are python objects). Here the conversion from Decimal to int does not overflow by default and one can’t turn it off with safe=False:
>>> import pyarrow as pa
>>> pa.array([Decimal('1234')]).cast(pa.int8(), safe=False)
<pyarrow.lib.Int8Array object at 0x7efbbd18e4b0>
[
-46
]
>>> pa.array([Decimal('1234')], pa.int8(), safe=False)
Traceback (most recent call last):
...
ArrowInvalid: Value 1234 too large to fit in C integer type
Another example is JSON data with nested data (list type). The conversion is taking python_to_arrow.cc code path where safe keyword is also ignored. See example in https://github.com/apache/arrow/issues/31402.
In some cases the safe keyword is ignored and does unsafe conversions
Nested case
>>> import pyarrow as pa
>>> pa.array(np.array([[1.5], [2.5, 3.5]], dtype=object), type=pa.list_(pa.int64()), safe=True)
<pyarrow.lib.ListArray object at 0x7f004fc74700>
[
[
1
],
[
2,
3
]
]
Primitive array case. Numpy array and python list take a different code path (numpy_to_array.cc vs python_to_arrow.cc):
>>> import pyarrow as pa
>>> pa.array(np.array([1.5, 2.5]), type=pa.int64(), safe=True)
...
ArrowInvalid: Float value 1.5 was truncated converting to int64
vs
>>> import pyarrow as pa
>>> pa.array([1.5, 2.5], type=pa.int64(), safe=True)
<pyarrow.lib.Int64Array object at 0x7f004fc72c40>
[
1,
2
]
Another example of wrong handling of safe keyword is when using nested data in pandas (an object thus taking the python_to_arrow.cc code path):
>>> import pandas as pd
>>> import pyarrow as pa
>>> int_dataframe = pd.DataFrame({"array": [[1, 2]]})
>>> float_dataframe = pd.DataFrame({"array": [[1.5, 2.3]]})
>>> int_table = pa.Table.from_pandas(int_dataframe)
>>> table = pa.Table.from_pandas(float_dataframe, schema=int_table.schema)
>>> table
pyarrow.Table
array: list<item: int64>
child 0, item: int64
----
array: [[[1,2]]]
cc @jorisvandenbossche @dane
Component(s)
Python
@AlenkaF I may not be the Dane you're looking for, but I can confirm this is Python. Best of luck on addressing the bug!
Oh sorry about that! :)
| gharchive/issue | 2023-06-21T14:44:15 | 2025-04-01T04:33:26.492985 | {
"authors": [
"AlenkaF"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/issues/36202",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2368041912 | [C++][Parquet] Add a RowGroup reader interface class
Currently the logic for interacting with row group metadata and constructing column decoders is embedded in the parquet_reader.cc executable here:
https://github.com/apache/parquet-cpp/blob/master/example/parquet_reader.cc
With PARQUET-434, we have a file reader container, which can then provide a row group reader container, something like
RowGroupReader* group_reader = file_reader->row_group(i);
Reporter: Wes McKinney / @wesm
Assignee: Wes McKinney](https://issues.apache.org/jira/secure/ViewProfile.jspa?name=wesm) / @wesm
Note: This issue was originally created as PARQUET-451. Please see the migration documentation for further details.
Wes McKinney / @wesm:
This is done in https://github.com/apache/parquet-cpp/pull/23
The API currently returns std::shared_ptr instances owned by the parent file reader. Feedback on this API is requested.
std::shared_ptr<RowGroupReader> reader = file_reader->RowGroup(i);
Nong Li / @nongli:
Issue resolved by pull request 23
https://github.com/apache/parquet-cpp/pull/23
| gharchive/issue | 2016-01-22T19:43:20 | 2025-04-01T04:33:26.498877 | {
"authors": [
"asfimport"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/issues/42307",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1119989343 | ARROW-3016: [Docs][C++] Memory profiling with perf
Put together some scripts and an example for using perf to understand memory allocations in an Arrow C++ program.
Thanks, this is pretty slick. Just two nits.
Well the commands you provided earlier help me get unstuck on learning how to use perf probes, so thank you @lidavidm! 🙏
| gharchive/pull-request | 2022-01-31T22:12:46 | 2025-04-01T04:33:26.500721 | {
"authors": [
"wjones127"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/12307",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
363411168 | [GLib] Import Parquet GLib
It's based on https://github.com/red-data-tools/parquet-glib at
32a2b2099f90c19d28322693dd853dc8643c55ec.
See also: https://github.com/red-data-tools/parquet-glib/commit/32a2b2099f90c19d28322693dd853dc8643c55ec
I started discussion at dev@parquet.apache.org: https://lists.apache.org/thread.html/72b2bf04d8035d854a44487d956322cbe6c8629488128bc326613fb1@<dev.parquet.apache.org>
Codecov Report
Merging #2622 into master will decrease coverage by 0.01%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #2622 +/- ##
==========================================
- Coverage 87.17% 87.15% -0.02%
==========================================
Files 380 380
Lines 59037 59037
==========================================
- Hits 51466 51456 -10
- Misses 7501 7507 +6
- Partials 70 74 +4
Impacted Files
Coverage Δ
go/arrow/math/int64_avx2_amd64.go
0% <0%> (-100%)
:arrow_down:
go/arrow/memory/memory_avx2_amd64.go
0% <0%> (-100%)
:arrow_down:
go/arrow/math/float64_avx2_amd64.go
0% <0%> (-100%)
:arrow_down:
go/arrow/math/uint64_avx2_amd64.go
0% <0%> (-100%)
:arrow_down:
go/arrow/memory/memory_amd64.go
28.57% <0%> (-14.29%)
:arrow_down:
go/arrow/math/math_amd64.go
31.57% <0%> (-5.27%)
:arrow_down:
go/arrow/math/float64_amd64.go
33.33% <0%> (ø)
:arrow_up:
go/arrow/math/int64_amd64.go
33.33% <0%> (ø)
:arrow_up:
go/arrow/math/uint64_amd64.go
33.33% <0%> (ø)
:arrow_up:
go/arrow/math/float64_sse4_amd64.go
100% <0%> (+100%)
:arrow_up:
... and 3 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update a20d281...43ae5c5. Read the comment docs.
Thanks @kou -- we should go through the IP clearance after having a vote on dev@parquet.a.o. Your ICLA should be sufficient for this (I see there is one other contributor with a 2 line change)?
we should go through the IP clearance after having a vote on dev@parquet.a.o.
OK.
Your ICLA should be sufficient for this (I see there is one other contributor with a 2 line change)?
Yes.
The other contributor has one commit: https://github.com/red-data-tools/parquet-glib/commit/8091c812aedd4c94af98d1e664b8bb424c36b992
But the change isn't included in this pull request because we don't need the change by merging this project to the Arrow repository: https://github.com/apache/arrow/pull/2622/files#diff-88333024fdffef0c691b5aa164aa1a37R110
Great work @kou . I looked through the code and everything looks fine. We should do the formal vote and than we can merge this.
Thanks for reviewing this.
I'll create a PARQUET JIRA ticket when the vote is passed.
@wesm Thanks for creating a JIRA ticket.
Is it OK to use ARROW ticket instead of PARQUET ticket for Parquet GLib bindings?
If it's OK, I'll also use ARROW ticket for new features.
OK. I'll also use ARROW project.
| gharchive/pull-request | 2018-09-25T05:03:14 | 2025-04-01T04:33:26.519873 | {
"authors": [
"codecov-io",
"kou",
"wesm",
"xhochy"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/2622",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
386721068 | ARROW-3916: [Python] Add support for filesystem kwarg in ParquetWriter
Implements ARROW 3916.
I'll resolve the flake8 failures this evening. Anyone know if I should be worried about the failing Java micro benchmark?
Codecov Report
Merging #3070 into master will increase coverage by 1.11%.
The diff coverage is 95%.
@@ Coverage Diff @@
## master #3070 +/- ##
==========================================
+ Coverage 87.06% 88.18% +1.11%
==========================================
Files 489 431 -58
Lines 68974 65330 -3644
==========================================
- Hits 60055 57610 -2445
+ Misses 8818 7720 -1098
+ Partials 101 0 -101
Impacted Files
Coverage Δ
python/pyarrow/parquet.py
93.89% <100%> (ø)
:arrow_up:
python/pyarrow/filesystem.py
52.91% <100%> (+2.08%)
:arrow_up:
python/pyarrow/tests/test_parquet.py
97.29% <92.85%> (-0.11%)
:arrow_down:
cpp/src/plasma/fling.cc
67.18% <0%> (-17.26%)
:arrow_down:
cpp/src/arrow/util/thread-pool-test.cc
98.91% <0%> (-0.55%)
:arrow_down:
cpp/src/plasma/thirdparty/dlmalloc.c
47.16% <0%> (-0.38%)
:arrow_down:
cpp/src/arrow/util/parsing.h
95.6% <0%> (ø)
:arrow_up:
go/arrow/array/table.go
go/arrow/math/uint64_amd64.go
go/arrow/internal/testing/tools/bool.go
... and 62 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 98bdde8...9f70cfc. Read the comment docs.
Are you able to rebase this?
| gharchive/pull-request | 2018-12-03T09:10:08 | 2025-04-01T04:33:26.533909 | {
"authors": [
"codecov-io",
"macks22",
"wesm"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/3070",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1897179622 | GH-34105: [R] Provide extra output for failed builds
Rationale for this change
This is a replacement for the previous PR https://github.com/apache/arrow/pull/37698. The rationale for this PR is providing extra output for R package builds where the C++ build fails
What changes are included in this PR?
Update the system call to save output when building Arrow C++ from the R package and output it if it's failed
Are these changes tested?
No automated tests but the changes have been tested manually.
Are there any user-facing changes?
Yes, but only for users building the R package from source which is hopefully not common.
Closes: #34105
I've updated the PR to address points brought up in https://github.com/apache/arrow/pull/37727#discussion_r1326477485 so this is ready for another look.
| gharchive/pull-request | 2023-09-14T19:30:40 | 2025-04-01T04:33:26.537223 | {
"authors": [
"amoeba"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/37727",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2291901373 | GH-41608: [C++][Python] Extends the add_key_value to parquet::arrow and PyArrow
Rationale for this change
The previous pr ( https://github.com/apache/arrow/pull/34889 ) add a AddKeyValueMetadata to FileWriter. And now we should export it to Parquet Arrow and Python API.
What changes are included in this PR?
Add AddKeyValueMetadata in parquet::arrow
Add add_key_value_metadata in pyarrow
testing
Are these changes tested?
Yes
Are there any user-facing changes?
New api allowing add key-value metadata to Parquet file
GitHub Issue: #41608
I've no idea why R language CI failed
@AlenkaF @jorisvandenbossche @pitrou Would you mind take a look?
Migrate to test_parquet_writer.py now. @pitrou would you mind revisit C++ part or merge this?
@pitrou comment fixed
@mapleFU Feel free to merge if CI is fine.
CI failed is unrelated, merge
| gharchive/pull-request | 2024-05-13T06:04:57 | 2025-04-01T04:33:26.542632 | {
"authors": [
"mapleFU",
"pitrou"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/41633",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
445576055 | ARROW-5360: [Rust] Update rustyline to fix build
This is placeholder to check if the pending update to rustyline will fix the build issue. I'll update with the correct crate once it's published
@wesm @sunchao PTAL, I'll wait to see if the build passes before I go to bed.
| gharchive/pull-request | 2019-05-17T19:00:34 | 2025-04-01T04:33:26.543685 | {
"authors": [
"nevi-me"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/4337",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
587985757 | ARROW-8220: [Python] Make dataset FileFormat objects serializable
Also did some refactoring for a more pleasant user API.
Also did some refactoring for a more pleasant user API.
I also don't like the ParquetFileFormatReaderOptions very much as user API, but, I am not sure we can just pass them all to ParquetFileFormat, since we are going to use that for both reading and writing, and mixing keywords for those all in a single constructor is going to get confusing.
I think we should rather give a better API in a parquet / reading specific API like parquet.read_table or ParquetDataset.
@jorisvandenbossche agree. May we defer your suggestion to a follow-up?
Well, my comment is kind of: we need to keep ParquetFileFormatReaderOptions, so since you are removing that, I would rather not defer that to a follow-up (but you don't need to agree with keeping it, of course :-))
ParquetFileFormatReaderOptions was still bound to the ParquetFormat, your proposal is more about making the reader and writer options independent from the ParquetFormat. So this PR doesn't change that dependency.
I can wire these options, but it's not entirely clear because we don't have a read() method on the datasets. Once we add support for writing we can refine the API.
ParquetFileFormatReaderOptions was still bound to the ParquetFormat, your proposal is more about making the reader and writer options independent from the ParquetFormat. So this PR doesn't change that dependency.
Yes, it is still bound to the format, but it splits its keywords in two groups:
format = ParquetFileFormat(reader_options=dict(...), writer_options=dict(...))
it's not entirely clear because we don't have a read() method on the datasets
I think to_table is the "read" method?
Once we add support for writing we can refine the API.
Yeah, I fully agree much of this discussion is a bit "up in the air", since we don't yet have writing, so don't yet know how we would want to make the API for writing.
But it's for that reason that I commented to keep it as is, as there is also no clear reason yet for changing IMO, since we don't know the final API with writing (but it was an explicit decision, at least on the C++ side, to have this a separate set of options instead of direct ParquetFileFormat options). But OK, since it is easy to put it back later, I won't block removing it if you prefer that :)
@jorisvandenbossche updated as you requested
Build failure is unrelated.
| gharchive/pull-request | 2020-03-25T20:56:05 | 2025-04-01T04:33:26.549545 | {
"authors": [
"jorisvandenbossche",
"kszucs"
],
"repo": "apache/arrow",
"url": "https://github.com/apache/arrow/pull/6720",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
202608973 | [BEAM-1292] PubSubIO: fix and improve testing for DisplayData
Also adds better type/nullability checking in the code.
R: @tgroh
Coverage decreased (-0.04%) to 69.743% when pulling fcee4e4d434d18e719a5c729a2712cbf57f423d4 on dhalperi:pubsub-display-data into 9db5f746ae81c4fa755a9a8f2c4888759c8e7042 on apache:master.
| gharchive/pull-request | 2017-01-23T18:09:46 | 2025-04-01T04:33:26.557067 | {
"authors": [
"coveralls",
"dhalperi"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/1817",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1260178897 | Add mockito configuration to support static mocking without mockito-i…
…nline
Following the tips here: https://stackoverflow.com/questions/65986197/difference-between-mockito-core-vs-mockito-inline
Removing extra dep added here: https://github.com/apache/beam/pull/17753/
Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily:
[ ] Choose reviewer(s) and mention them in a comment (R: @username).
[ ] Add a link to the appropriate issue in your description, if applicable. This will automatically link the pull request to the issue.
[ ] Update CHANGES.md with noteworthy changes.
[ ] If this contribution is large, please file an Apache Individual Contributor License Agreement.
See the Contributor Guide for more tips on how to make review process smoother.
To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md
GitHub Actions Tests Status (on master branch)
See CI.md for more information about GitHub Actions CI.
hm this seems to definitely have broken something, hah.
| gharchive/pull-request | 2022-06-03T17:38:03 | 2025-04-01T04:33:26.564412 | {
"authors": [
"pabloem"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/18385",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
207300666 | [BEAM-1471]: Add cross-sdk tests for IterableCoder
Be sure to do all of the following to help us incorporate your contribution
quickly and easily:
[ ] Make sure the PR title is formatted like:
[BEAM-<Jira issue #>] Description of pull request
[ ] Make sure tests pass via mvn clean verify. (Even better, enable
Travis-CI on your fork and ensure the whole test matrix passes).
[ ] Replace <Jira issue #> in the title with the actual Jira issue
number, if there is one.
[ ] If this contribution is large, please file an Apache
Individual Contributor License Agreement.
R: @dhalperi @robertwb
Coverage decreased (-0.003%) to 69.713% when pulling 965160c0c0bada4500d51ca44c6c8e5e1c0476bf on vikkyrk:common_iterable_coder into 30cb93ced35494db79c81c49e07a298f560c3905 on apache:master.
retest this please
Changes Unknown when pulling cab5e6347f3fcece2cac4819e268488d7ce66830 on vikkyrk:common_iterable_coder into ** on apache:master**.
| gharchive/pull-request | 2017-02-13T18:28:38 | 2025-04-01T04:33:26.569698 | {
"authors": [
"coveralls",
"dhalperi",
"vikkyrk"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/1996",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
214208676 | [BEAM-1721] Do not shift Timestamps forwards in Reshuffle
Be sure to do all of the following to help us incorporate your contribution
quickly and easily:
[x] Make sure the PR title is formatted like:
[BEAM-<Jira issue #>] Description of pull request
[x] Make sure tests pass via mvn clean verify. (Even better, enable
Travis-CI on your fork and ensure the whole test matrix passes).
[x] Replace <Jira issue #> in the title with the actual Jira issue
number, if there is one.
[x] If this contribution is large, please file an Apache
Individual Contributor License Agreement.
Timestamps can be shifted forwards after the fact, but cannot generally
be shifted backwards. Because reshuffle outputs "as quickly as
possible", only elements that arrive approximately simulatenously with
each other will have their timestamps shifted.
There is currently no way to output all input elements with their
original timestamps without explicitly reifying those timestamps and
reassigning them on the output elements.
Coverage decreased (-0.003%) to 70.177% when pulling 347250b80fafecbdb12059233255e526370a1623 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master.
R: @jkff
Coverage decreased (-0.008%) to 70.173% when pulling da185f577c8193ea2dd9071c183b385fe6f83560 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master.
Coverage decreased (-0.01%) to 70.17% when pulling 2a7a4adc36c058f966ba235ae2da01e0796e96c7 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master.
Coverage decreased (-0.001%) to 70.179% when pulling 2a7a4adc36c058f966ba235ae2da01e0796e96c7 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master.
Coverage increased (+0.002%) to 70.183% when pulling 2a7a4adc36c058f966ba235ae2da01e0796e96c7 on tgroh:reshuffle_output_time_fn into cc12fd37834ce2b0f85cce79dae8821f7c3c1607 on apache:master.
Started to explicitly reify timestamps in Reshuffle and restore them after the fact.
CC: @dpmills
retest this please
retest this please
Coverage decreased (-0.05%) to 70.137% when pulling 4763f74d325a5c22e82236a8225ec5248ba2a5cb on tgroh:reshuffle_output_time_fn into 49c29f1faebddebedb151013cc18c4c4f2237644 on apache:master.
retest this please
Coverage decreased (-0.04%) to 70.156% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master.
Coverage decreased (-0.04%) to 70.156% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master.
Coverage decreased (-0.03%) to 70.159% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master.
retest this please
Coverage decreased (-0.04%) to 70.156% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master.
retest this please
Coverage decreased (-0.03%) to 70.159% when pulling e0eef51695e8bb749c8d83426b7af16e2e916435 on tgroh:reshuffle_output_time_fn into a0b8a854df6129e2dbbbd4bad683e53d975f3146 on apache:master.
| gharchive/pull-request | 2017-03-14T21:14:26 | 2025-04-01T04:33:26.587758 | {
"authors": [
"coveralls",
"jkff",
"tgroh"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/2246",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1412365582 | GA Migration Java Precommit IO AWS, AWS2, Azure, Cassandra, CDAP ( Java Precommit Sharding)
This PR is part of the migration from jenkins to github actions.
This PR is part of the Precommit Java Sharding
This PR contains this tests
Precommit Java IO AWS
Precommit Java IO AWS2
Precommit Java IO Azure
Precommit Java IO Cassandra
Precommit Java IO CDAP
4 Files were added to the workflows folder of github actions
job-precommit-java-io-aws.yml
job-precommit-java-io-aws2.yml
job-precommit-java-io-azure.yml
job-precommit-java-io-cassandra.yml
job-precommit-java-io-cdap.yml
CI.md was modified adding the new GA workflows
Waiting for this PR before merge https://github.com/apache/beam/pull/22703
Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily:
[ ] Choose reviewer(s) and mention them in a comment (R: @username).
[ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead.
[ ] Update CHANGES.md with noteworthy changes.
[ ] If this contribution is large, please file an Apache Individual Contributor License Agreement.
See the Contributor Guide for more tips on how to make review process smoother.
To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md
GitHub Actions Tests Status (on master branch)
See CI.md for more information about GitHub Actions CI.
R: @damccorm
This PR was missing in the document with all the other PR's I'll add it ( sorry for that).
It's also part of the Java Precommit Sharding tests.
| gharchive/pull-request | 2022-10-17T23:29:34 | 2025-04-01T04:33:26.597782 | {
"authors": [
"andoni-guzman"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/23683",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1433143103 | [Task]: PR Bot will push commits only if they are non-empty
fixes #23851
Currently a number of non-empty commits are pushed for the PR bot.
This pr provides for modifying the bot such that it commits & pushes only if there are any changes made.
Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily:
[ ] Choose reviewer(s) and mention them in a comment (R: @username).
[ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead.
[ ] Update CHANGES.md with noteworthy changes.
[ ] If this contribution is large, please file an Apache Individual Contributor License Agreement.
See the Contributor Guide for more tips on how to make review process smoother.
To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md
GitHub Actions Tests Status (on master branch)
See CI.md for more information about GitHub Actions CI.
R: @pabloem
This looks great! Thanks for sticking with this one!
Thanks for helping me so much . I also learnt about exitcodes in git and how they can be used with OR operator by using your idea. 🙂
| gharchive/pull-request | 2022-11-02T13:24:16 | 2025-04-01T04:33:26.606169 | {
"authors": [
"ayushthe1"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/23937",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1846665445 | [Website] add octo case-study
add octo case-study with related files
before:
after:
Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily:
[ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead.
[ ] Update CHANGES.md with noteworthy changes.
[ ] If this contribution is large, please file an Apache Individual Contributor License Agreement.
See the Contributor Guide for more tips on how to make review process smoother.
To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md
GitHub Actions Tests Status (on master branch)
See CI.md for more information about GitHub Actions CI or the workflows README to see a list of phrases to trigger workflows.
R: @aaltay, could you kindly, review current PR?
| gharchive/pull-request | 2023-08-11T11:30:50 | 2025-04-01T04:33:26.613966 | {
"authors": [
"bullet03"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/27967",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1953289679 | [YAML] Improved pipeline schema definition.
Slightly stricter definitions for catching more errors, as well as avoding the use of anyOf which often makes it difficult to deduce what the true error is.
This does mean a pipeline must have a transform (or source/sink) block rather than simply be itself a list of transforms.
Thank you for your contribution! Follow this checklist to help us incorporate your contribution quickly and easily:
[ ] Mention the appropriate issue in your description (for example: addresses #123), if applicable. This will automatically add a link to the pull request in the issue. If you would like the issue to automatically close on merging the pull request, comment fixes #<ISSUE NUMBER> instead.
[ ] Update CHANGES.md with noteworthy changes.
[ ] If this contribution is large, please file an Apache Individual Contributor License Agreement.
See the Contributor Guide for more tips on how to make review process smoother.
To check the build health, please visit https://github.com/apache/beam/blob/master/.test-infra/BUILD_STATUS.md
GitHub Actions Tests Status (on master branch)
See CI.md for more information about GitHub Actions CI or the workflows README to see a list of phrases to trigger workflows.
R: @damccorm
| gharchive/pull-request | 2023-10-20T00:54:00 | 2025-04-01T04:33:26.621287 | {
"authors": [
"robertwb"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/29083",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
228432462 | [BEAM-1345] Remove FileSystems.setDefaultConfigInWorkers since Dataflow no longer depends on this
Be sure to do all of the following to help us incorporate your contribution
quickly and easily:
[x] Make sure the PR title is formatted like:
[BEAM-<Jira issue #>] Description of pull request
[ ] Make sure tests pass via mvn clean verify.
[x] Replace <Jira issue #> in the title with the actual Jira issue
number, if there is one.
[x] If this contribution is large, please file an Apache
Individual Contributor License Agreement.
R: @dhalperi
Dataflow worker for master still needs to be updated, this is to be held till that occurs
LGTM, once Dataflow worker for master is updated and examples precommit is green.
Coverage decreased (-0.002%) to 70.669% when pulling e6650d9d4a5fb68bc45acc9286bec52c44a83a4f on lukecwik:cleanup into 49245080ad6393b5678077caa047f7a6ef0efffc on apache:master.
retest this please
Coverage increased (+0.005%) to 70.676% when pulling e6650d9d4a5fb68bc45acc9286bec52c44a83a4f on lukecwik:cleanup into 49245080ad6393b5678077caa047f7a6ef0efffc on apache:master.
| gharchive/pull-request | 2017-05-12T23:46:53 | 2025-04-01T04:33:26.626852 | {
"authors": [
"coveralls",
"dhalperi",
"lukecwik"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/3134",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2489574014 | Support ordered list states in python sdk and fnapi runner
This is a follow-up PR of #30317, where we support ordered list state in java sdk.
Here we implement ordered list state for python sdk. We also add the support of ordered list state into Python FnApi runner for debug and testing purpose.
R: @robertwb @acrites
@robertwb Could you please take another look? Thanks!
| gharchive/pull-request | 2024-08-27T14:36:46 | 2025-04-01T04:33:26.628308 | {
"authors": [
"shunping"
],
"repo": "apache/beam",
"url": "https://github.com/apache/beam/pull/32326",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
800358958 | Build failure when artifact cache is incorrectly set up
See original issue on GitLab
In GitLab by [Gitlab user @raoul].hidalgocharman on Nov 14, 2018, 15:46
Summary
If you specify an artifact cache incorrectly, you eventually get something along these lines
[00:00:08][f0c3789e][build:hello.bst] FAILURE push_directory was called, but no remote artifact servers are configured as push remotes.
Either more useful information or a warning that continues despite not successfully pushing to a directory would be preferable.
Steps to reproduce
Add an artifact cache in project.conf that isn't actually set up, and a remote-execution service (but it doesn't need to be configured correctly).
For example, provided you don't have an artifact server running locally on 50052:
artifacts:
- url: https://localhost:50052
push: true
remote-execution:
url: http://remote-execution-server:50051
Then try and build an artifact.
What is the current bug behavior?
It warns at the beginning that it failed to initialise the remote:
[--:--:--][][] WARNING Failed to initialize remote http://localhost:50052: Connect Failed
Then when trying to push an artifact after building it, the entire build fails.
What is the expected correct behavior?
Highlighting that it failed to initialise the remote in the build failure would be more useful , but it doesn't seem appropriate that an incorrectly configure artifact cache (or one that is currently down) should cause an entire build to fail.
When remote execution isn't specified the build will still continue, and it should also do this when remote execution is specified.
Relevant logs and/or screenshots
Using the following project.conf for the autotools example project, with artifacts and remote-execution specified:
# Unique project name
name: autotools
# Required BuildStream format version
format-version: 18
# Subdirectory where elements are stored
element-path: elements
# Define some aliases for the tarballs we download
aliases:
alpine: https://gnome7.codethink.co.uk/tarballs/
gnu: http://ftpmirror.gnu.org/gnu/automake/
artifacts:
- url: http://localhost:50052
push: true
remote-execution:
url: http://remote-execution-server:50051
When trying to build you get the following failure message on hello.bst:
[00:00:06][f0c3789e][build:hello.bst ] FAILURE Running configure-commands
[--:--:--][f0c3789e][build:hello.bst ] START Caching artifact
[00:00:00][f0c3789e][build:hello.bst ] SUCCESS Caching artifact
[00:00:09][f0c3789e][build:hello.bst ] FAILURE push_directory was called, but no remote artifact servers are configured as push remotes.
Printing the last 20 lines from log file:
/home/raoulhidalgo/.cache/buildstream/logs/autotools/hello/f0c3789e-build.26301.log
======================================================================
[--:--:--] START hello.bst: Integrating sandbox
[00:00:00] SUCCESS hello.bst: Integrating sandbox
[--:--:--] START hello.bst: Staging sources
[00:00:00] SUCCESS hello.bst: Staging sources
[--:--:--] START hello.bst: Running configure-commands
[--:--:--] STATUS hello.bst: Running configure-commands
export NOCONFIGURE=1;
if [ -x ./configure ]; then true;
elif [ -x ./autogen ]; then ./autogen;
elif [ -x ./autogen.sh ]; then ./autogen.sh;
elif [ -x ./bootstrap ]; then ./bootstrap;
elif [ -x ./bootstrap.sh ]; then ./bootstrap.sh;
else autoreconf -ivf .;
fi
[00:00:06] FAILURE hello.bst: Running configure-commands
[--:--:--] START hello.bst: Caching artifact
[00:00:00] SUCCESS hello.bst: Caching artifact
[00:00:09] FAILURE hello.bst: push_directory was called, but no remote artifact servers are configured as push remotes.
======================================================================
Build failure on element: hello.bst
Possible fixes
The exception is raised in buildstream._artifactcache.artifact_cache.py:735.
Maybe make sure it doesn't attempt to use push_directory when no artifact caches get successfully set up, or make push_directory silently pass.
This needs to be re-verified, I'm quite sure this is not the case anymore.
| gharchive/issue | 2021-02-03T14:03:05 | 2025-04-01T04:33:26.637109 | {
"authors": [
"BuildStream-Migration-Bot",
"gtristan"
],
"repo": "apache/buildstream",
"url": "https://github.com/apache/buildstream/issues/766",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1160648302 | Some logging improvements
See individual commits for details
This looks like a good improvement to me, and I've run the tests locally for good measure.
| gharchive/pull-request | 2022-03-06T16:05:51 | 2025-04-01T04:33:26.638251 | {
"authors": [
"abderrahim",
"gtristan"
],
"repo": "apache/buildstream",
"url": "https://github.com/apache/buildstream/pull/1611",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2342774658 | TLS support of Knative
Requirement
Knative Eventing has recently released TLS support for event delivery, which is currently in the alpha stage. Adding support for this feature in Camel-K would ensure encrypted event delivery.
The annoucement link can be found here
The corresponding github issue can be found here
The public design proposal can be found here
Knative Eventing TLS project board can be found here.
Problem
Currently, Camel-K does not support the newly released TLS feature for Knative Eventing components. This limitation prevents users from leveraging encrypted event delivery within their clusters and third-party sinks when they enabled transport-encryption feature flag in Knative.
Proposal
No response
Open questions
No response
Thanks for the feature request. Just a little notice though. The Knative feature is still marked as "experimental". Any contribution is welcome as usual anyway.
/remove-stale
| gharchive/issue | 2024-06-10T03:47:11 | 2025-04-01T04:33:26.642210 | {
"authors": [
"Leo6Leo",
"squakez"
],
"repo": "apache/camel-k",
"url": "https://github.com/apache/camel-k/issues/5607",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2346790751 | Trait annotations saved on the IntegrationKit resource
What happened?
The IntegrationKit resource seems to save the trait annotations that have been used to build the kit. Some trait annotation configs do not influence the kit and get saved, too. This is a problem because subsequent integrations that resolve to that IntegrationKit inherit the trait annotation config and this influences the integration runtime.
As an example you can use trait.camel.apache.org/knative-service.enabled: 'false' annotation. When the integration that triggers the kit build uses this annotation it gets saved to the IntegrationKit resource:
apiVersion: camel.apache.org/v1
kind: IntegrationKit
metadata:
annotations:
camel.apache.org/operator.id: camel-k
trait.camel.apache.org/knative-service.enabled: 'false'
This is a problem because every integration that uses this IntegrationKit will also inherit the annotation trait config leading to the Knative service trait being disabled for the all integrations that use the kit.
Steps to reproduce
Start an integration that uses trait.camel.apache.org/knative-service.enabled: 'false'
Make sure that the IntegrationKit is build for this integration and see the Knative service condition being disabled explicitly
status:
conditions:
- lastTransitionTime: '2024-06-11T15:59:40Z'
lastUpdateTime: '2024-06-11T15:59:40Z'
message: explicitly disabled
reason: KnativeServiceNotAvailable
status: 'False'
type: KnativeServiceKnativeServiceAvailable
Start an integration that matches the previously built kit but does not use the trait annotation at all
See the Knative service trait also being disabled for this 2nd integration because it is explicitly disabled by trait config inherited from the integration kit
We would expect the Knative service being enabled for the 2nd integration because this is the default behavior when Knative serving is available on the cluster
Relevant log output
No response
Camel K version
2.3.3
Yeah, that's a bad side effect of using trait into annotations. Probably those annotations should be completely filtered out. If you notice, the IntegrationKit, by design has only a subset of traits:
https://github.com/apache/camel-k/blob/489f42ede99ccd91b64bb9efdc28e9593d3770f0/pkg/apis/camel/v1/integrationkit_types.go#L82 and they are populated in the Quarkus trait when creating a new IntegrationKit.
I'm working on the fix, and I've realized the usage of traits in annotations can be simplified a lot in general. However, I am wondering if for Integrations, IntegrationKits and IntegrationPlatforms it does make sense to have the possibility to use also annotations to provide such a configuration. I think it's fine in Pipes, but cannot figure it out the use case for Integrations (ie, if the user can create an annotation, it should be able to edit easily the Integration spec as well).
@lburgazzoli I see this was introduced to support Pipes (well, KameletBindings [1]) but was moved automatically to the others custom resources. I'd be inclined to deprecate, unless there is any valid reason to keep it.
[1] https://github.com/apache/camel-k/pull/2338
I'm working on the fix, and I've realized the usage of traits in annotations can be simplified a lot in general. However, I am wondering if for Integrations, IntegrationKits and IntegrationPlatforms it does make sense to have the possibility to use also annotations to provide such a configuration. I think it's fine in Pipes, but cannot figure it out the use case for Integrations (ie, if the user can create an annotation, it should be able to edit easily the Integration spec as well).
@lburgazzoli I see this was introduced to support Pipes (well, KameletBindings [1]) but was moved automatically to the others custom resources. I'd be inclined to deprecate, unless there is any valid reason to keep it.
No, I don't think there is any more reason to use traits as annotations on resources other than Pipes
| gharchive/issue | 2024-06-11T16:01:30 | 2025-04-01T04:33:26.649843 | {
"authors": [
"christophd",
"lburgazzoli",
"squakez"
],
"repo": "apache/camel-k",
"url": "https://github.com/apache/camel-k/issues/5620",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1695981970 | Upgraded snmp4j to another provider to avoid flaky tests
follows https://github.com/apache/camel-quarkus/pull/4870
follows https://github.com/apache/camel-quarkus/pull/4867#issuecomment-1534290658
Camel in 2.13.x brings org.apache.servicemix.bundles.snmp4j which causes random errors in snmp tests (with low resources). For example:
java.nio.BufferOverflowException
java.util.ConcurrentModificationException
The actual length of the SEQUENCE object org.snmp4j.PDUv1 is 11, but 14 was expected
Solution is to switch to org.snmp4j.snmp4j (by forcing in CQ BOM - which is purpose of this PR).
There is different method signature (during call of SecurityModels.getInstance().addSecurityModel(this.usm);, therefore both producers have to be recompiled in CQ. Because both classes contains a lot of private fileds, it was not possible to override only 1 method. Therefore both classes are copied from the Camel (without any change). If Camel 3.x in future contains org.snmp4j.snmp4j. This PR could be reverted.
@ppalaga @jamesnetherton FYI
Great work, @JiriOndrusek! I am totally happy that we can make this work also on CQ 2.13.
I have a couple of questions/suggestions:
Could you please check what is the impact on the generated catalog in catalog/target/classes/org/apache/camel/catalog/quarkus after building the whole source tree before and after the change? Ideally, there should be no changes. Esp. snmp.json should still be there and the metadata in it should mirror the real state.
We should inform users in the docs that Camel Quarkus uses a different protocol implementation than plain Camel 3.18. Explain reasons, possible gotchas (if any) and explain that org.snmp4j.snmp4j will be used in Camel 4 and CQ 3 anyway.
Practical: Could please the code copied from Camel keep the original java source file structure, so that it is easier to compare what changes have been done between the two and so that it is easier to port any future changes from Camel 3.18?
Question: so now the SNMP tests in CQ 2.13.x and main are 100% identical, right?
@ppalaga thanks for the suggestions!
1 - There is no difference (I used diff with the generated stuff before this PR and after PR)
2,3- Following PR (https://github.com/apache/camel-quarkus/pull/4873) adds doc, refactors class files from Camel (I had to change the name of the files to avoid duplicate java class error)
4 - PR for 2 and 3 with a small upstream change (https://github.com/apache/camel-quarkus/pull/4872), makes tests in CQ 2.13.x and main 100% identical (of course there is a difference javax vs jakarta)
@JiriOndrusek thanks for adressing the suggestions!
I dared to fixup some small glitches in this PR by creating a new PR to avoid another review round: https://github.com/apache/camel-quarkus/pull/4876
First, snm in the org.apache.camel.quarkus.component.snm.graal package name looked like a typo. I changed it to simply org.apache.camel.quarkus.component.snmp because we tend to use graal packages for substitutions. I also moved the camel classes to org.apache.camel.quarkus.component.snmp because I think there is no reason to introduce a duplicate package.
Other than that, i rephrased the docs text a bit. I hope I have not broken anything.
| gharchive/pull-request | 2023-05-04T13:09:00 | 2025-04-01T04:33:26.659448 | {
"authors": [
"JiriOndrusek",
"ppalaga"
],
"repo": "apache/camel-quarkus",
"url": "https://github.com/apache/camel-quarkus/pull/4871",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
176924254 | Camel 10319: SNMP Producer
https://issues.apache.org/jira/browse/CAMEL-10319
Hello @Fabryprog
Can you please amend the commit messages with a reference to the issue CAMEL-10319?
This way we can always track the history of an issue.
Please, run mvn -Psourcecheck too
oscerd@ghost:~/workspace/apache-camel/camel/components/camel-snmp$ mvn -Psourcecheck
[WARNING]
[WARNING] Some problems were encountered while building the effective settings
[WARNING] 'profiles.profile.id' must be unique but found duplicate profile with id jboss-public-repository @ /opt/apache-maven-3.3.9/conf/settings.xml
[WARNING] 'profiles.profile.id' must be unique but found duplicate profile with id redhat-earlyaccess-repository @ /opt/apache-maven-3.3.9/conf/settings.xml
[WARNING]
[INFO] Scanning for projects...
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Building Camel :: SNMP 2.18.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] --- maven-enforcer-plugin:1.4.1:enforce (enforce-maven) @ camel-snmp ---
[INFO]
[INFO] --- maven-bundle-plugin:3.2.0:cleanVersions (versions) @ camel-snmp ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.17:checkstyle (validate) @ camel-snmp ---
[INFO] Starting audit...
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpEndpoint.java:114:5: File contains tab characters (this is the first instance). [FileTabCharacter]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:2: Line does not match expected header line of ' * Licensed to the Apache Software Foundation (ASF) under one or more'. [Header]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:106:17: '{' at column 17 should have line break after. [LeftCurly]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:106:50: 'catch' is not followed by whitespace. [WhitespaceAround]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:106:65: '{' is not followed by whitespace. [WhitespaceAround]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:107:17: '{' at column 17 should have line break after. [LeftCurly]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:107:45: 'catch' is not followed by whitespace. [WhitespaceAround]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/main/java/org/apache/camel/component/snmp/SnmpProducer.java:107:60: '{' is not followed by whitespace. [WhitespaceAround]
[ERROR] /home/oscerd/workspace/apache-camel/camel/components/camel-snmp/src/test/java/org/apache/camel/component/snmp/ProducerTest.java:2: Line does not match expected header line of ' * Licensed to the Apache Software Foundation (ASF) under one or more'. [Header]
Audit done.
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.998 s
[INFO] Finished at: 2016-09-14T17:06:47+02:00
[INFO] Final Memory: 30M/319M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-checkstyle-plugin:2.17:checkstyle (validate) on project camel-snmp: An error has occurred in Checkstyle report generation. Failed during checkstyle execution: There are 9 errors reported by Checkstyle 6.17 with camel-checkstyle.xml ruleset. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
Maybe, squash all the commit into one only if possible.
Thanks.
NOW SOURCE CHECK IS OK!
Can you please amend the commit messages with a reference to the issue CAMEL-10319?
Or squash all the commit into only one?
Thanks!
@lburgazzolii have fix variable into consumer!
I meant if the initialization of targetAddress, address and usm can be moved in onStart and related cleanup onStop , in general the number of object created should be reduced to the min.
@lburgazzoli I entered your valuable guidance
Can you please squash commit once again?
Thanks. After that I'll merge. If there will be any change you can simply track your update on a new JIRA and open a PR.
:-)
@oscerd is it mandatory? :-) i merged the master to test it and now is a very tedious work! :-1:
It isn't mandatory but read "merge branch 'master'" in the git log is so ugly...
Merged! :-)
Many thanks! Can you close this PR?
Thanks :100: !!!!
| gharchive/pull-request | 2016-09-14T14:34:56 | 2025-04-01T04:33:26.667462 | {
"authors": [
"Fabryprog",
"lburgazzoli",
"oscerd"
],
"repo": "apache/camel",
"url": "https://github.com/apache/camel/pull/1177",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
316782457 | [CARBONDATA-2387]Primitive AVRO Datatype Checking
Test case added to check all the primitive datatype support of AVRO and its corresponding mapping with carbon datatypes.
Supported Primitive Datatype.
1. Boolean
2. Int
3. long
4. float -> To carbon Internally it is double.
5. double
6. String
Not Supported
1.NULL Datatype
2.Bytes
[x] Any interfaces changed?
[x] Any backward compatibility impacted?
[x] Document update required?
[x] Testing done
Please provide details on
- Whether new unit test cases have been added or why no new tests are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance test report.
- Any additional information to help reviewers in testing this change.
[x] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA.
Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/4160/
SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/4485/
Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/5344/
| gharchive/pull-request | 2018-04-23T12:03:51 | 2025-04-01T04:33:26.673196 | {
"authors": [
"CarbonDataQA",
"ravipesala",
"sounakr"
],
"repo": "apache/carbondata",
"url": "https://github.com/apache/carbondata/pull/2214",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
345113831 | [wip]disable local dictionary by default
Be sure to do all of the following checklist to help us incorporate
your contribution quickly and easily:
[ ] Any interfaces changed?
[ ] Any backward compatibility impacted?
[ ] Document update required?
[ ] Testing done
Please provide details on
- Whether new unit test cases have been added or why no new tests are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance test report.
- Any additional information to help reviewers in testing this change.
[ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA.
Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7555/
Build Failed with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6309/
Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7564/
Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7568/
SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6033/
retest this please
retest sdv please
Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7603/
retest this please
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7627/
retest this please
retest this please
retest this please
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7639/
Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6364/
SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6070/
Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6380/
retest sdv please
SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6077/
| gharchive/pull-request | 2018-07-27T07:04:12 | 2025-04-01T04:33:26.684042 | {
"authors": [
"CarbonDataQA",
"akashrn5",
"brijoobopanna",
"ravipesala"
],
"repo": "apache/carbondata",
"url": "https://github.com/apache/carbondata/pull/2570",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
392083144 | [CARBONDATA-3161]Pipe dilimiter is not working for streaming table
Be sure to do all of the following checklist to help us incorporate
your contribution quickly and easily:
[Issue]: During data load in streaming scenario the delimiter was taking from CSVInputformat but it was not set in CarbonAppendableStreamSink Class.
[Solution]: Set parameters like Delimiters, Escape char, Max columns in CSVInputFormat for streaming also
[No] Any interfaces changed?
[No] Any backward compatibility impacted?
[No ] Document update required?
[Yes] Testing done
Please provide details on
- Whether new unit test cases have been added or why no new tests are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance test report.
- Any additional information to help reviewers in testing this change.
[No] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA.
Build Success with Spark 2.3.2, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/10082/
Build Success with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/2035/
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/1827/
LGTM
| gharchive/pull-request | 2018-12-18T10:11:45 | 2025-04-01T04:33:26.689096 | {
"authors": [
"CarbonDataQA",
"QiangCai",
"brijoobopanna"
],
"repo": "apache/carbondata",
"url": "https://github.com/apache/carbondata/pull/2997",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1093249249 | [WIP] Fix clean files removing wrong delta files
Why is this PR needed?
In the case where there are multiple delete delta files in a partition in a partition table, some delta files were being ignored and deleted, thus changing the value during the query
What changes were proposed in this PR?
Fixed the logic which checks which delta file to delete. Now checking the deltaStartTime and comparing it with deltaEndTime to check consider all the delta files during clean files.
Does this PR introduce any user interface change?
No
Is any new testcase added?
Yes, one test case has been added.
Build Failed with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4449/
Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6193/
Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/584/
Build Success with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6194/
Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/585/
retest this please
Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/586/
Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6195/
Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4451/
retest this please
Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4452/
Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/587/
Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6197/
Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4453/
Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/588/
Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6198/
Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4454/
Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/589/
Build Failed with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4455/
Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6199/
Build Failed with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/590/
retest this please
Build Failed with Spark 2.3.4, Please check CI http://121.244.95.60:12602/job/ApacheCarbonPRBuilder2.3/6200/
Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_2.4.5/4457/
Build Success with Spark 3.1, Please check CI http://121.244.95.60:12602/job/ApacheCarbon_PR_Builder_3.1/592/
failure is not related to PR
LGTM
| gharchive/pull-request | 2022-01-04T11:12:05 | 2025-04-01T04:33:26.705789 | {
"authors": [
"CarbonDataQA2",
"akashrn5",
"kunal642",
"vikramahuja1001",
"ydvpankaj99"
],
"repo": "apache/carbondata",
"url": "https://github.com/apache/carbondata/pull/4246",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2210701406 | [CELEBORN-1359] Support Netty Logging at the network layer
What changes were proposed in this pull request?
Supporting Netty level logging at the network layer. To configure Netty level logging a LogHandler must be added to the channel pipeline. NettyLogger is introduced as a new class which is able to construct a log handler depending on the log level:
In case of : a custom log handler is created which does not dump the message contents. This way the log is a bit more compact. Moreover when network level encryption is switched on this level might be sufficient.
In case of : Netty's own log handler is used which dumps the message contents.
Otherwise (when the logger is not TRACE or DEBUG) the pipeline does not contain a log handler (there is no runtime penalty for the default setting but a long running app/service must be restarted along with the new log level to have an effect).
Backport:
[SPARK-36719][CORE] Supporting Netty Logging at the network layer
[SPARK-45377][CORE] Handle InputStream in NettyLogger
Why are the changes needed?
This level of logging proved to be sufficient during debugging some external shuffle related problem. Compared with the tcpdump this log lines can be more easily correlated with the Celeborn internal calls. Moreover the log layout can be configured to contain the thread names that way for a timeout a busy thread could be identified.
Does this PR introduce any user-facing change?
No.
How was this patch tested?
Local manually test.
Code change LGTM, please update the PR description, we are using LOG4J2 with log4j2.xml rahter than LOG4J1 with log4j.properties
log4j.logger.XXX
@pan3793, I have updated the description of this pull request. PTAL.
| gharchive/pull-request | 2024-03-27T12:42:08 | 2025-04-01T04:33:26.712589 | {
"authors": [
"SteNicholas",
"pan3793"
],
"repo": "apache/celeborn",
"url": "https://github.com/apache/celeborn/pull/2423",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
528004028 | creating template from snapshot does not work from snapshot list
If you go to STORAGE (main section menu on the left), then select VOLUMES from the dropdown list, select a particular volume, than see snapshots, then select a snaphost - then clicking on button WILL work fine and popup window is show as follows:
But if you go to STORAGE (main section menu on the left), then select SNAPSHOTS from the dropdown list and find one and click the button - then an exception happens in the javascript code:
ISSUE TYPE
Bug Report
COMPONENT NAME
UI
CLOUDSTACK VERSION
4.13
CONFIGURATION
OS / ENVIRONMENT
SUMMARY
STEPS TO REPRODUCE
EXPECTED RESULTS
ACTUAL RESULTS
during the latest update of jquery the createTemplate functions for volumes and snapshots were unified into a shared function without fully abstracting out the entities 'volumes' and 'snapshots'
We can duplicate the function as a quick-fix or put effort into the abstraction. I am choosing the first as the current UI is to be phased out.
@andrijapanicsb @rhtyd @rafaelweingartner
sounds good @DaanHoogland
I've also seen (probably NOT related, but to mention...) that when you register the template from URL, previously there was a spinning circle just until the template was added to the DB - then you can actually click on it and see it being downloaded etc. In current master when you register the template, the circle is spinning all the time, until the template is actually completely downloaded, which is insane. - not sure if that's also a quick fix...
actually @andrijapanicsb half an hour extra search gave me a location to add a conditional, testing is probably going to be more effort.
What you are saying seems a different location in the code. i'll keep it in mind. Did you create an issue for that already?
No @DaanHoogland, I was lazy...
| gharchive/issue | 2019-11-25T10:49:55 | 2025-04-01T04:33:26.727905 | {
"authors": [
"DaanHoogland",
"andrijapanicsb"
],
"repo": "apache/cloudstack",
"url": "https://github.com/apache/cloudstack/issues/3716",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
228969233 | [4.9+] CLOUDSTACK-9860: Power off VMs when force stop is provided
The 'force' option provided with the stopVirtualMachine API command is
often assumed to be a hard shutdown sent to the hypervisor, when in fact
it is for CloudStacks' internal use. CloudStack should be able to send
the 'hard' power-off request to the hosts.
When forced parameter on the stopVM API is true, power off (hard shutdown)
a VM. This uses initial changes from #1635 to pass the forced parameter
to hypervisor plugin via the StopCommand, and fixes force stop (poweroff)
handling for KVM, VMware and XenServer.
This will speed up VM stops with forced=true. Useful for CI-like environments.
The global setting vm.destroy.forcestop is now also honoured to force stop VMs while destroying them, if it is set to true.
Pinging for review @DaanHoogland @abhinandanprateek @nathanejohnson @kiwiflyer @wido @karuturi and others.
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos6 ✔centos7 ✔debian. JID-727
@blueorangutan test
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
Nice @rhtyd thanks, will test it soon
Trillian test result (tid-1099)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 29534 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1099-kvm-centos7.zip
Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py
Test completed. 47 look ok, 2 have error(s)
Test
Result
Time (s)
Test File
test_04_rvpc_privategw_static_routes
Failure
376.63
test_privategw_acl.py
test_02_list_snapshots_with_removed_data_store
Error
0.04
test_snapshots.py
test_01_vpc_site2site_vpn
Success
180.96
test_vpc_vpn.py
test_01_vpc_remote_access_vpn
Success
71.17
test_vpc_vpn.py
test_01_redundant_vpc_site2site_vpn
Success
265.85
test_vpc_vpn.py
test_02_VPC_default_routes
Success
269.30
test_vpc_router_nics.py
test_01_VPC_nics_after_destroy
Success
588.43
test_vpc_router_nics.py
test_05_rvpc_multi_tiers
Success
527.95
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Success
1306.50
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Success
574.37
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Success
786.83
test_vpc_redundant.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Success
1327.73
test_vpc_redundant.py
test_09_delete_detached_volume
Success
156.58
test_volumes.py
test_08_resize_volume
Success
156.41
test_volumes.py
test_07_resize_fail
Success
161.51
test_volumes.py
test_06_download_detached_volume
Success
156.39
test_volumes.py
test_05_detach_volume
Success
150.71
test_volumes.py
test_04_delete_attached_volume
Success
151.24
test_volumes.py
test_03_download_attached_volume
Success
156.31
test_volumes.py
test_02_attach_volume
Success
89.42
test_volumes.py
test_01_create_volume
Success
712.21
test_volumes.py
test_deploy_vm_multiple
Success
318.15
test_vm_life_cycle.py
test_deploy_vm
Success
0.03
test_vm_life_cycle.py
test_advZoneVirtualRouter
Success
0.02
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Success
36.71
test_vm_life_cycle.py
test_09_expunge_vm
Success
125.29
test_vm_life_cycle.py
test_08_migrate_vm
Success
40.96
test_vm_life_cycle.py
test_07_restore_vm
Success
0.13
test_vm_life_cycle.py
test_06_destroy_vm
Success
130.90
test_vm_life_cycle.py
test_03_reboot_vm
Success
125.87
test_vm_life_cycle.py
test_02_start_vm
Success
10.20
test_vm_life_cycle.py
test_01_stop_vm
Success
35.30
test_vm_life_cycle.py
test_CreateTemplateWithDuplicateName
Success
95.77
test_templates.py
test_08_list_system_templates
Success
0.03
test_templates.py
test_07_list_public_templates
Success
0.04
test_templates.py
test_05_template_permissions
Success
0.06
test_templates.py
test_04_extract_template
Success
5.16
test_templates.py
test_03_delete_template
Success
5.12
test_templates.py
test_02_edit_template
Success
90.16
test_templates.py
test_01_create_template
Success
105.85
test_templates.py
test_10_destroy_cpvm
Success
191.71
test_ssvm.py
test_09_destroy_ssvm
Success
168.56
test_ssvm.py
test_08_reboot_cpvm
Success
101.57
test_ssvm.py
test_07_reboot_ssvm
Success
133.60
test_ssvm.py
test_06_stop_cpvm
Success
136.74
test_ssvm.py
test_05_stop_ssvm
Success
133.67
test_ssvm.py
test_04_cpvm_internals
Success
1.20
test_ssvm.py
test_03_ssvm_internals
Success
3.34
test_ssvm.py
test_02_list_cpvm_vm
Success
0.12
test_ssvm.py
test_01_list_sec_storage_vm
Success
0.13
test_ssvm.py
test_01_snapshot_root_disk
Success
16.18
test_snapshots.py
test_04_change_offering_small
Success
211.01
test_service_offerings.py
test_03_delete_service_offering
Success
0.05
test_service_offerings.py
test_02_edit_service_offering
Success
0.08
test_service_offerings.py
test_01_create_service_offering
Success
0.11
test_service_offerings.py
test_02_sys_template_ready
Success
0.14
test_secondary_storage.py
test_01_sys_vm_start
Success
0.23
test_secondary_storage.py
test_09_reboot_router
Success
40.36
test_routers.py
test_08_start_router
Success
35.36
test_routers.py
test_07_stop_router
Success
10.17
test_routers.py
test_06_router_advanced
Success
0.06
test_routers.py
test_05_router_basic
Success
0.04
test_routers.py
test_04_restart_network_wo_cleanup
Success
5.80
test_routers.py
test_03_restart_network_cleanup
Success
65.55
test_routers.py
test_02_router_internal_adv
Success
1.11
test_routers.py
test_01_router_internal_basic
Success
0.52
test_routers.py
test_router_dns_guestipquery
Success
77.20
test_router_dns.py
test_router_dns_externalipquery
Success
0.08
test_router_dns.py
test_router_dhcphosts
Success
272.78
test_router_dhcphosts.py
test_router_dhcp_opts
Success
21.81
test_router_dhcphosts.py
test_01_updatevolumedetail
Success
5.15
test_resource_detail.py
test_01_reset_vm_on_reboot
Success
181.44
test_reset_vm_on_reboot.py
test_createRegion
Success
0.05
test_regions.py
test_create_pvlan_network
Success
5.35
test_pvlan.py
test_dedicatePublicIpRange
Success
0.67
test_public_ip_range.py
test_03_vpc_privategw_restart_vpc_cleanup
Success
510.53
test_privategw_acl.py
test_02_vpc_privategw_static_routes
Success
446.30
test_privategw_acl.py
test_01_vpc_privategw_acl
Success
97.33
test_privategw_acl.py
test_01_primary_storage_nfs
Success
35.86
test_primary_storage.py
test_createPortablePublicIPRange
Success
10.19
test_portable_publicip.py
test_createPortablePublicIPAcquire
Success
15.55
test_portable_publicip.py
test_isolate_network_password_server
Success
60.31
test_password_server.py
test_UpdateStorageOverProvisioningFactor
Success
0.12
test_over_provisioning.py
test_oobm_zchange_password
Success
30.72
test_outofbandmanagement.py
test_oobm_multiple_mgmt_server_ownership
Success
16.37
test_outofbandmanagement.py
test_oobm_issue_power_status
Success
5.26
test_outofbandmanagement.py
test_oobm_issue_power_soft
Success
15.35
test_outofbandmanagement.py
test_oobm_issue_power_reset
Success
15.37
test_outofbandmanagement.py
test_oobm_issue_power_on
Success
15.40
test_outofbandmanagement.py
test_oobm_issue_power_off
Success
10.39
test_outofbandmanagement.py
test_oobm_issue_power_cycle
Success
10.38
test_outofbandmanagement.py
test_oobm_enabledisable_across_clusterzones
Success
82.73
test_outofbandmanagement.py
test_oobm_enable_feature_valid
Success
5.16
test_outofbandmanagement.py
test_oobm_enable_feature_invalid
Success
0.10
test_outofbandmanagement.py
test_oobm_disable_feature_valid
Success
5.25
test_outofbandmanagement.py
test_oobm_disable_feature_invalid
Success
0.14
test_outofbandmanagement.py
test_oobm_configure_invalid_driver
Success
0.12
test_outofbandmanagement.py
test_oobm_configure_default_driver
Success
0.13
test_outofbandmanagement.py
test_oobm_background_powerstate_sync
Success
23.52
test_outofbandmanagement.py
test_extendPhysicalNetworkVlan
Success
15.47
test_non_contigiousvlan.py
test_01_nic
Success
439.34
test_nic.py
test_releaseIP
Success
192.59
test_network.py
test_reboot_router
Success
403.51
test_network.py
test_public_ip_user_account
Success
10.28
test_network.py
test_public_ip_admin_account
Success
40.29
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Success
66.95
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Success
61.78
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Success
125.04
test_network.py
test_delete_account
Success
318.06
test_network.py
test_02_port_fwd_on_non_src_nat
Success
55.71
test_network.py
test_01_port_fwd_on_src_nat
Success
111.87
test_network.py
test_nic_secondaryip_add_remove
Success
298.28
test_multipleips_per_nic.py
test_list_zones_metrics
Success
0.27
test_metrics_api.py
test_list_volumes_metrics
Success
5.56
test_metrics_api.py
test_list_vms_metrics
Success
191.85
test_metrics_api.py
test_list_pstorage_metrics
Success
0.40
test_metrics_api.py
test_list_infrastructure_metrics
Success
0.55
test_metrics_api.py
test_list_hosts_metrics
Success
0.36
test_metrics_api.py
test_list_clusters_metrics
Success
0.38
test_metrics_api.py
login_test_saml_user
Success
19.17
test_login.py
test_assign_and_removal_lb
Success
133.74
test_loadbalance.py
test_02_create_lb_rule_non_nat
Success
187.41
test_loadbalance.py
test_01_create_lb_rule_src_nat
Success
219.22
test_loadbalance.py
test_03_list_snapshots
Success
0.08
test_list_ids_parameter.py
test_02_list_templates
Success
0.04
test_list_ids_parameter.py
test_01_list_volumes
Success
0.05
test_list_ids_parameter.py
test_07_list_default_iso
Success
0.06
test_iso.py
test_05_iso_permissions
Success
0.07
test_iso.py
test_04_extract_Iso
Success
5.19
test_iso.py
test_03_delete_iso
Success
95.17
test_iso.py
test_02_edit_iso
Success
0.06
test_iso.py
test_01_create_iso
Success
21.06
test_iso.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Success
228.94
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Success
148.70
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Success
556.22
test_internal_lb.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Success
431.23
test_internal_lb.py
test_dedicateGuestVlanRange
Success
10.28
test_guest_vlan_range.py
test_UpdateConfigParamWithScope
Success
0.14
test_global_settings.py
test_rolepermission_lifecycle_update
Success
6.14
test_dynamicroles.py
test_rolepermission_lifecycle_list
Success
6.01
test_dynamicroles.py
test_rolepermission_lifecycle_delete
Success
5.85
test_dynamicroles.py
test_rolepermission_lifecycle_create
Success
5.89
test_dynamicroles.py
test_rolepermission_lifecycle_concurrent_updates
Success
6.02
test_dynamicroles.py
test_role_lifecycle_update_role_inuse
Success
5.89
test_dynamicroles.py
test_role_lifecycle_update
Success
10.99
test_dynamicroles.py
test_role_lifecycle_list
Success
5.92
test_dynamicroles.py
test_role_lifecycle_delete
Success
10.95
test_dynamicroles.py
test_role_lifecycle_create
Success
5.93
test_dynamicroles.py
test_role_inuse_deletion
Success
5.92
test_dynamicroles.py
test_role_account_acls_multiple_mgmt_servers
Success
8.13
test_dynamicroles.py
test_role_account_acls
Success
8.28
test_dynamicroles.py
test_default_role_deletion
Success
6.05
test_dynamicroles.py
test_04_create_fat_type_disk_offering
Success
0.11
test_disk_offerings.py
test_03_delete_disk_offering
Success
0.05
test_disk_offerings.py
test_02_edit_disk_offering
Success
0.07
test_disk_offerings.py
test_02_create_sparse_type_disk_offering
Success
0.10
test_disk_offerings.py
test_01_create_disk_offering
Success
0.12
test_disk_offerings.py
test_deployvm_userdispersing
Success
20.59
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userconcentrated
Success
50.78
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_firstfit
Success
60.71
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userdata_post
Success
10.39
test_deploy_vm_with_userdata.py
test_deployvm_userdata
Success
70.84
test_deploy_vm_with_userdata.py
test_02_deploy_vm_root_resize
Success
5.96
test_deploy_vm_root_resize.py
test_01_deploy_vm_root_resize
Success
6.00
test_deploy_vm_root_resize.py
test_00_deploy_vm_root_resize
Success
283.01
test_deploy_vm_root_resize.py
test_deploy_vm_from_iso
Success
217.56
test_deploy_vm_iso.py
test_DeployVmAntiAffinityGroup
Success
76.08
test_affinity_groups.py
test_03_delete_vm_snapshots
Skipped
0.00
test_vm_snapshots.py
test_02_revert_vm_snapshots
Skipped
0.00
test_vm_snapshots.py
test_01_create_vm_snapshots
Skipped
0.00
test_vm_snapshots.py
test_06_copy_template
Skipped
0.00
test_templates.py
test_static_role_account_acls
Skipped
0.02
test_staticroles.py
test_01_scale_vm
Skipped
0.00
test_scale_vm.py
test_01_primary_storage_iscsi
Skipped
0.06
test_primary_storage.py
test_06_copy_iso
Skipped
0.00
test_iso.py
test_deploy_vgpu_enabled_vm
Skipped
0.01
test_deploy_vgpu_enabled_vm.py
Looks good @rhtyd !
LGTM based on the code
@rhtyd
In cloudstack api doc, forced means "Force stop the VM (vm is marked as Stopped even when command fails to be send to the backend). The caller knows the VM is stopped."
If you change the process, the api doc needs to be changed as well
@ustcweizhou thanks, fixed the api doc for stopVirtualMachine API and updated the JIRA id.
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos6 ✔centos7 ✔debian. JID-729
@blueorangutan test centos7 vmware-55u3
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + vmware-55u3) has been kicked to run smoke tests
code LGTM.
Trillian test result (tid-1100)
Environment: vmware-55u3 (x2), Advanced Networking with Mgmt server 7
Total time taken: 55717 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1100-vmware-55u3.zip
Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermitten failure detected: /marvin/tests/smoke/test_routers.py
Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py
Intermitten failure detected: /marvin/tests/smoke/test_ssvm.py
Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermitten failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Test completed. 44 look ok, 5 have error(s)
Test
Result
Time (s)
Test File
test_01_vpc_remote_access_vpn
Failure
961.01
test_vpc_vpn.py
test_04_cpvm_internals
Failure
0.65
test_ssvm.py
test_04_rvpc_privategw_static_routes
Failure
899.97
test_privategw_acl.py
test_01_vpc_privategw_acl
Failure
977.00
test_privategw_acl.py
test_04_rvpc_network_garbage_collector_nics
Error
961.96
test_vpc_redundant.py
test_02_list_snapshots_with_removed_data_store
Error
86.03
test_snapshots.py
test_02_list_snapshots_with_removed_data_store
Error
91.13
test_snapshots.py
test_01_vpc_site2site_vpn
Success
376.46
test_vpc_vpn.py
test_01_redundant_vpc_site2site_vpn
Success
624.65
test_vpc_vpn.py
test_02_VPC_default_routes
Success
340.94
test_vpc_router_nics.py
test_01_VPC_nics_after_destroy
Success
726.70
test_vpc_router_nics.py
test_05_rvpc_multi_tiers
Success
689.97
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Success
716.86
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Success
730.98
test_vpc_redundant.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Success
1421.53
test_vpc_redundant.py
test_09_delete_detached_volume
Success
31.03
test_volumes.py
test_06_download_detached_volume
Success
70.73
test_volumes.py
test_05_detach_volume
Success
100.25
test_volumes.py
test_04_delete_attached_volume
Success
10.22
test_volumes.py
test_03_download_attached_volume
Success
15.34
test_volumes.py
test_02_attach_volume
Success
96.00
test_volumes.py
test_01_create_volume
Success
520.84
test_volumes.py
test_03_delete_vm_snapshots
Success
275.22
test_vm_snapshots.py
test_02_revert_vm_snapshots
Success
232.23
test_vm_snapshots.py
test_01_create_vm_snapshots
Success
161.79
test_vm_snapshots.py
test_deploy_vm_multiple
Success
273.00
test_vm_life_cycle.py
test_deploy_vm
Success
0.03
test_vm_life_cycle.py
test_advZoneVirtualRouter
Success
0.03
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Success
26.88
test_vm_life_cycle.py
test_09_expunge_vm
Success
125.26
test_vm_life_cycle.py
test_08_migrate_vm
Success
81.18
test_vm_life_cycle.py
test_07_restore_vm
Success
0.25
test_vm_life_cycle.py
test_06_destroy_vm
Success
10.18
test_vm_life_cycle.py
test_03_reboot_vm
Success
5.18
test_vm_life_cycle.py
test_02_start_vm
Success
25.32
test_vm_life_cycle.py
test_01_stop_vm
Success
10.18
test_vm_life_cycle.py
test_CreateTemplateWithDuplicateName
Success
302.16
test_templates.py
test_08_list_system_templates
Success
0.03
test_templates.py
test_07_list_public_templates
Success
0.04
test_templates.py
test_05_template_permissions
Success
0.07
test_templates.py
test_04_extract_template
Success
15.23
test_templates.py
test_03_delete_template
Success
5.11
test_templates.py
test_02_edit_template
Success
90.13
test_templates.py
test_01_create_template
Success
141.13
test_templates.py
test_10_destroy_cpvm
Success
267.05
test_ssvm.py
test_09_destroy_ssvm
Success
269.14
test_ssvm.py
test_08_reboot_cpvm
Success
156.62
test_ssvm.py
test_07_reboot_ssvm
Success
158.69
test_ssvm.py
test_06_stop_cpvm
Success
177.04
test_ssvm.py
test_05_stop_ssvm
Success
203.94
test_ssvm.py
test_03_ssvm_internals
Success
3.15
test_ssvm.py
test_02_list_cpvm_vm
Success
0.15
test_ssvm.py
test_01_list_sec_storage_vm
Success
0.17
test_ssvm.py
test_01_snapshot_root_disk
Success
66.55
test_snapshots.py
test_04_change_offering_small
Success
92.11
test_service_offerings.py
test_03_delete_service_offering
Success
0.06
test_service_offerings.py
test_02_edit_service_offering
Success
0.12
test_service_offerings.py
test_01_create_service_offering
Success
0.12
test_service_offerings.py
test_02_sys_template_ready
Success
0.16
test_secondary_storage.py
test_01_sys_vm_start
Success
0.24
test_secondary_storage.py
test_09_reboot_router
Success
156.01
test_routers.py
test_08_start_router
Success
135.91
test_routers.py
test_07_stop_router
Success
25.27
test_routers.py
test_06_router_advanced
Success
0.06
test_routers.py
test_05_router_basic
Success
0.04
test_routers.py
test_04_restart_network_wo_cleanup
Success
5.67
test_routers.py
test_03_restart_network_cleanup
Success
191.42
test_routers.py
test_02_router_internal_adv
Success
1.18
test_routers.py
test_01_router_internal_basic
Success
0.58
test_routers.py
test_router_dns_guestipquery
Success
76.85
test_router_dns.py
test_router_dns_externalipquery
Success
0.07
test_router_dns.py
test_router_dhcphosts
Success
121.81
test_router_dhcphosts.py
test_router_dhcp_opts
Success
21.76
test_router_dhcphosts.py
test_01_updatevolumedetail
Success
0.12
test_resource_detail.py
test_01_reset_vm_on_reboot
Success
65.62
test_reset_vm_on_reboot.py
test_createRegion
Success
0.04
test_regions.py
test_create_pvlan_network
Success
5.27
test_pvlan.py
test_dedicatePublicIpRange
Success
0.54
test_public_ip_range.py
test_03_vpc_privategw_restart_vpc_cleanup
Success
1135.21
test_privategw_acl.py
test_02_vpc_privategw_static_routes
Success
722.96
test_privategw_acl.py
test_01_primary_storage_nfs
Success
37.52
test_primary_storage.py
test_createPortablePublicIPRange
Success
15.25
test_portable_publicip.py
test_createPortablePublicIPAcquire
Success
15.57
test_portable_publicip.py
test_isolate_network_password_server
Success
64.85
test_password_server.py
test_UpdateStorageOverProvisioningFactor
Success
0.15
test_over_provisioning.py
test_oobm_zchange_password
Success
30.85
test_outofbandmanagement.py
test_oobm_multiple_mgmt_server_ownership
Success
16.47
test_outofbandmanagement.py
test_oobm_issue_power_status
Success
10.30
test_outofbandmanagement.py
test_oobm_issue_power_soft
Success
15.42
test_outofbandmanagement.py
test_oobm_issue_power_reset
Success
15.52
test_outofbandmanagement.py
test_oobm_issue_power_on
Success
15.35
test_outofbandmanagement.py
test_oobm_issue_power_off
Success
15.38
test_outofbandmanagement.py
test_oobm_issue_power_cycle
Success
15.55
test_outofbandmanagement.py
test_oobm_enabledisable_across_clusterzones
Success
82.98
test_outofbandmanagement.py
test_oobm_enable_feature_valid
Success
5.19
test_outofbandmanagement.py
test_oobm_enable_feature_invalid
Success
0.12
test_outofbandmanagement.py
test_oobm_disable_feature_valid
Success
5.20
test_outofbandmanagement.py
test_oobm_disable_feature_invalid
Success
0.13
test_outofbandmanagement.py
test_oobm_configure_invalid_driver
Success
0.12
test_outofbandmanagement.py
test_oobm_configure_default_driver
Success
0.10
test_outofbandmanagement.py
test_oobm_background_powerstate_sync
Success
23.53
test_outofbandmanagement.py
test_extendPhysicalNetworkVlan
Success
15.39
test_non_contigiousvlan.py
test_01_nic
Success
490.48
test_nic.py
test_releaseIP
Success
338.83
test_network.py
test_reboot_router
Success
625.22
test_network.py
test_public_ip_user_account
Success
10.29
test_network.py
test_public_ip_admin_account
Success
40.32
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Success
77.04
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Success
61.91
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Success
125.33
test_network.py
test_delete_account
Success
363.77
test_network.py
test_02_port_fwd_on_non_src_nat
Success
55.83
test_network.py
test_01_port_fwd_on_src_nat
Success
111.89
test_network.py
test_nic_secondaryip_add_remove
Success
218.15
test_multipleips_per_nic.py
test_list_zones_metrics
Success
0.25
test_metrics_api.py
test_list_volumes_metrics
Success
5.53
test_metrics_api.py
test_list_vms_metrics
Success
252.40
test_metrics_api.py
test_list_pstorage_metrics
Success
0.29
test_metrics_api.py
test_list_infrastructure_metrics
Success
0.45
test_metrics_api.py
test_list_hosts_metrics
Success
0.48
test_metrics_api.py
test_list_clusters_metrics
Success
0.41
test_metrics_api.py
login_test_saml_user
Success
19.41
test_login.py
test_assign_and_removal_lb
Success
148.48
test_loadbalance.py
test_02_create_lb_rule_non_nat
Success
207.42
test_loadbalance.py
test_01_create_lb_rule_src_nat
Success
207.97
test_loadbalance.py
test_03_list_snapshots
Success
0.09
test_list_ids_parameter.py
test_02_list_templates
Success
0.05
test_list_ids_parameter.py
test_01_list_volumes
Success
0.03
test_list_ids_parameter.py
test_07_list_default_iso
Success
0.07
test_iso.py
test_05_iso_permissions
Success
0.07
test_iso.py
test_04_extract_Iso
Success
5.14
test_iso.py
test_03_delete_iso
Success
95.18
test_iso.py
test_02_edit_iso
Success
0.06
test_iso.py
test_01_create_iso
Success
21.01
test_iso.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Success
586.29
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Success
460.80
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Success
1073.94
test_internal_lb.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Success
802.49
test_internal_lb.py
test_dedicateGuestVlanRange
Success
10.31
test_guest_vlan_range.py
test_UpdateConfigParamWithScope
Success
0.14
test_global_settings.py
test_rolepermission_lifecycle_update
Success
6.36
test_dynamicroles.py
test_rolepermission_lifecycle_list
Success
6.09
test_dynamicroles.py
test_rolepermission_lifecycle_delete
Success
5.99
test_dynamicroles.py
test_rolepermission_lifecycle_create
Success
5.92
test_dynamicroles.py
test_rolepermission_lifecycle_concurrent_updates
Success
6.19
test_dynamicroles.py
test_role_lifecycle_update_role_inuse
Success
6.00
test_dynamicroles.py
test_role_lifecycle_update
Success
6.05
test_dynamicroles.py
test_role_lifecycle_list
Success
5.96
test_dynamicroles.py
test_role_lifecycle_delete
Success
5.98
test_dynamicroles.py
test_role_lifecycle_create
Success
5.95
test_dynamicroles.py
test_role_inuse_deletion
Success
5.92
test_dynamicroles.py
test_role_account_acls_multiple_mgmt_servers
Success
8.46
test_dynamicroles.py
test_role_account_acls
Success
8.78
test_dynamicroles.py
test_default_role_deletion
Success
6.04
test_dynamicroles.py
test_04_create_fat_type_disk_offering
Success
0.07
test_disk_offerings.py
test_03_delete_disk_offering
Success
0.04
test_disk_offerings.py
test_02_edit_disk_offering
Success
0.05
test_disk_offerings.py
test_02_create_sparse_type_disk_offering
Success
0.08
test_disk_offerings.py
test_01_create_disk_offering
Success
0.11
test_disk_offerings.py
test_deployvm_userdispersing
Success
65.98
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userconcentrated
Success
106.19
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_firstfit
Success
201.69
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userdata_post
Success
30.54
test_deploy_vm_with_userdata.py
test_deployvm_userdata
Success
156.46
test_deploy_vm_with_userdata.py
test_02_deploy_vm_root_resize
Success
5.96
test_deploy_vm_root_resize.py
test_01_deploy_vm_root_resize
Success
5.99
test_deploy_vm_root_resize.py
test_00_deploy_vm_root_resize
Success
6.39
test_deploy_vm_root_resize.py
test_deploy_vm_from_iso
Success
207.58
test_deploy_vm_iso.py
test_DeployVmAntiAffinityGroup
Success
282.58
test_affinity_groups.py
test_08_resize_volume
Skipped
5.12
test_volumes.py
test_07_resize_fail
Skipped
10.31
test_volumes.py
test_06_copy_template
Skipped
0.00
test_templates.py
test_static_role_account_acls
Skipped
0.04
test_staticroles.py
test_01_scale_vm
Skipped
66.35
test_scale_vm.py
test_01_primary_storage_iscsi
Skipped
0.04
test_primary_storage.py
test_06_copy_iso
Skipped
0.00
test_iso.py
test_deploy_vgpu_enabled_vm
Skipped
0.01
test_deploy_vgpu_enabled_vm.py
@blueorangutan test
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
There seems to be some new (intermittent) errors that don't seem to be related to this PR, however I would like to investigate and possibly fix them (in case they are caused some other recent changes to 4.9/master). I'll also kick tests for xenserver:
@blueorangutan test centos7 xenserver-65sp1
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + xenserver-65sp1) has been kicked to run smoke tests
Trillian test result (tid-1104)
Environment: xenserver-65sp1 (x2), Advanced Networking with Mgmt server 7
Total time taken: 49407 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1104-xenserver-65sp1.zip
Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py
Intermitten failure detected: /marvin/tests/smoke/test_templates.py
Intermitten failure detected: /marvin/tests/smoke/test_volumes.py
Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Test completed. 45 look ok, 4 have error(s)
Test
Result
Time (s)
Test File
test_05_rvpc_multi_tiers
Failure
493.10
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Failure
1427.86
test_vpc_redundant.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Failure
585.31
test_vpc_redundant.py
test_04_rvpc_privategw_static_routes
Failure
846.93
test_privategw_acl.py
test_01_create_template
Error
80.55
test_templates.py
test_02_list_snapshots_with_removed_data_store
Error
98.86
test_snapshots.py
test_02_list_snapshots_with_removed_data_store
Error
103.92
test_snapshots.py
test_01_vpc_site2site_vpn
Success
395.67
test_vpc_vpn.py
test_01_vpc_remote_access_vpn
Success
206.46
test_vpc_vpn.py
test_01_redundant_vpc_site2site_vpn
Success
566.05
test_vpc_vpn.py
test_02_VPC_default_routes
Success
570.10
test_vpc_router_nics.py
test_01_VPC_nics_after_destroy
Success
838.70
test_vpc_router_nics.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Success
927.60
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Success
1135.49
test_vpc_redundant.py
test_09_delete_detached_volume
Success
10.45
test_volumes.py
test_08_resize_volume
Success
115.84
test_volumes.py
test_07_resize_fail
Success
125.92
test_volumes.py
test_06_download_detached_volume
Success
20.28
test_volumes.py
test_05_detach_volume
Success
110.27
test_volumes.py
test_04_delete_attached_volume
Success
15.17
test_volumes.py
test_03_download_attached_volume
Success
15.22
test_volumes.py
test_02_attach_volume
Success
15.87
test_volumes.py
test_01_create_volume
Success
393.08
test_volumes.py
test_03_delete_vm_snapshots
Success
280.20
test_vm_snapshots.py
test_02_revert_vm_snapshots
Success
232.46
test_vm_snapshots.py
test_01_create_vm_snapshots
Success
130.94
test_vm_snapshots.py
test_deploy_vm_multiple
Success
317.28
test_vm_life_cycle.py
test_deploy_vm
Success
0.02
test_vm_life_cycle.py
test_advZoneVirtualRouter
Success
0.02
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Success
118.16
test_vm_life_cycle.py
test_09_expunge_vm
Success
125.14
test_vm_life_cycle.py
test_08_migrate_vm
Success
90.95
test_vm_life_cycle.py
test_07_restore_vm
Success
0.07
test_vm_life_cycle.py
test_06_destroy_vm
Success
15.17
test_vm_life_cycle.py
test_03_reboot_vm
Success
20.18
test_vm_life_cycle.py
test_02_start_vm
Success
25.18
test_vm_life_cycle.py
test_01_stop_vm
Success
35.24
test_vm_life_cycle.py
test_CreateTemplateWithDuplicateName
Success
196.08
test_templates.py
test_08_list_system_templates
Success
0.02
test_templates.py
test_07_list_public_templates
Success
0.03
test_templates.py
test_05_template_permissions
Success
0.04
test_templates.py
test_04_extract_template
Success
5.24
test_templates.py
test_03_delete_template
Success
5.08
test_templates.py
test_02_edit_template
Success
90.05
test_templates.py
test_10_destroy_cpvm
Success
196.40
test_ssvm.py
test_09_destroy_ssvm
Success
234.40
test_ssvm.py
test_08_reboot_cpvm
Success
156.51
test_ssvm.py
test_07_reboot_ssvm
Success
304.67
test_ssvm.py
test_06_stop_cpvm
Success
181.77
test_ssvm.py
test_05_stop_ssvm
Success
204.90
test_ssvm.py
test_04_cpvm_internals
Success
1.32
test_ssvm.py
test_03_ssvm_internals
Success
3.69
test_ssvm.py
test_02_list_cpvm_vm
Success
0.10
test_ssvm.py
test_01_list_sec_storage_vm
Success
0.10
test_ssvm.py
test_01_snapshot_root_disk
Success
36.05
test_snapshots.py
test_04_change_offering_small
Success
121.29
test_service_offerings.py
test_03_delete_service_offering
Success
0.03
test_service_offerings.py
test_02_edit_service_offering
Success
0.04
test_service_offerings.py
test_01_create_service_offering
Success
0.06
test_service_offerings.py
test_02_sys_template_ready
Success
0.12
test_secondary_storage.py
test_01_sys_vm_start
Success
0.14
test_secondary_storage.py
test_01_scale_vm
Success
5.13
test_scale_vm.py
test_09_reboot_router
Success
105.55
test_routers.py
test_08_start_router
Success
70.39
test_routers.py
test_07_stop_router
Success
20.16
test_routers.py
test_06_router_advanced
Success
0.04
test_routers.py
test_05_router_basic
Success
0.03
test_routers.py
test_04_restart_network_wo_cleanup
Success
5.56
test_routers.py
test_03_restart_network_cleanup
Success
160.81
test_routers.py
test_02_router_internal_adv
Success
0.70
test_routers.py
test_01_router_internal_basic
Success
0.37
test_routers.py
test_router_dns_guestipquery
Success
48.84
test_router_dns.py
test_router_dns_externalipquery
Success
0.06
test_router_dns.py
test_router_dhcphosts
Success
116.97
test_router_dhcphosts.py
test_router_dhcp_opts
Success
31.23
test_router_dhcphosts.py
test_01_updatevolumedetail
Success
5.08
test_resource_detail.py
test_01_reset_vm_on_reboot
Success
201.09
test_reset_vm_on_reboot.py
test_createRegion
Success
0.03
test_regions.py
test_create_pvlan_network
Success
5.16
test_pvlan.py
test_dedicatePublicIpRange
Success
0.27
test_public_ip_range.py
test_03_vpc_privategw_restart_vpc_cleanup
Success
764.48
test_privategw_acl.py
test_02_vpc_privategw_static_routes
Success
785.96
test_privategw_acl.py
test_01_vpc_privategw_acl
Success
252.27
test_privategw_acl.py
test_01_primary_storage_nfs
Success
38.45
test_primary_storage.py
test_01_primary_storage_iscsi
Success
77.64
test_primary_storage.py
test_createPortablePublicIPRange
Success
15.15
test_portable_publicip.py
test_createPortablePublicIPAcquire
Success
10.29
test_portable_publicip.py
test_isolate_network_password_server
Success
65.95
test_password_server.py
test_UpdateStorageOverProvisioningFactor
Success
0.09
test_over_provisioning.py
test_oobm_zchange_password
Success
25.56
test_outofbandmanagement.py
test_oobm_multiple_mgmt_server_ownership
Success
16.28
test_outofbandmanagement.py
test_oobm_issue_power_status
Success
10.19
test_outofbandmanagement.py
test_oobm_issue_power_soft
Success
10.24
test_outofbandmanagement.py
test_oobm_issue_power_reset
Success
15.26
test_outofbandmanagement.py
test_oobm_issue_power_on
Success
15.28
test_outofbandmanagement.py
test_oobm_issue_power_off
Success
15.28
test_outofbandmanagement.py
test_oobm_issue_power_cycle
Success
15.29
test_outofbandmanagement.py
test_oobm_enabledisable_across_clusterzones
Success
102.34
test_outofbandmanagement.py
test_oobm_enable_feature_valid
Success
0.10
test_outofbandmanagement.py
test_oobm_enable_feature_invalid
Success
0.07
test_outofbandmanagement.py
test_oobm_disable_feature_valid
Success
5.13
test_outofbandmanagement.py
test_oobm_disable_feature_invalid
Success
0.08
test_outofbandmanagement.py
test_oobm_configure_invalid_driver
Success
0.06
test_outofbandmanagement.py
test_oobm_configure_default_driver
Success
0.06
test_outofbandmanagement.py
test_oobm_background_powerstate_sync
Success
29.31
test_outofbandmanagement.py
test_extendPhysicalNetworkVlan
Success
15.24
test_non_contigiousvlan.py
test_01_nic
Success
574.15
test_nic.py
test_releaseIP
Success
297.41
test_network.py
test_reboot_router
Success
624.00
test_network.py
test_public_ip_user_account
Success
10.20
test_network.py
test_public_ip_admin_account
Success
40.21
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Success
86.50
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Success
76.22
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Success
102.65
test_network.py
test_delete_account
Success
398.07
test_network.py
test_02_port_fwd_on_non_src_nat
Success
70.68
test_network.py
test_01_port_fwd_on_src_nat
Success
83.76
test_network.py
test_nic_secondaryip_add_remove
Success
126.90
test_multipleips_per_nic.py
test_list_zones_metrics
Success
0.21
test_metrics_api.py
test_list_volumes_metrics
Success
0.35
test_metrics_api.py
test_list_vms_metrics
Success
222.91
test_metrics_api.py
test_list_pstorage_metrics
Success
0.53
test_metrics_api.py
test_list_infrastructure_metrics
Success
0.55
test_metrics_api.py
test_list_hosts_metrics
Success
0.41
test_metrics_api.py
test_list_clusters_metrics
Success
0.62
test_metrics_api.py
login_test_saml_user
Success
18.09
test_login.py
test_assign_and_removal_lb
Success
148.48
test_loadbalance.py
test_02_create_lb_rule_non_nat
Success
207.38
test_loadbalance.py
test_01_create_lb_rule_src_nat
Success
207.63
test_loadbalance.py
test_03_list_snapshots
Success
0.06
test_list_ids_parameter.py
test_02_list_templates
Success
0.03
test_list_ids_parameter.py
test_01_list_volumes
Success
0.03
test_list_ids_parameter.py
test_07_list_default_iso
Success
0.05
test_iso.py
test_05_iso_permissions
Success
0.05
test_iso.py
test_04_extract_Iso
Success
5.21
test_iso.py
test_03_delete_iso
Success
95.14
test_iso.py
test_02_edit_iso
Success
0.04
test_iso.py
test_01_create_iso
Success
20.74
test_iso.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Success
464.00
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Success
317.96
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Success
1051.89
test_internal_lb.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Success
699.84
test_internal_lb.py
test_dedicateGuestVlanRange
Success
10.20
test_guest_vlan_range.py
test_UpdateConfigParamWithScope
Success
0.10
test_global_settings.py
test_rolepermission_lifecycle_update
Success
5.83
test_dynamicroles.py
test_rolepermission_lifecycle_list
Success
5.72
test_dynamicroles.py
test_rolepermission_lifecycle_delete
Success
5.58
test_dynamicroles.py
test_rolepermission_lifecycle_create
Success
5.59
test_dynamicroles.py
test_rolepermission_lifecycle_concurrent_updates
Success
5.68
test_dynamicroles.py
test_role_lifecycle_update_role_inuse
Success
5.67
test_dynamicroles.py
test_role_lifecycle_update
Success
5.64
test_dynamicroles.py
test_role_lifecycle_list
Success
5.59
test_dynamicroles.py
test_role_lifecycle_delete
Success
5.66
test_dynamicroles.py
test_role_lifecycle_create
Success
5.64
test_dynamicroles.py
test_role_inuse_deletion
Success
5.58
test_dynamicroles.py
test_role_account_acls_multiple_mgmt_servers
Success
6.93
test_dynamicroles.py
test_role_account_acls
Success
7.17
test_dynamicroles.py
test_default_role_deletion
Success
5.69
test_dynamicroles.py
test_04_create_fat_type_disk_offering
Success
0.05
test_disk_offerings.py
test_03_delete_disk_offering
Success
0.08
test_disk_offerings.py
test_02_edit_disk_offering
Success
0.06
test_disk_offerings.py
test_02_create_sparse_type_disk_offering
Success
0.07
test_disk_offerings.py
test_01_create_disk_offering
Success
0.07
test_disk_offerings.py
test_deployvm_userdispersing
Success
30.49
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userconcentrated
Success
80.82
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_firstfit
Success
140.91
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userdata_post
Success
25.37
test_deploy_vm_with_userdata.py
test_deployvm_userdata
Success
171.13
test_deploy_vm_with_userdata.py
test_02_deploy_vm_root_resize
Success
5.56
test_deploy_vm_root_resize.py
test_01_deploy_vm_root_resize
Success
5.59
test_deploy_vm_root_resize.py
test_00_deploy_vm_root_resize
Success
5.68
test_deploy_vm_root_resize.py
test_deploy_vm_from_iso
Success
196.84
test_deploy_vm_iso.py
test_DeployVmAntiAffinityGroup
Success
151.11
test_affinity_groups.py
test_06_copy_template
Skipped
0.00
test_templates.py
test_static_role_account_acls
Skipped
0.02
test_staticroles.py
test_06_copy_iso
Skipped
0.00
test_iso.py
test_deploy_vgpu_enabled_vm
Skipped
0.01
test_deploy_vgpu_enabled_vm.py
Trillian test result (tid-1105)
Environment: vmware-55u3 (x2), Advanced Networking with Mgmt server 7
Total time taken: 53273 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1105-vmware-55u3.zip
Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py
Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Test completed. 47 look ok, 2 have error(s)
Test
Result
Time (s)
Test File
test_04_rvpc_privategw_static_routes
Failure
1087.47
test_privategw_acl.py
test_02_list_snapshots_with_removed_data_store
Error
131.60
test_snapshots.py
test_02_list_snapshots_with_removed_data_store
Error
136.75
test_snapshots.py
test_01_vpc_site2site_vpn
Success
578.92
test_vpc_vpn.py
test_01_vpc_remote_access_vpn
Success
272.61
test_vpc_vpn.py
test_01_redundant_vpc_site2site_vpn
Success
885.64
test_vpc_vpn.py
test_02_VPC_default_routes
Success
471.84
test_vpc_router_nics.py
test_01_VPC_nics_after_destroy
Success
947.31
test_vpc_router_nics.py
test_05_rvpc_multi_tiers
Success
896.47
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Success
1691.65
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Success
893.04
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Success
909.00
test_vpc_redundant.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Success
1605.68
test_vpc_redundant.py
test_09_delete_detached_volume
Success
36.03
test_volumes.py
test_06_download_detached_volume
Success
106.02
test_volumes.py
test_05_detach_volume
Success
105.34
test_volumes.py
test_04_delete_attached_volume
Success
15.24
test_volumes.py
test_03_download_attached_volume
Success
20.37
test_volumes.py
test_02_attach_volume
Success
63.82
test_volumes.py
test_01_create_volume
Success
525.89
test_volumes.py
test_03_delete_vm_snapshots
Success
280.29
test_vm_snapshots.py
test_02_revert_vm_snapshots
Success
237.50
test_vm_snapshots.py
test_01_create_vm_snapshots
Success
167.11
test_vm_snapshots.py
test_deploy_vm_multiple
Success
278.17
test_vm_life_cycle.py
test_deploy_vm
Success
0.04
test_vm_life_cycle.py
test_advZoneVirtualRouter
Success
0.03
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Success
26.85
test_vm_life_cycle.py
test_09_expunge_vm
Success
125.28
test_vm_life_cycle.py
test_08_migrate_vm
Success
91.31
test_vm_life_cycle.py
test_07_restore_vm
Success
0.11
test_vm_life_cycle.py
test_06_destroy_vm
Success
10.18
test_vm_life_cycle.py
test_03_reboot_vm
Success
5.15
test_vm_life_cycle.py
test_02_start_vm
Success
25.27
test_vm_life_cycle.py
test_01_stop_vm
Success
5.14
test_vm_life_cycle.py
test_CreateTemplateWithDuplicateName
Success
498.59
test_templates.py
test_08_list_system_templates
Success
0.03
test_templates.py
test_07_list_public_templates
Success
0.04
test_templates.py
test_05_template_permissions
Success
0.06
test_templates.py
test_04_extract_template
Success
15.24
test_templates.py
test_03_delete_template
Success
5.12
test_templates.py
test_02_edit_template
Success
90.29
test_templates.py
test_01_create_template
Success
247.00
test_templates.py
test_10_destroy_cpvm
Success
327.12
test_ssvm.py
test_09_destroy_ssvm
Success
268.97
test_ssvm.py
test_08_reboot_cpvm
Success
186.76
test_ssvm.py
test_07_reboot_ssvm
Success
158.61
test_ssvm.py
test_06_stop_cpvm
Success
207.03
test_ssvm.py
test_05_stop_ssvm
Success
208.88
test_ssvm.py
test_04_cpvm_internals
Success
1.25
test_ssvm.py
test_03_ssvm_internals
Success
3.62
test_ssvm.py
test_02_list_cpvm_vm
Success
0.15
test_ssvm.py
test_01_list_sec_storage_vm
Success
0.15
test_ssvm.py
test_01_snapshot_root_disk
Success
61.83
test_snapshots.py
test_04_change_offering_small
Success
91.92
test_service_offerings.py
test_03_delete_service_offering
Success
0.04
test_service_offerings.py
test_02_edit_service_offering
Success
0.09
test_service_offerings.py
test_01_create_service_offering
Success
0.12
test_service_offerings.py
test_02_sys_template_ready
Success
0.15
test_secondary_storage.py
test_01_sys_vm_start
Success
0.19
test_secondary_storage.py
test_09_reboot_router
Success
176.28
test_routers.py
test_08_start_router
Success
156.31
test_routers.py
test_07_stop_router
Success
25.32
test_routers.py
test_06_router_advanced
Success
0.07
test_routers.py
test_05_router_basic
Success
0.04
test_routers.py
test_04_restart_network_wo_cleanup
Success
5.75
test_routers.py
test_03_restart_network_cleanup
Success
151.27
test_routers.py
test_02_router_internal_adv
Success
1.06
test_routers.py
test_01_router_internal_basic
Success
0.62
test_routers.py
test_router_dns_guestipquery
Success
77.24
test_router_dns.py
test_router_dns_externalipquery
Success
0.09
test_router_dns.py
test_router_dhcphosts
Success
191.38
test_router_dhcphosts.py
test_router_dhcp_opts
Success
27.16
test_router_dhcphosts.py
test_01_updatevolumedetail
Success
0.13
test_resource_detail.py
test_01_reset_vm_on_reboot
Success
45.62
test_reset_vm_on_reboot.py
test_createRegion
Success
0.05
test_regions.py
test_create_pvlan_network
Success
5.24
test_pvlan.py
test_dedicatePublicIpRange
Success
0.55
test_public_ip_range.py
test_03_vpc_privategw_restart_vpc_cleanup
Success
1301.80
test_privategw_acl.py
test_02_vpc_privategw_static_routes
Success
799.16
test_privategw_acl.py
test_01_vpc_privategw_acl
Success
198.59
test_privategw_acl.py
test_01_primary_storage_nfs
Success
39.15
test_primary_storage.py
test_createPortablePublicIPRange
Success
15.26
test_portable_publicip.py
test_createPortablePublicIPAcquire
Success
15.52
test_portable_publicip.py
test_isolate_network_password_server
Success
99.69
test_password_server.py
test_UpdateStorageOverProvisioningFactor
Success
0.13
test_over_provisioning.py
test_oobm_zchange_password
Success
25.93
test_outofbandmanagement.py
test_oobm_multiple_mgmt_server_ownership
Success
16.37
test_outofbandmanagement.py
test_oobm_issue_power_status
Success
10.36
test_outofbandmanagement.py
test_oobm_issue_power_soft
Success
15.39
test_outofbandmanagement.py
test_oobm_issue_power_reset
Success
15.36
test_outofbandmanagement.py
test_oobm_issue_power_on
Success
15.37
test_outofbandmanagement.py
test_oobm_issue_power_off
Success
15.37
test_outofbandmanagement.py
test_oobm_issue_power_cycle
Success
10.34
test_outofbandmanagement.py
test_oobm_enabledisable_across_clusterzones
Success
92.82
test_outofbandmanagement.py
test_oobm_enable_feature_valid
Success
0.14
test_outofbandmanagement.py
test_oobm_enable_feature_invalid
Success
0.10
test_outofbandmanagement.py
test_oobm_disable_feature_valid
Success
5.20
test_outofbandmanagement.py
test_oobm_disable_feature_invalid
Success
0.11
test_outofbandmanagement.py
test_oobm_configure_invalid_driver
Success
0.09
test_outofbandmanagement.py
test_oobm_configure_default_driver
Success
0.11
test_outofbandmanagement.py
test_oobm_background_powerstate_sync
Success
23.43
test_outofbandmanagement.py
test_extendPhysicalNetworkVlan
Success
15.35
test_non_contigiousvlan.py
test_01_nic
Success
622.63
test_nic.py
test_releaseIP
Success
400.02
test_network.py
test_reboot_router
Success
710.96
test_network.py
test_public_ip_user_account
Success
10.26
test_network.py
test_public_ip_admin_account
Success
40.31
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Success
77.48
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Success
61.84
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Success
125.79
test_network.py
test_delete_account
Success
313.39
test_network.py
test_02_port_fwd_on_non_src_nat
Success
61.14
test_network.py
test_01_port_fwd_on_src_nat
Success
111.89
test_network.py
test_nic_secondaryip_add_remove
Success
227.95
test_multipleips_per_nic.py
test_list_zones_metrics
Success
0.36
test_metrics_api.py
test_list_volumes_metrics
Success
5.56
test_metrics_api.py
test_list_vms_metrics
Success
257.37
test_metrics_api.py
test_list_pstorage_metrics
Success
0.28
test_metrics_api.py
test_list_infrastructure_metrics
Success
0.42
test_metrics_api.py
test_list_hosts_metrics
Success
0.52
test_metrics_api.py
test_list_clusters_metrics
Success
0.39
test_metrics_api.py
login_test_saml_user
Success
19.42
test_login.py
test_assign_and_removal_lb
Success
150.29
test_loadbalance.py
test_02_create_lb_rule_non_nat
Success
207.97
test_loadbalance.py
test_01_create_lb_rule_src_nat
Success
208.85
test_loadbalance.py
test_03_list_snapshots
Success
0.10
test_list_ids_parameter.py
test_02_list_templates
Success
0.05
test_list_ids_parameter.py
test_01_list_volumes
Success
0.03
test_list_ids_parameter.py
test_07_list_default_iso
Success
0.06
test_iso.py
test_05_iso_permissions
Success
0.13
test_iso.py
test_04_extract_Iso
Success
5.18
test_iso.py
test_03_delete_iso
Success
95.14
test_iso.py
test_02_edit_iso
Success
0.06
test_iso.py
test_01_create_iso
Success
21.01
test_iso.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Success
702.24
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Success
460.02
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Success
1119.03
test_internal_lb.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Success
844.21
test_internal_lb.py
test_dedicateGuestVlanRange
Success
10.29
test_guest_vlan_range.py
test_UpdateConfigParamWithScope
Success
0.15
test_global_settings.py
test_rolepermission_lifecycle_update
Success
6.22
test_dynamicroles.py
test_rolepermission_lifecycle_list
Success
5.99
test_dynamicroles.py
test_rolepermission_lifecycle_delete
Success
5.88
test_dynamicroles.py
test_rolepermission_lifecycle_create
Success
5.90
test_dynamicroles.py
test_rolepermission_lifecycle_concurrent_updates
Success
6.02
test_dynamicroles.py
test_role_lifecycle_update_role_inuse
Success
5.91
test_dynamicroles.py
test_role_lifecycle_update
Success
11.00
test_dynamicroles.py
test_role_lifecycle_list
Success
6.09
test_dynamicroles.py
test_role_lifecycle_delete
Success
5.93
test_dynamicroles.py
test_role_lifecycle_create
Success
5.96
test_dynamicroles.py
test_role_inuse_deletion
Success
5.87
test_dynamicroles.py
test_role_account_acls_multiple_mgmt_servers
Success
8.14
test_dynamicroles.py
test_role_account_acls
Success
8.39
test_dynamicroles.py
test_default_role_deletion
Success
5.99
test_dynamicroles.py
test_04_create_fat_type_disk_offering
Success
0.07
test_disk_offerings.py
test_03_delete_disk_offering
Success
0.06
test_disk_offerings.py
test_02_edit_disk_offering
Success
0.05
test_disk_offerings.py
test_02_create_sparse_type_disk_offering
Success
0.08
test_disk_offerings.py
test_01_create_disk_offering
Success
0.10
test_disk_offerings.py
test_deployvm_userdispersing
Success
65.96
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userconcentrated
Success
136.56
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_firstfit
Success
221.86
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userdata_post
Success
50.73
test_deploy_vm_with_userdata.py
test_deployvm_userdata
Success
186.59
test_deploy_vm_with_userdata.py
test_02_deploy_vm_root_resize
Success
6.14
test_deploy_vm_root_resize.py
test_01_deploy_vm_root_resize
Success
5.84
test_deploy_vm_root_resize.py
test_00_deploy_vm_root_resize
Success
6.00
test_deploy_vm_root_resize.py
test_deploy_vm_from_iso
Success
252.86
test_deploy_vm_iso.py
test_DeployVmAntiAffinityGroup
Success
307.74
test_affinity_groups.py
test_08_resize_volume
Skipped
5.13
test_volumes.py
test_07_resize_fail
Skipped
10.30
test_volumes.py
test_06_copy_template
Skipped
0.00
test_templates.py
test_static_role_account_acls
Skipped
0.02
test_staticroles.py
test_01_scale_vm
Skipped
66.35
test_scale_vm.py
test_01_primary_storage_iscsi
Skipped
0.04
test_primary_storage.py
test_06_copy_iso
Skipped
0.00
test_iso.py
test_deploy_vgpu_enabled_vm
Skipped
0.01
test_deploy_vgpu_enabled_vm.py
VMware and KVM look okay. Some new errors in xensever, will re-kick to see if we reproduce them again;
@blueorangutan test centos7 xenserver-65sp1
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + xenserver-65sp1) has been kicked to run smoke tests
Trillian test result (tid-1114)
Environment: xenserver-65sp1 (x2), Advanced Networking with Mgmt server 7
Total time taken: 46141 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1114-xenserver-65sp1.zip
Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py
Intermitten failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Test completed. 46 look ok, 3 have error(s)
Test
Result
Time (s)
Test File
test_05_rvpc_multi_tiers
Failure
579.18
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Failure
1385.93
test_vpc_redundant.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Failure
585.29
test_vpc_redundant.py
test_04_rvpc_privategw_static_routes
Failure
833.77
test_privategw_acl.py
test_02_list_snapshots_with_removed_data_store
Error
89.31
test_snapshots.py
test_02_list_snapshots_with_removed_data_store
Error
94.40
test_snapshots.py
test_01_vpc_site2site_vpn
Success
341.00
test_vpc_vpn.py
test_01_vpc_remote_access_vpn
Success
166.68
test_vpc_vpn.py
test_01_redundant_vpc_site2site_vpn
Success
567.94
test_vpc_vpn.py
test_02_VPC_default_routes
Success
394.93
test_vpc_router_nics.py
test_01_VPC_nics_after_destroy
Success
703.29
test_vpc_router_nics.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Success
953.46
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Success
1101.33
test_vpc_redundant.py
test_09_delete_detached_volume
Success
15.64
test_volumes.py
test_08_resize_volume
Success
85.79
test_volumes.py
test_07_resize_fail
Success
100.88
test_volumes.py
test_06_download_detached_volume
Success
20.30
test_volumes.py
test_05_detach_volume
Success
100.26
test_volumes.py
test_04_delete_attached_volume
Success
10.23
test_volumes.py
test_03_download_attached_volume
Success
15.26
test_volumes.py
test_02_attach_volume
Success
10.77
test_volumes.py
test_01_create_volume
Success
392.58
test_volumes.py
test_03_delete_vm_snapshots
Success
280.25
test_vm_snapshots.py
test_02_revert_vm_snapshots
Success
181.32
test_vm_snapshots.py
test_01_create_vm_snapshots
Success
105.84
test_vm_snapshots.py
test_deploy_vm_multiple
Success
227.38
test_vm_life_cycle.py
test_deploy_vm
Success
0.03
test_vm_life_cycle.py
test_advZoneVirtualRouter
Success
0.02
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Success
41.88
test_vm_life_cycle.py
test_09_expunge_vm
Success
125.30
test_vm_life_cycle.py
test_08_migrate_vm
Success
81.20
test_vm_life_cycle.py
test_07_restore_vm
Success
0.10
test_vm_life_cycle.py
test_06_destroy_vm
Success
15.18
test_vm_life_cycle.py
test_03_reboot_vm
Success
20.21
test_vm_life_cycle.py
test_02_start_vm
Success
25.24
test_vm_life_cycle.py
test_01_stop_vm
Success
30.24
test_vm_life_cycle.py
test_CreateTemplateWithDuplicateName
Success
171.15
test_templates.py
test_08_list_system_templates
Success
0.03
test_templates.py
test_07_list_public_templates
Success
0.04
test_templates.py
test_05_template_permissions
Success
0.05
test_templates.py
test_04_extract_template
Success
5.13
test_templates.py
test_03_delete_template
Success
5.11
test_templates.py
test_02_edit_template
Success
90.13
test_templates.py
test_01_create_template
Success
95.67
test_templates.py
test_10_destroy_cpvm
Success
196.53
test_ssvm.py
test_09_destroy_ssvm
Success
198.78
test_ssvm.py
test_08_reboot_cpvm
Success
141.41
test_ssvm.py
test_07_reboot_ssvm
Success
143.74
test_ssvm.py
test_06_stop_cpvm
Success
191.61
test_ssvm.py
test_05_stop_ssvm
Success
163.93
test_ssvm.py
test_04_cpvm_internals
Success
0.91
test_ssvm.py
test_03_ssvm_internals
Success
3.55
test_ssvm.py
test_02_list_cpvm_vm
Success
0.12
test_ssvm.py
test_01_list_sec_storage_vm
Success
0.16
test_ssvm.py
test_01_snapshot_root_disk
Success
21.17
test_snapshots.py
test_04_change_offering_small
Success
119.12
test_service_offerings.py
test_03_delete_service_offering
Success
0.04
test_service_offerings.py
test_02_edit_service_offering
Success
0.05
test_service_offerings.py
test_01_create_service_offering
Success
0.11
test_service_offerings.py
test_02_sys_template_ready
Success
0.12
test_secondary_storage.py
test_01_sys_vm_start
Success
0.18
test_secondary_storage.py
test_01_scale_vm
Success
5.23
test_scale_vm.py
test_09_reboot_router
Success
65.48
test_routers.py
test_08_start_router
Success
45.36
test_routers.py
test_07_stop_router
Success
15.18
test_routers.py
test_06_router_advanced
Success
0.05
test_routers.py
test_05_router_basic
Success
0.04
test_routers.py
test_04_restart_network_wo_cleanup
Success
5.52
test_routers.py
test_03_restart_network_cleanup
Success
115.79
test_routers.py
test_02_router_internal_adv
Success
0.66
test_routers.py
test_01_router_internal_basic
Success
0.41
test_routers.py
test_router_dns_guestipquery
Success
48.86
test_router_dns.py
test_router_dns_externalipquery
Success
0.08
test_router_dns.py
test_router_dhcphosts
Success
107.49
test_router_dhcphosts.py
test_router_dhcp_opts
Success
31.22
test_router_dhcphosts.py
test_01_updatevolumedetail
Success
5.11
test_resource_detail.py
test_01_reset_vm_on_reboot
Success
45.42
test_reset_vm_on_reboot.py
test_createRegion
Success
0.04
test_regions.py
test_create_pvlan_network
Success
5.21
test_pvlan.py
test_dedicatePublicIpRange
Success
0.41
test_public_ip_range.py
test_03_vpc_privategw_restart_vpc_cleanup
Success
998.03
test_privategw_acl.py
test_02_vpc_privategw_static_routes
Success
732.25
test_privategw_acl.py
test_01_vpc_privategw_acl
Success
187.76
test_privategw_acl.py
test_01_primary_storage_nfs
Success
39.71
test_primary_storage.py
test_01_primary_storage_iscsi
Success
79.23
test_primary_storage.py
test_createPortablePublicIPRange
Success
15.20
test_portable_publicip.py
test_createPortablePublicIPAcquire
Success
15.46
test_portable_publicip.py
test_isolate_network_password_server
Success
36.32
test_password_server.py
test_UpdateStorageOverProvisioningFactor
Success
0.12
test_over_provisioning.py
test_oobm_zchange_password
Success
30.71
test_outofbandmanagement.py
test_oobm_multiple_mgmt_server_ownership
Success
11.29
test_outofbandmanagement.py
test_oobm_issue_power_status
Success
10.25
test_outofbandmanagement.py
test_oobm_issue_power_soft
Success
15.39
test_outofbandmanagement.py
test_oobm_issue_power_reset
Success
15.37
test_outofbandmanagement.py
test_oobm_issue_power_on
Success
15.34
test_outofbandmanagement.py
test_oobm_issue_power_off
Success
15.32
test_outofbandmanagement.py
test_oobm_issue_power_cycle
Success
15.56
test_outofbandmanagement.py
test_oobm_enabledisable_across_clusterzones
Success
92.69
test_outofbandmanagement.py
test_oobm_enable_feature_valid
Success
5.14
test_outofbandmanagement.py
test_oobm_enable_feature_invalid
Success
0.09
test_outofbandmanagement.py
test_oobm_disable_feature_valid
Success
5.17
test_outofbandmanagement.py
test_oobm_disable_feature_invalid
Success
5.17
test_outofbandmanagement.py
test_oobm_configure_invalid_driver
Success
0.08
test_outofbandmanagement.py
test_oobm_configure_default_driver
Success
0.07
test_outofbandmanagement.py
test_oobm_background_powerstate_sync
Success
23.41
test_outofbandmanagement.py
test_extendPhysicalNetworkVlan
Success
15.28
test_non_contigiousvlan.py
test_01_nic
Success
650.19
test_nic.py
test_releaseIP
Success
318.29
test_network.py
test_reboot_router
Success
589.53
test_network.py
test_public_ip_user_account
Success
10.29
test_network.py
test_public_ip_admin_account
Success
40.32
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Success
86.49
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Success
76.56
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Success
100.54
test_network.py
test_delete_account
Success
262.65
test_network.py
test_02_port_fwd_on_non_src_nat
Success
70.78
test_network.py
test_01_port_fwd_on_src_nat
Success
83.82
test_network.py
test_nic_secondaryip_add_remove
Success
252.73
test_multipleips_per_nic.py
test_list_zones_metrics
Success
0.32
test_metrics_api.py
test_list_volumes_metrics
Success
5.50
test_metrics_api.py
test_list_vms_metrics
Success
126.40
test_metrics_api.py
test_list_pstorage_metrics
Success
0.38
test_metrics_api.py
test_list_infrastructure_metrics
Success
0.51
test_metrics_api.py
test_list_hosts_metrics
Success
0.40
test_metrics_api.py
test_list_clusters_metrics
Success
0.36
test_metrics_api.py
login_test_saml_user
Success
19.15
test_login.py
test_assign_and_removal_lb
Success
148.83
test_loadbalance.py
test_02_create_lb_rule_non_nat
Success
207.47
test_loadbalance.py
test_01_create_lb_rule_src_nat
Success
208.09
test_loadbalance.py
test_03_list_snapshots
Success
0.05
test_list_ids_parameter.py
test_02_list_templates
Success
0.04
test_list_ids_parameter.py
test_01_list_volumes
Success
0.03
test_list_ids_parameter.py
test_07_list_default_iso
Success
0.08
test_iso.py
test_05_iso_permissions
Success
0.09
test_iso.py
test_04_extract_Iso
Success
5.14
test_iso.py
test_03_delete_iso
Success
95.21
test_iso.py
test_02_edit_iso
Success
0.07
test_iso.py
test_01_create_iso
Success
21.00
test_iso.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Success
494.53
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Success
324.00
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Success
1073.40
test_internal_lb.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Success
722.96
test_internal_lb.py
test_dedicateGuestVlanRange
Success
10.29
test_guest_vlan_range.py
test_UpdateConfigParamWithScope
Success
0.13
test_global_settings.py
test_rolepermission_lifecycle_update
Success
6.24
test_dynamicroles.py
test_rolepermission_lifecycle_list
Success
6.03
test_dynamicroles.py
test_rolepermission_lifecycle_delete
Success
5.93
test_dynamicroles.py
test_rolepermission_lifecycle_create
Success
6.14
test_dynamicroles.py
test_rolepermission_lifecycle_concurrent_updates
Success
6.08
test_dynamicroles.py
test_role_lifecycle_update_role_inuse
Success
5.97
test_dynamicroles.py
test_role_lifecycle_update
Success
10.97
test_dynamicroles.py
test_role_lifecycle_list
Success
5.88
test_dynamicroles.py
test_role_lifecycle_delete
Success
10.94
test_dynamicroles.py
test_role_lifecycle_create
Success
5.94
test_dynamicroles.py
test_role_inuse_deletion
Success
5.89
test_dynamicroles.py
test_role_account_acls_multiple_mgmt_servers
Success
8.30
test_dynamicroles.py
test_role_account_acls
Success
8.34
test_dynamicroles.py
test_default_role_deletion
Success
6.01
test_dynamicroles.py
test_04_create_fat_type_disk_offering
Success
0.07
test_disk_offerings.py
test_03_delete_disk_offering
Success
0.04
test_disk_offerings.py
test_02_edit_disk_offering
Success
0.05
test_disk_offerings.py
test_02_create_sparse_type_disk_offering
Success
0.06
test_disk_offerings.py
test_01_create_disk_offering
Success
0.08
test_disk_offerings.py
test_deployvm_userdispersing
Success
121.17
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userconcentrated
Success
50.81
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_firstfit
Success
221.68
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userdata_post
Success
20.48
test_deploy_vm_with_userdata.py
test_deployvm_userdata
Success
217.05
test_deploy_vm_with_userdata.py
test_02_deploy_vm_root_resize
Success
5.84
test_deploy_vm_root_resize.py
test_01_deploy_vm_root_resize
Success
5.83
test_deploy_vm_root_resize.py
test_00_deploy_vm_root_resize
Success
5.97
test_deploy_vm_root_resize.py
test_deploy_vm_from_iso
Success
333.13
test_deploy_vm_iso.py
test_DeployVmAntiAffinityGroup
Success
141.59
test_affinity_groups.py
test_06_copy_template
Skipped
0.00
test_templates.py
test_static_role_account_acls
Skipped
0.02
test_staticroles.py
test_06_copy_iso
Skipped
0.00
test_iso.py
test_deploy_vgpu_enabled_vm
Skipped
0.00
test_deploy_vgpu_enabled_vm.py
Test results LGTM, I don't see any new regressions. Some of the failing tests are known intermittent failures around rvpcs, snapshot/disk (snapshot failures/tests are fixed on master, backported to 4.9 here: #2052 ) etc.
Pinging - @DaanHoogland @abhinandanprateek @ustcweizhou - please review and see if this is okay?
/cc @karuturi
@rhtyd Please add some integration tests for hard shutdown scenario. Also consider the case of restoring a VM after hard shutdown.
The CS documentation should be updated to point to the relevant HV specific hard shutdown pros and cons.
PS: A similar thing was done in Cloudplatform but for the Destroy VM API. See if you would like to include that scenario as well in this PR or want it to be done separately.
@koushik-das I've added a stop_vm_forced test case in the vm lifecycle tests however it will difficult to validate a hard-shutdown scenarios. It's an operational improvement and I'm not sure what we will validate/verify when testing a case of starting a hard shutdown VM?
CloudStack api docs will automatically be generated and updated during the release/website-update. If needed, we can mention this change in the release notes under 'what's new/changed' etc. I see that the destroy VM codebase already uses a global setting to force shutdown the VM on destroy -- however, in ACS the global setting vm.destroy.forcestop is not honoured to force stop VMs while destroying them, if it is set to true. This changes fixes that.
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos6 ✔centos7 ✔debian. JID-743
@blueorangutan test centos7 vmware-55u3
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + vmware-55u3) has been kicked to run smoke tests
Trillian test result (tid-1128)
Environment: vmware-55u3 (x2), Advanced Networking with Mgmt server 7
Total time taken: 45568 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr2108-t1128-vmware-55u3.zip
Intermitten failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermitten failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermitten failure detected: /marvin/tests/smoke/test_snapshots.py
Test completed. 47 look ok, 2 have error(s)
Test
Result
Time (s)
Test File
test_04_rvpc_privategw_static_routes
Failure
1029.96
test_privategw_acl.py
test_02_list_snapshots_with_removed_data_store
Error
131.16
test_snapshots.py
test_02_list_snapshots_with_removed_data_store
Error
136.24
test_snapshots.py
test_01_vpc_site2site_vpn
Success
461.84
test_vpc_vpn.py
test_01_vpc_remote_access_vpn
Success
186.84
test_vpc_vpn.py
test_01_redundant_vpc_site2site_vpn
Success
673.22
test_vpc_vpn.py
test_02_VPC_default_routes
Success
571.00
test_vpc_router_nics.py
test_01_VPC_nics_after_destroy
Success
867.93
test_vpc_router_nics.py
test_05_rvpc_multi_tiers
Success
794.01
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Success
1665.02
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Success
867.67
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Success
785.84
test_vpc_redundant.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Success
1451.45
test_vpc_redundant.py
test_09_delete_detached_volume
Success
35.93
test_volumes.py
test_06_download_detached_volume
Success
95.77
test_volumes.py
test_05_detach_volume
Success
110.32
test_volumes.py
test_04_delete_attached_volume
Success
20.26
test_volumes.py
test_03_download_attached_volume
Success
25.36
test_volumes.py
test_02_attach_volume
Success
58.87
test_volumes.py
test_01_create_volume
Success
520.70
test_volumes.py
test_03_delete_vm_snapshots
Success
280.31
test_vm_snapshots.py
test_02_revert_vm_snapshots
Success
205.01
test_vm_snapshots.py
test_01_create_vm_snapshots
Success
161.77
test_vm_snapshots.py
test_deploy_vm_multiple
Success
302.90
test_vm_life_cycle.py
test_deploy_vm
Success
0.03
test_vm_life_cycle.py
test_advZoneVirtualRouter
Success
0.02
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Success
26.85
test_vm_life_cycle.py
test_09_expunge_vm
Success
125.23
test_vm_life_cycle.py
test_08_migrate_vm
Success
81.21
test_vm_life_cycle.py
test_07_restore_vm
Success
0.10
test_vm_life_cycle.py
test_06_destroy_vm
Success
10.15
test_vm_life_cycle.py
test_03_reboot_vm
Success
5.17
test_vm_life_cycle.py
test_02_start_vm
Success
25.26
test_vm_life_cycle.py
test_01_stop_vm_forced
Success
5.14
test_vm_life_cycle.py
test_01_stop_vm
Success
10.15
test_vm_life_cycle.py
test_CreateTemplateWithDuplicateName
Success
377.36
test_templates.py
test_08_list_system_templates
Success
0.03
test_templates.py
test_07_list_public_templates
Success
0.04
test_templates.py
test_05_template_permissions
Success
0.05
test_templates.py
test_04_extract_template
Success
25.38
test_templates.py
test_03_delete_template
Success
5.11
test_templates.py
test_02_edit_template
Success
90.14
test_templates.py
test_01_create_template
Success
196.28
test_templates.py
test_10_destroy_cpvm
Success
241.81
test_ssvm.py
test_09_destroy_ssvm
Success
268.86
test_ssvm.py
test_08_reboot_cpvm
Success
156.54
test_ssvm.py
test_07_reboot_ssvm
Success
158.51
test_ssvm.py
test_06_stop_cpvm
Success
206.94
test_ssvm.py
test_05_stop_ssvm
Success
178.76
test_ssvm.py
test_04_cpvm_internals
Success
1.17
test_ssvm.py
test_03_ssvm_internals
Success
3.91
test_ssvm.py
test_02_list_cpvm_vm
Success
0.11
test_ssvm.py
test_01_list_sec_storage_vm
Success
0.12
test_ssvm.py
test_01_snapshot_root_disk
Success
31.28
test_snapshots.py
test_04_change_offering_small
Success
101.96
test_service_offerings.py
test_03_delete_service_offering
Success
0.04
test_service_offerings.py
test_02_edit_service_offering
Success
0.08
test_service_offerings.py
test_01_create_service_offering
Success
0.19
test_service_offerings.py
test_02_sys_template_ready
Success
0.13
test_secondary_storage.py
test_01_sys_vm_start
Success
0.17
test_secondary_storage.py
test_09_reboot_router
Success
166.17
test_routers.py
test_08_start_router
Success
140.92
test_routers.py
test_07_stop_router
Success
25.24
test_routers.py
test_06_router_advanced
Success
0.05
test_routers.py
test_05_router_basic
Success
0.05
test_routers.py
test_04_restart_network_wo_cleanup
Success
5.67
test_routers.py
test_03_restart_network_cleanup
Success
171.16
test_routers.py
test_02_router_internal_adv
Success
1.02
test_routers.py
test_01_router_internal_basic
Success
0.57
test_routers.py
test_router_dns_guestipquery
Success
76.77
test_router_dns.py
test_router_dns_externalipquery
Success
0.08
test_router_dns.py
test_router_dhcphosts
Success
149.10
test_router_dhcphosts.py
test_router_dhcp_opts
Success
21.71
test_router_dhcphosts.py
test_01_updatevolumedetail
Success
0.08
test_resource_detail.py
test_01_reset_vm_on_reboot
Success
35.38
test_reset_vm_on_reboot.py
test_createRegion
Success
0.04
test_regions.py
test_create_pvlan_network
Success
5.25
test_pvlan.py
test_dedicatePublicIpRange
Success
0.47
test_public_ip_range.py
test_03_vpc_privategw_restart_vpc_cleanup
Success
1113.47
test_privategw_acl.py
test_02_vpc_privategw_static_routes
Success
742.65
test_privategw_acl.py
test_01_vpc_privategw_acl
Success
208.06
test_privategw_acl.py
test_01_primary_storage_nfs
Success
38.76
test_primary_storage.py
test_createPortablePublicIPRange
Success
15.21
test_portable_publicip.py
test_createPortablePublicIPAcquire
Success
15.44
test_portable_publicip.py
test_isolate_network_password_server
Success
94.41
test_password_server.py
test_UpdateStorageOverProvisioningFactor
Success
0.12
test_over_provisioning.py
test_oobm_zchange_password
Success
30.75
test_outofbandmanagement.py
test_oobm_multiple_mgmt_server_ownership
Success
16.35
test_outofbandmanagement.py
test_oobm_issue_power_status
Success
10.28
test_outofbandmanagement.py
test_oobm_issue_power_soft
Success
15.41
test_outofbandmanagement.py
test_oobm_issue_power_reset
Success
15.35
test_outofbandmanagement.py
test_oobm_issue_power_on
Success
15.44
test_outofbandmanagement.py
test_oobm_issue_power_off
Success
15.33
test_outofbandmanagement.py
test_oobm_issue_power_cycle
Success
15.36
test_outofbandmanagement.py
test_oobm_enabledisable_across_clusterzones
Success
87.60
test_outofbandmanagement.py
test_oobm_enable_feature_valid
Success
5.17
test_outofbandmanagement.py
test_oobm_enable_feature_invalid
Success
0.11
test_outofbandmanagement.py
test_oobm_disable_feature_valid
Success
5.18
test_outofbandmanagement.py
test_oobm_disable_feature_invalid
Success
0.10
test_outofbandmanagement.py
test_oobm_configure_invalid_driver
Success
0.07
test_outofbandmanagement.py
test_oobm_configure_default_driver
Success
0.08
test_outofbandmanagement.py
test_oobm_background_powerstate_sync
Success
23.40
test_outofbandmanagement.py
test_extendPhysicalNetworkVlan
Success
15.39
test_non_contigiousvlan.py
test_01_nic
Success
510.38
test_nic.py
test_releaseIP
Success
353.73
test_network.py
test_reboot_router
Success
715.72
test_network.py
test_public_ip_user_account
Success
10.28
test_network.py
test_public_ip_admin_account
Success
40.31
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Success
76.92
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Success
61.82
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Success
125.61
test_network.py
test_delete_account
Success
343.41
test_network.py
test_02_port_fwd_on_non_src_nat
Success
55.77
test_network.py
test_01_port_fwd_on_src_nat
Success
111.83
test_network.py
test_nic_secondaryip_add_remove
Success
222.67
test_multipleips_per_nic.py
test_list_zones_metrics
Success
0.27
test_metrics_api.py
test_list_volumes_metrics
Success
5.46
test_metrics_api.py
test_list_vms_metrics
Success
242.35
test_metrics_api.py
test_list_pstorage_metrics
Success
0.38
test_metrics_api.py
test_list_infrastructure_metrics
Success
0.51
test_metrics_api.py
test_list_hosts_metrics
Success
0.49
test_metrics_api.py
test_list_clusters_metrics
Success
0.34
test_metrics_api.py
login_test_saml_user
Success
19.42
test_login.py
test_assign_and_removal_lb
Success
148.57
test_loadbalance.py
test_02_create_lb_rule_non_nat
Success
207.26
test_loadbalance.py
test_01_create_lb_rule_src_nat
Success
208.51
test_loadbalance.py
test_03_list_snapshots
Success
0.09
test_list_ids_parameter.py
test_02_list_templates
Success
0.07
test_list_ids_parameter.py
test_01_list_volumes
Success
0.03
test_list_ids_parameter.py
test_07_list_default_iso
Success
0.06
test_iso.py
test_05_iso_permissions
Success
0.06
test_iso.py
test_04_extract_Iso
Success
5.18
test_iso.py
test_03_delete_iso
Success
95.16
test_iso.py
test_02_edit_iso
Success
0.06
test_iso.py
test_01_create_iso
Success
21.12
test_iso.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Success
608.01
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Success
404.84
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Success
1120.61
test_internal_lb.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Success
862.56
test_internal_lb.py
test_dedicateGuestVlanRange
Success
10.30
test_guest_vlan_range.py
test_UpdateConfigParamWithScope
Success
0.15
test_global_settings.py
test_rolepermission_lifecycle_update
Success
6.23
test_dynamicroles.py
test_rolepermission_lifecycle_list
Success
5.97
test_dynamicroles.py
test_rolepermission_lifecycle_delete
Success
5.86
test_dynamicroles.py
test_rolepermission_lifecycle_create
Success
5.91
test_dynamicroles.py
test_rolepermission_lifecycle_concurrent_updates
Success
6.04
test_dynamicroles.py
test_role_lifecycle_update_role_inuse
Success
5.94
test_dynamicroles.py
test_role_lifecycle_update
Success
11.00
test_dynamicroles.py
test_role_lifecycle_list
Success
5.94
test_dynamicroles.py
test_role_lifecycle_delete
Success
10.93
test_dynamicroles.py
test_role_lifecycle_create
Success
5.93
test_dynamicroles.py
test_role_inuse_deletion
Success
5.91
test_dynamicroles.py
test_role_account_acls_multiple_mgmt_servers
Success
8.35
test_dynamicroles.py
test_role_account_acls
Success
8.23
test_dynamicroles.py
test_default_role_deletion
Success
5.97
test_dynamicroles.py
test_04_create_fat_type_disk_offering
Success
0.07
test_disk_offerings.py
test_03_delete_disk_offering
Success
0.04
test_disk_offerings.py
test_02_edit_disk_offering
Success
0.05
test_disk_offerings.py
test_02_create_sparse_type_disk_offering
Success
0.08
test_disk_offerings.py
test_01_create_disk_offering
Success
0.10
test_disk_offerings.py
test_deployvm_userdispersing
Success
71.11
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userconcentrated
Success
126.25
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_firstfit
Success
292.28
test_deploy_vms_with_varied_deploymentplanners.py
test_deployvm_userdata_post
Success
35.56
test_deploy_vm_with_userdata.py
test_deployvm_userdata
Success
236.82
test_deploy_vm_with_userdata.py
test_02_deploy_vm_root_resize
Success
5.86
test_deploy_vm_root_resize.py
test_01_deploy_vm_root_resize
Success
5.87
test_deploy_vm_root_resize.py
test_00_deploy_vm_root_resize
Success
6.10
test_deploy_vm_root_resize.py
test_deploy_vm_from_iso
Success
222.52
test_deploy_vm_iso.py
test_DeployVmAntiAffinityGroup
Success
196.99
test_affinity_groups.py
test_08_resize_volume
Skipped
10.14
test_volumes.py
test_07_resize_fail
Skipped
15.32
test_volumes.py
test_06_copy_template
Skipped
0.00
test_templates.py
test_static_role_account_acls
Skipped
0.03
test_staticroles.py
test_01_scale_vm
Skipped
66.34
test_scale_vm.py
test_01_primary_storage_iscsi
Skipped
0.03
test_primary_storage.py
test_06_copy_iso
Skipped
0.00
test_iso.py
test_deploy_vgpu_enabled_vm
Skipped
0.00
test_deploy_vgpu_enabled_vm.py
Test LGTM. This is ready for merge, has enough test results, a new marvin test, reviews, LGTMs. /cc @karuturi
Thanks @rhtyd. Code changes LGTM
@karuturi can you consider reviewing/merging this? this has several test results and lgtms
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos6 ✔centos7 ✔debian. JID-757
@karuturi thanks for merging, however as @borisstoyanov has pointed out, there seems to be regression caused due to some refactoring changes around stopVM code for KVM. I'll send a separate FR swiftly.
@karuturi here's the reboot/regression fix: https://github.com/apache/cloudstack/pull/2135/files
@borisstoyanov please review/verify ^^
| gharchive/pull-request | 2017-05-16T09:30:05 | 2025-04-01T04:33:27.493428 | {
"authors": [
"blueorangutan",
"koushik-das",
"remibergsma",
"resmo",
"rhtyd",
"ustcweizhou",
"wido"
],
"repo": "apache/cloudstack",
"url": "https://github.com/apache/cloudstack/pull/2108",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
548935717 | Display capability info in listNetwork response
Description
Display some properties (eg RedundantRouter) in listNetworks response which are initializd in VirtualRouterElement.setCapabilities()
Types of changes
[ ] Breaking change (fix or feature that would cause existing functionality to change)
[ ] New feature (non-breaking change which adds functionality)
[ ] Bug fix (non-breaking change which fixes an issue)
[X] Enhancement (improves an existing feature and functionality)
[ ] Cleanup (Code refactoring and cleanup, that may add test cases)
Screenshots (if appropriate):
How Has This Been Tested?
listNetworks response for StaticNat service before the change
{
"name": "StaticNat"
},
listNetworks response for StaticNat after the change
{
"capability": [
{
"canchooseservicecapability": false,
"name": "ElasticIp",
"value": "false"
},
{
"canchooseservicecapability": false,
"name": "AssociatePublicIP",
"value": "true"
}
],
"name": "StaticNat",
"provider": [
{
"name": "VirtualRouter"
}
]
},
@ravening Moving to 4.16 based on base branch. Please update if needed
@ravening Moving to 4.16 based on base branch. Please update if needed
@ravening Moving to 4.16 based on base branch. Please update if needed
made it 4.15 (didn't try 4.14, making assumptions)
@ravening Moving to 4.16 based on base branch. Please update if needed
made it 4.15 (didn't try 4.14, making assumptions)
@blueorangutan package
@shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2781
@blueorangutan package
@shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2787
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2797
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2800
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2809
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 29
@blueorangutan package
@Pearl1594 a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 239
@Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
Trillian test result (tid-252)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 53122 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t252-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py
Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py
Smoke tests completed. 76 look OK, 10 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_DeployVmAntiAffinityGroup_in_project
Error
58.94
test_affinity_groups_projects.py
test_DeployVmAntiAffinityGroup
Error
36.39
test_affinity_groups.py
test_04_rvpc_privategw_static_routes
Error
0.44
test_privategw_acl.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
23.64
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
34.23
test_kubernetes_clusters.py
ContextSuite context=TestRAMCPUResourceAccounting>:setup
Error
0.00
test_resource_accounting.py
ContextSuite context=TestRouterDHCPHosts>:setup
Error
0.00
test_router_dhcphosts.py
ContextSuite context=TestRouterDHCPOpts>:setup
Error
0.00
test_router_dhcphosts.py
ContextSuite context=TestIsolatedNetworks>:setup
Error
0.00
test_routers_network_ops.py
ContextSuite context=TestRedundantIsolateNetworks>:setup
Error
0.00
test_routers_network_ops.py
ContextSuite context=TestCpuCapServiceOfferings>:setup
Error
0.00
test_service_offerings.py
ContextSuite context=TestServiceOfferings>:setup
Error
0.10
test_service_offerings.py
ContextSuite context=TestSnapshotRootDisk>:setup
Error
0.00
test_snapshots.py
test_hostha_kvm_host_degraded
Failure
677.99
test_hostha_kvm.py
test_hostha_kvm_host_fencing
Failure
644.32
test_hostha_kvm.py
test_hostha_kvm_host_recovering
Failure
648.34
test_hostha_kvm.py
@blueorangutan test
@Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
Trillian test result (tid-258)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 40933 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t258-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_virtio_scsi_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py
Smoke tests completed. 82 look OK, 4 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_DeployVmAntiAffinityGroup_in_project
Error
59.90
test_affinity_groups_projects.py
test_DeployVmAntiAffinityGroup
Error
35.36
test_affinity_groups.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
24.69
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
35.16
test_kubernetes_clusters.py
test_hostha_kvm_host_degraded
Failure
678.33
test_hostha_kvm.py
test_hostha_kvm_host_fencing
Failure
643.83
test_hostha_kvm.py
test_hostha_kvm_host_recovering
Failure
643.73
test_hostha_kvm.py
@blueorangutan package
@Pearl1594 a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 272
@blueorangutan test
@Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
Code LGTM. verified the output of listnetworks. Waiting for the test results.
Trillian test result (tid-280)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 46771 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t280-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_usage.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Smoke tests completed. 81 look OK, 5 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
85.98
test_kubernetes_clusters.py
test_01_isolate_network_FW_PF_default_routes_egress_true
Failure
65.59
test_routers_network_ops.py
test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false
Failure
121.82
test_routers_network_ops.py
ContextSuite context=TestISOUsage>:setup
Error
0.00
test_usage.py
test_01_migrate_VM_and_root_volume
Error
70.32
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
50.05
test_vm_life_cycle.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Failure
494.01
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Failure
481.57
test_vpc_redundant.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 321
@blueorangutan test
@Pearl1594 a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
Trillian test result (tid-347)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36633 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t347-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
69.95
test_kubernetes_clusters.py
test_01_migrate_VM_and_root_volume
Error
69.26
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
50.04
test_vm_life_cycle.py
@ravening @weizhouapache cc @Pearl1594 should this information be restricted to just admins or to all users?
Re-ping @weizhouapache @ravening
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 101
Trillian test result (tid-766)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32216 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t766-kvm-centos7.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ping @ravening - can you address outstanding remarks and advise if this is ready for merging? I see last rounds of tests have passeed, on the design side - do we need those capabilities as part of the list API response (even for non-admin users?)
re-kick travis
@ravening @weizhouapache cc @Pearl1594 should this information be restricted to just admins or to all users?
@rhtyd this can be displayed to both users and admins
@blueorangutan package
@DaanHoogland a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: el7 :heavy_check_mark: el8 :heavy_check_mark: debian. SL-JID 947
@blueorangutan test
@DaanHoogland a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
Trillian test result (tid-1735)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32476 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t1735-kvm-centos7.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Ping @Pearl1594 @DaanHoogland can you review/test this, would it cause any leaks or regression (for UI?)
Moving to 4.16, as this introduces changes in API response - we'll have more time to review/test this against UI/clients.
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_multiplication_x: el7 :heavy_check_mark: el8 :heavy_multiplication_x: debian :heavy_check_mark: suse15. SL-JID 1161
@blueorangutan package
@sureshanaparti a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_multiplication_x: el7 :heavy_check_mark: el8 :heavy_check_mark: debian :heavy_check_mark: suse15. SL-JID 1173
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: el7 :heavy_check_mark: el8 :heavy_check_mark: debian :heavy_check_mark: suse15. SL-JID 1238
@blueorangutan test
@sureshanaparti a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
@blueorangutan test
@rhtyd a Trillian-Jenkins test job (centos7 mgmt + kvm-centos7) has been kicked to run smoke tests
@blueorangutan test matrix
@sureshanaparti a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-2122)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36862 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t2122-kvm-centos7.zip
Smoke tests completed. 89 look OK, 0 have errors
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-2121)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 37333 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t2121-xenserver-71.zip
Smoke tests completed. 89 look OK, 0 have errors
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-2123)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 38750 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr3804-t2123-vmware-65u2.zip
Smoke tests completed. 89 look OK, 0 have errors
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Hi @Pearl1594 can you test this PR for any regressions in the UI given the change on the networks response?
I'll have a look at it @nvazquez
LGTM
| gharchive/pull-request | 2020-01-13T13:37:48 | 2025-04-01T04:33:27.572532 | {
"authors": [
"DaanHoogland",
"Pearl1594",
"blueorangutan",
"nvazquez",
"ravening",
"rhtyd",
"shwstppr",
"sureshanaparti"
],
"repo": "apache/cloudstack",
"url": "https://github.com/apache/cloudstack/pull/3804",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
811830319 | [HEALTH][4.15] Health Check Run
Dummy PR for tracking health on 4.15 branch
@blueorangutan package
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos7 ✔centos8 ✔debian. JID-2814
@blueorangutan test matrix
@DaanHoogland a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-3606)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 31484 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t3606-kvm-centos7.zip
Smoke tests completed. 86 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-3605)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32355 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t3605-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 85 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_scale_vm
Failure
9.28
test_scale_vm.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos7 ✔centos8 ✔debian. JID-2820
@blueorangutan package
@blueorangutan package
@shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2845
@blueorangutan package
@shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✔centos7 ✔centos8 ✔debian. JID-2856
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: ✖centos7 ✖centos8 ✖debian. JID-2883
[S] Trillian test result (tid-75)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 34945 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t75-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 85 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_migrate_VM_and_root_volume
Error
77.47
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
50.19
test_vm_life_cycle.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S]
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 86
[S] Trillian test result (tid-74)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 319816 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t74-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py
Intermittent failure detected: /marvin/tests/smoke/test_async_job.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_iso.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_root_resize.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vms_with_varied_deploymentplanners.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_with_userdata.py
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_domain_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_domain_vpc_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py
Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py
Intermittent failure detected: /marvin/tests/smoke/test_metrics_api.py
Intermittent failure detected: /marvin/tests/smoke/test_multipleips_per_nic.py
Intermittent failure detected: /marvin/tests/smoke/test_network_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_nic.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_portforwardingrules.py
Intermittent failure detected: /marvin/tests/smoke/test_primary_storage.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_projects.py
Intermittent failure detected: /marvin/tests/smoke/test_public_ip_range.py
Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py
Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_routers.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_secondary_storage.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py
Intermittent failure detected: /marvin/tests/smoke/test_storage_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_templates.py
Intermittent failure detected: /marvin/tests/smoke/test_usage.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_deployment_planner.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 38 look OK, 48 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S]
Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 104
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S]
Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_multiplication_x: debian. SL-JID 106
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S]
@blueorangutan package
@davidjumani a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S]
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress. [S]
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 194
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests [S]
[S] Trillian test result (tid-190)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 90748 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t190-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vgpu_enabled_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py
Intermittent failure detected: /marvin/tests/smoke/test_human_readable_logs.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py
Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py
Intermittent failure detected: /marvin/tests/smoke/test_network.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_routers.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Smoke tests completed. 60 look OK, 26 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_create_lb_rule_src_nat
Failure
309.69
test_loadbalance.py
test_02_create_lb_rule_non_nat
Failure
36.55
test_loadbalance.py
test_assign_and_removal_lb
Failure
36.54
test_loadbalance.py
ContextSuite context=TestDeployvGPUenabledVM>:setup
Error
0.00
test_deploy_vgpu_enabled_vm.py
test_07_deploy_vm_with_extraconfig_xenserver
Error
659.01
test_deploy_vm_extra_config_data.py
test_01_port_fwd_on_src_nat
Failure
607.31
test_network.py
test_02_port_fwd_on_non_src_nat
Failure
609.07
test_network.py
test_reboot_router
Failure
386.76
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Error
609.38
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Error
610.16
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Error
613.22
test_network.py
test_isolate_network_password_server
Failure
161.16
test_password_server.py
test_02_vpc_privategw_static_routes
Failure
817.02
test_privategw_acl.py
test_03_vpc_privategw_restart_vpc_cleanup
Failure
815.13
test_privategw_acl.py
test_04_rvpc_privategw_static_routes
Failure
919.01
test_privategw_acl.py
test_01_disableHumanReadableLogs
Error
602.72
test_human_readable_logs.py
test_02_enableHumanReadableLogs
Error
602.73
test_human_readable_logs.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Failure
307.91
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Failure
386.90
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Error
168.94
test_internal_lb.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Error
220.07
test_internal_lb.py
test_01_deploy_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
0.01
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_06_deploy_invalid_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_01_add_delete_kubernetes_supported_version
Error
0.01
test_kubernetes_supported_versions.py
test_router_dhcphosts
Failure
168.68
test_router_dhcphosts.py
ContextSuite context=TestRouterDHCPHosts>:teardown
Error
182.03
test_router_dhcphosts.py
test_router_dhcp_opts
Error
613.47
test_router_dhcphosts.py
test_router_dns_guestipquery
Failure
456.87
test_router_dns.py
test_router_dns_guestipquery
Failure
456.63
test_router_dnsservice.py
test_02_routervm_iptables_policies
Error
689.05
test_routers_iptables_default_policy.py
test_01_single_VPC_iptables_policies
Error
751.75
test_routers_iptables_default_policy.py
test_01_isolate_network_FW_PF_default_routes_egress_true
Failure
238.18
test_routers_network_ops.py
test_02_isolate_network_FW_PF_default_routes_egress_false
Failure
248.52
test_routers_network_ops.py
test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true
Failure
277.54
test_routers_network_ops.py
test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false
Failure
276.85
test_routers_network_ops.py
test_03_RVR_Network_check_router_state
Error
720.81
test_routers_network_ops.py
test_01_router_internal_basic
Error
602.86
test_routers.py
test_02_router_internal_adv
Error
602.78
test_routers.py
test_04_restart_network_wo_cleanup
Error
604.98
test_routers.py
test_01_scale_vm
Failure
10.24
test_scale_vm.py
test_04_change_offering_small
Failure
711.70
test_service_offerings.py
test_01_snapshot_root_disk
Error
611.99
test_snapshots.py
test_03_ssvm_internals
Error
602.84
test_ssvm.py
test_04_cpvm_internals
Error
602.79
test_ssvm.py
test_05_stop_ssvm
Error
676.77
test_ssvm.py
test_06_stop_cpvm
Error
657.54
test_ssvm.py
test_07_reboot_ssvm
Error
727.29
test_ssvm.py
test_08_reboot_cpvm
Error
634.19
test_ssvm.py
test_09_destroy_ssvm
Error
683.58
test_ssvm.py
test_10_destroy_cpvm
Error
657.51
test_ssvm.py
test_10_attachAndDetach_iso
Failure
609.62
test_vm_life_cycle.py
test_change_service_offering_for_vm_with_snapshots
Failure
695.43
test_vm_snapshots.py
test_01_create_vm_snapshots
Failure
602.39
test_vm_snapshots.py
test_02_revert_vm_snapshots
Failure
601.87
test_vm_snapshots.py
test_03_delete_vm_snapshots
Failure
0.02
test_vm_snapshots.py
test_01_create_volume
Failure
608.39
test_volumes.py
test_02_attach_volume
Failure
618.85
test_volumes.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Error
942.30
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Error
937.70
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Error
793.40
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Error
739.49
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
847.81
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
885.91
test_vpc_redundant.py
test_01_VPC_nics_after_destroy
Failure
783.44
test_vpc_router_nics.py
test_02_VPC_default_routes
Failure
770.89
test_vpc_router_nics.py
test_01_redundant_vpc_site2site_vpn
Failure
431.44
test_vpc_vpn.py
test_01_vpc_site2site_vpn_multiple_options
Error
282.70
test_vpc_vpn.py
test_01_vpc_site2site_vpn
Error
305.28
test_vpc_vpn.py
[S] Trillian test result (tid-192)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 96345 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t192-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_human_readable_logs.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py
Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py
Intermittent failure detected: /marvin/tests/smoke/test_nested_virtualization.py
Intermittent failure detected: /marvin/tests/smoke/test_network.py
Intermittent failure detected: /marvin/tests/smoke/test_nic.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_routers.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 59 look OK, 27 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_create_lb_rule_src_nat
Failure
308.54
test_loadbalance.py
test_02_create_lb_rule_non_nat
Failure
34.65
test_loadbalance.py
test_assign_and_removal_lb
Failure
34.66
test_loadbalance.py
test_01_port_fwd_on_src_nat
Failure
605.88
test_network.py
test_02_port_fwd_on_non_src_nat
Failure
607.81
test_network.py
test_reboot_router
Failure
400.82
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Error
607.58
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Error
610.18
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Error
612.17
test_network.py
test_05_deploy_vm_with_extraconfig_vmware
Error
687.45
test_deploy_vm_extra_config_data.py
test_01_nic
Error
744.28
test_nic.py
test_isolate_network_password_server
Failure
157.36
test_password_server.py
test_02_vpc_privategw_static_routes
Failure
818.75
test_privategw_acl.py
test_03_vpc_privategw_restart_vpc_cleanup
Failure
816.90
test_privategw_acl.py
test_04_rvpc_privategw_static_routes
Failure
941.68
test_privategw_acl.py
test_01_disableHumanReadableLogs
Error
602.73
test_human_readable_logs.py
test_02_enableHumanReadableLogs
Error
602.72
test_human_readable_logs.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Failure
325.26
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Failure
385.99
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Error
173.62
test_internal_lb.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Error
229.63
test_internal_lb.py
test_01_deploy_kubernetes_cluster
Failure
0.01
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
0.01
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
0.01
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_06_deploy_invalid_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_01_add_delete_kubernetes_supported_version
Error
0.01
test_kubernetes_supported_versions.py
test_nested_virtualization_vmware
Error
756.30
test_nested_virtualization.py
test_router_dhcphosts
Failure
157.19
test_router_dhcphosts.py
ContextSuite context=TestRouterDHCPHosts>:teardown
Error
169.51
test_router_dhcphosts.py
test_router_dhcp_opts
Error
609.25
test_router_dhcphosts.py
test_router_dns_guestipquery
Failure
456.44
test_router_dns.py
test_router_dns_guestipquery
Failure
455.30
test_router_dnsservice.py
test_02_routervm_iptables_policies
Error
685.08
test_routers_iptables_default_policy.py
test_01_single_VPC_iptables_policies
Error
761.96
test_routers_iptables_default_policy.py
test_01_isolate_network_FW_PF_default_routes_egress_true
Failure
234.08
test_routers_network_ops.py
test_02_isolate_network_FW_PF_default_routes_egress_false
Failure
229.31
test_routers_network_ops.py
test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true
Failure
285.89
test_routers_network_ops.py
test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false
Failure
278.31
test_routers_network_ops.py
test_03_RVR_Network_check_router_state
Error
734.63
test_routers_network_ops.py
test_01_router_internal_basic
Error
602.85
test_routers.py
test_02_router_internal_adv
Error
602.86
test_routers.py
test_04_restart_network_wo_cleanup
Error
605.01
test_routers.py
test_01_scale_vm
Error
604.53
test_scale_vm.py
test_04_change_offering_small
Failure
719.49
test_service_offerings.py
test_01_snapshot_root_disk
Error
650.81
test_snapshots.py
test_03_ssvm_internals
Error
602.85
test_ssvm.py
test_04_cpvm_internals
Error
602.91
test_ssvm.py
test_05_stop_ssvm
Error
675.52
test_ssvm.py
test_06_stop_cpvm
Error
673.38
test_ssvm.py
test_07_reboot_ssvm
Error
703.26
test_ssvm.py
test_08_reboot_cpvm
Error
611.04
test_ssvm.py
test_09_destroy_ssvm
Error
705.68
test_ssvm.py
test_10_destroy_cpvm
Error
666.71
test_ssvm.py
ContextSuite context=TestVAppsVM>:setup
Error
41.18
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Failure
609.21
test_vm_life_cycle.py
test_change_service_offering_for_vm_with_snapshots
Failure
698.23
test_vm_snapshots.py
test_01_create_vm_snapshots
Failure
604.54
test_vm_snapshots.py
test_02_revert_vm_snapshots
Failure
601.39
test_vm_snapshots.py
test_03_delete_vm_snapshots
Failure
0.01
test_vm_snapshots.py
test_01_create_volume
Failure
609.12
test_volumes.py
test_02_attach_volume
Failure
606.96
test_volumes.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Error
905.83
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Error
905.08
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Error
791.44
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Error
751.93
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
869.92
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
905.03
test_vpc_redundant.py
test_01_VPC_nics_after_destroy
Failure
769.04
test_vpc_router_nics.py
test_02_VPC_default_routes
Failure
766.56
test_vpc_router_nics.py
test_01_redundant_vpc_site2site_vpn
Failure
437.55
test_vpc_vpn.py
test_01_vpc_site2site_vpn_multiple_options
Error
298.69
test_vpc_vpn.py
test_01_vpc_site2site_vpn
Error
311.23
test_vpc_vpn.py
[S] Trillian test result (tid-191)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 101140 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t191-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_deploy_virtio_scsi_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_extra_config_data.py
Intermittent failure detected: /marvin/tests/smoke/test_human_readable_logs.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py
Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py
Intermittent failure detected: /marvin/tests/smoke/test_network.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_routers.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_ssvm.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py
Smoke tests completed. 60 look OK, 26 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_create_lb_rule_src_nat
Failure
307.78
test_loadbalance.py
test_02_create_lb_rule_non_nat
Failure
35.59
test_loadbalance.py
test_assign_and_removal_lb
Failure
34.60
test_loadbalance.py
test_01_verify_libvirt
Error
602.64
test_deploy_virtio_scsi_vm.py
test_02_verify_libvirt_after_restart
Error
609.98
test_deploy_virtio_scsi_vm.py
test_03_verify_libvirt_attach_disk
Error
605.85
test_deploy_virtio_scsi_vm.py
test_04_verify_guest_lspci
Error
602.08
test_deploy_virtio_scsi_vm.py
test_05_change_vm_ostype_restart
Error
610.07
test_deploy_virtio_scsi_vm.py
test_06_verify_guest_lspci_again
Error
602.06
test_deploy_virtio_scsi_vm.py
ContextSuite context=TestAddConfigtoDeployVM>:setup
Error
0.00
test_deploy_vm_extra_config_data.py
test_01_port_fwd_on_src_nat
Failure
605.08
test_network.py
test_02_port_fwd_on_non_src_nat
Failure
607.91
test_network.py
ContextSuite context=TestPrivateVlansL2Networks>:setup
Error
1230.16
test_network.py
test_reboot_router
Failure
358.50
test_network.py
test_network_rules_acquired_public_ip_1_static_nat_rule
Error
607.44
test_network.py
test_network_rules_acquired_public_ip_2_nat_rule
Error
609.15
test_network.py
test_network_rules_acquired_public_ip_3_Load_Balancer_Rule
Error
613.23
test_network.py
test_isolate_network_password_server
Failure
157.18
test_password_server.py
test_02_vpc_privategw_static_routes
Failure
828.76
test_privategw_acl.py
test_03_vpc_privategw_restart_vpc_cleanup
Failure
754.69
test_privategw_acl.py
test_04_rvpc_privategw_static_routes
Failure
851.45
test_privategw_acl.py
test_01_disableHumanReadableLogs
Error
602.87
test_human_readable_logs.py
test_02_enableHumanReadableLogs
Error
602.61
test_human_readable_logs.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Failure
268.37
test_internal_lb.py
test_02_internallb_roundrobin_1RVPC_3VM_HTTP_port80
Failure
317.03
test_internal_lb.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Error
143.52
test_internal_lb.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Error
172.93
test_internal_lb.py
test_01_deploy_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_06_deploy_invalid_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_01_add_delete_kubernetes_supported_version
Error
0.01
test_kubernetes_supported_versions.py
test_router_dhcphosts
Failure
157.09
test_router_dhcphosts.py
ContextSuite context=TestRouterDHCPHosts>:teardown
Error
168.40
test_router_dhcphosts.py
test_router_dhcp_opts
Error
609.10
test_router_dhcphosts.py
test_router_dns_guestipquery
Failure
454.70
test_router_dns.py
test_router_dns_guestipquery
Failure
454.62
test_router_dnsservice.py
test_02_routervm_iptables_policies
Error
660.60
test_routers_iptables_default_policy.py
test_01_single_VPC_iptables_policies
Error
720.34
test_routers_iptables_default_policy.py
test_01_isolate_network_FW_PF_default_routes_egress_true
Failure
209.90
test_routers_network_ops.py
test_02_isolate_network_FW_PF_default_routes_egress_false
Failure
210.91
test_routers_network_ops.py
test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true
Failure
243.33
test_routers_network_ops.py
test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false
Failure
243.11
test_routers_network_ops.py
test_03_RVR_Network_check_router_state
Error
689.27
test_routers_network_ops.py
test_01_router_internal_basic
Error
602.80
test_routers.py
test_02_router_internal_adv
Error
602.75
test_routers.py
test_04_restart_network_wo_cleanup
Error
604.99
test_routers.py
test_01_service_offering_cpu_limit_use
Error
100.50
test_service_offerings.py
test_04_change_offering_small
Failure
715.21
test_service_offerings.py
test_01_snapshot_root_disk
Error
606.04
test_snapshots.py
test_03_ssvm_internals
Error
602.83
test_ssvm.py
test_04_cpvm_internals
Error
602.81
test_ssvm.py
test_05_stop_ssvm
Error
676.59
test_ssvm.py
test_06_stop_cpvm
Error
652.34
test_ssvm.py
test_07_reboot_ssvm
Error
720.59
test_ssvm.py
test_08_reboot_cpvm
Error
630.16
test_ssvm.py
test_09_destroy_ssvm
Error
672.43
test_ssvm.py
test_10_destroy_cpvm
Error
653.40
test_ssvm.py
test_01_migrate_VM_and_root_volume
Error
72.69
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
54.24
test_vm_life_cycle.py
test_01_secure_vm_migration
Error
703.25
test_vm_life_cycle.py
test_02_unsecure_vm_migration
Error
637.16
test_vm_life_cycle.py
test_03_secured_to_nonsecured_vm_migration
Error
637.97
test_vm_life_cycle.py
test_04_nonsecured_to_secured_vm_migration
Error
636.14
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Failure
608.16
test_vm_life_cycle.py
test_01_create_vm_snapshots
Failure
604.51
test_vm_snapshots.py
test_02_revert_vm_snapshots
Failure
601.21
test_vm_snapshots.py
test_03_delete_vm_snapshots
Failure
0.01
test_vm_snapshots.py
test_01_create_volume
Failure
606.75
test_volumes.py
test_02_attach_volume
Failure
605.85
test_volumes.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Error
825.80
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Error
824.98
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Error
743.03
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Error
709.98
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
801.08
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
827.08
test_vpc_redundant.py
test_01_VPC_nics_after_destroy
Failure
752.13
test_vpc_router_nics.py
test_02_VPC_default_routes
Failure
724.23
test_vpc_router_nics.py
test_01_redundant_vpc_site2site_vpn
Failure
345.91
test_vpc_vpn.py
test_01_vpc_site2site_vpn_multiple_options
Error
249.38
test_vpc_vpn.py
test_01_vpc_site2site_vpn
Error
286.15
test_vpc_vpn.py
test_hostha_enable_ha_when_host_disabled
Error
0.35
test_hostha_kvm.py
test_hostha_enable_ha_when_host_disconected
Error
605.03
test_hostha_kvm.py
test_hostha_enable_ha_when_host_in_maintenance
Error
303.11
test_hostha_kvm.py
test_hostha_kvm_host_fencing
Error
7.53
test_hostha_kvm.py
Trillian test result (tid-202)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 66060 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t202-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups_projects.py
Intermittent failure detected: /marvin/tests/smoke/test_affinity_groups.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vms_with_varied_deploymentplanners.py
Intermittent failure detected: /marvin/tests/smoke/test_deploy_vm_with_userdata.py
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_iso.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_supported_versions.py
Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py
Intermittent failure detected: /marvin/tests/smoke/test_metrics_api.py
Intermittent failure detected: /marvin/tests/smoke/test_multipleips_per_nic.py
Intermittent failure detected: /marvin/tests/smoke/test_nic_adapter_type.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_portforwardingrules.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_projects.py
Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py
Smoke tests completed. 71 look OK, 15 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_DeployVmAntiAffinityGroup_in_project
Error
58.76
test_affinity_groups_projects.py
test_DeployVmAntiAffinityGroup
Error
34.34
test_affinity_groups.py
test_list_clusters_metrics
Error
1527.61
test_metrics_api.py
test_list_vms_metrics
Error
0.26
test_metrics_api.py
ContextSuite context=TestDeployVmWithVariedPlanners>:setup
Error
0.00
test_deploy_vms_with_varied_deploymentplanners.py
ContextSuite context=TestDeployVmWithUserData>:setup
Error
0.00
test_deploy_vm_with_userdata.py
ContextSuite context=TestRemoteDiagnostics>:setup
Error
0.00
test_diagnostics.py
ContextSuite context=TestAdapterTypeForNic>:setup
Error
0.00
test_nic_adapter_type.py
test_03_vpc_internallb_haproxy_stats_on_all_interfaces
Error
197.81
test_internal_lb.py
test_04_rvpc_internallb_haproxy_stats_on_all_interfaces
Error
240.88
test_internal_lb.py
test_01_create_iso_with_checksum_sha1
Error
65.33
test_iso.py
test_02_create_iso_with_checksum_sha256
Error
65.37
test_iso.py
test_03_create_iso_with_checksum_md5
Error
65.30
test_iso.py
test_04_create_iso_with_no_checksum
Error
65.37
test_iso.py
test_01_create_iso
Failure
1512.00
test_iso.py
ContextSuite context=TestISO>:setup
Error
3025.08
test_iso.py
test_01_deploy_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.00
test_kubernetes_clusters.py
test_06_deploy_invalid_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.00
test_kubernetes_clusters.py
test_01_add_delete_kubernetes_supported_version
Error
60.58
test_kubernetes_supported_versions.py
ContextSuite context=TestListIdsParams>:setup
Error
0.00
test_list_ids_parameter.py
ContextSuite context=TestPortForwardingRules>:setup
Error
0.00
test_portforwardingrules.py
ContextSuite context=TestPrivateGwACL>:setup
Error
0.00
test_privategw_acl.py
test_hostha_kvm_host_degraded
Failure
690.01
test_hostha_kvm.py
test_hostha_kvm_host_fencing
Failure
644.69
test_hostha_kvm.py
test_hostha_kvm_host_recovering
Failure
648.20
test_hostha_kvm.py
@blueorangutan test matrix
@davidjumani a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian Build Failed (tid-213)
Trillian Build Failed (tid-215)
Trillian Build Failed (tid-214)
@blueorangutan test matrix
@davidjumani a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian Build Failed (tid-216)
Trillian Build Failed (tid-219)
Trillian test result (tid-217)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 23343 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t217-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_secondary_storage.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py
Smoke tests completed. 76 look OK, 10 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_isolate_network_FW_PF_default_routes_egress_true
Failure
131.08
test_routers_network_ops.py
test_02_isolate_network_FW_PF_default_routes_egress_false
Failure
109.27
test_routers_network_ops.py
test_01_RVR_Network_FW_PF_SSH_default_routes_egress_true
Failure
204.51
test_routers_network_ops.py
test_02_RVR_Network_FW_PF_SSH_default_routes_egress_false
Failure
211.38
test_routers_network_ops.py
test_01_sys_vm_start
Failure
0.04
test_secondary_storage.py
ContextSuite context=TestDeployVM>:setup
Error
0.00
test_vm_life_cycle.py
test_01_migrate_VM_and_root_volume
Error
1.19
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
1.13
test_vm_life_cycle.py
test_01_secure_vm_migration
Error
0.06
test_vm_life_cycle.py
test_02_unsecure_vm_migration
Error
1.07
test_vm_life_cycle.py
test_03_secured_to_nonsecured_vm_migration
Error
0.06
test_vm_life_cycle.py
test_04_nonsecured_to_secured_vm_migration
Error
1.08
test_vm_life_cycle.py
ContextSuite context=TestVMLifeCycle>:setup
Error
1.63
test_vm_life_cycle.py
ContextSuite context=TestVmSnapshot>:setup
Error
1.60
test_vm_snapshots.py
ContextSuite context=TestCreateVolume>:setup
Error
0.00
test_volumes.py
ContextSuite context=TestVolumes>:setup
Error
0.00
test_volumes.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Error
6.65
test_vpc_redundant.py
test_02_redundant_VPC_default_routes
Error
8.69
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Error
7.66
test_vpc_redundant.py
test_04_rvpc_network_garbage_collector_nics
Error
7.78
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
7.73
test_vpc_redundant.py
test_01_VPC_nics_after_destroy
Error
4.65
test_vpc_router_nics.py
test_02_VPC_default_routes
Error
5.66
test_vpc_router_nics.py
test_01_redundant_vpc_site2site_vpn
Failure
7.31
test_vpc_vpn.py
test_01_vpc_site2site_vpn_multiple_options
Failure
5.26
test_vpc_vpn.py
test_01_vpc_remote_access_vpn
Failure
3.15
test_vpc_vpn.py
test_01_vpc_site2site_vpn
Failure
5.27
test_vpc_vpn.py
test_01_cancel_host_maintenace_with_no_migration_jobs
Error
0.04
test_host_maintenance.py
test_02_cancel_host_maintenace_with_migration_jobs
Error
0.04
test_host_maintenance.py
test_03_cancel_host_maintenace_with_migration_jobs_failure
Error
0.04
test_host_maintenance.py
test_01_cancel_host_maintenance_ssh_enabled_agent_connected
Error
0.01
test_host_maintenance.py
test_03_cancel_host_maintenance_ssh_disabled_agent_connected
Error
0.01
test_host_maintenance.py
test_04_cancel_host_maintenance_ssh_disabled_agent_disconnected
Error
0.01
test_host_maintenance.py
test_hostha_enable_ha_when_host_in_maintenance
Error
303.52
test_hostha_kvm.py
Trillian test result (tid-220)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36258 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t220-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 85 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_scale_vm
Failure
11.27
test_scale_vm.py
Trillian test result (tid-218)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 51077 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t218-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 82 look OK, 4 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_10_traceroute_in_vr
Failure
61.13
test_diagnostics.py
test_01_internallb_roundrobin_1VPC_3VM_HTTP_port80
Failure
406.03
test_internal_lb.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
800.50
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
3615.34
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
87.96
test_kubernetes_clusters.py
ContextSuite context=TestVAppsVM>:setup
Error
46.39
test_vm_life_cycle.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_multiplication_x: debian. SL-JID 223
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_multiplication_x: debian. SL-JID 229
@blueorangutan package
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 288
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 284
@blueorangutan test matrix
cc @davidjumani
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-310)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 37814 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t310-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
80.36
test_kubernetes_clusters.py
test_01_migrate_VM_and_root_volume
Error
75.35
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
50.13
test_vm_life_cycle.py
Trillian test result (tid-311)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 38159 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t311-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 83 look OK, 3 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_10_traceroute_in_vr
Failure
61.19
test_diagnostics.py
test_01_deploy_kubernetes_cluster
Failure
118.00
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
108.44
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
65.53
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
77.75
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
84.81
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
66.41
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
82.90
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
71.83
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
112.23
test_kubernetes_clusters.py
ContextSuite context=TestVAppsVM>:setup
Error
45.53
test_vm_life_cycle.py
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 310
Trillian test result (tid-332)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 35328 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t332-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
70.07
test_kubernetes_clusters.py
test_01_migrate_VM_and_root_volume
Error
68.19
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
51.08
test_vm_life_cycle.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 316
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-339)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36700 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t339-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Smoke tests completed. 83 look OK, 3 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
89.11
test_kubernetes_clusters.py
test_01_scale_vm
Failure
10.29
test_scale_vm.py
test_01_cancel_host_maintenace_with_no_migration_jobs
Failure
306.06
test_host_maintenance.py
test_02_cancel_host_maintenace_with_migration_jobs
Failure
306.08
test_host_maintenance.py
test_03_cancel_host_maintenace_with_migration_jobs_failure
Failure
0.19
test_host_maintenance.py
Trillian test result (tid-341)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 37255 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t341-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_deploy_kubernetes_cluster
Failure
89.25
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
95.24
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
58.45
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
68.63
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
62.51
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
72.65
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
95.15
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
56.43
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
99.09
test_kubernetes_clusters.py
ContextSuite context=TestVAppsVM>:setup
Error
58.95
test_vm_life_cycle.py
Trillian test result (tid-340)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 38569 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t340-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
78.23
test_kubernetes_clusters.py
test_01_migrate_VM_and_root_volume
Error
67.26
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
49.09
test_vm_life_cycle.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 327
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-355)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 35150 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t355-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_deploy_kubernetes_cluster
Failure
106.86
test_kubernetes_clusters.py
test_02_invalid_upgrade_kubernetes_cluster
Failure
101.29
test_kubernetes_clusters.py
test_03_deploy_and_upgrade_kubernetes_cluster
Failure
59.63
test_kubernetes_clusters.py
test_04_deploy_and_scale_kubernetes_cluster
Failure
54.75
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
52.15
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
59.38
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
55.18
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
51.22
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
90.47
test_kubernetes_clusters.py
ContextSuite context=TestVAppsVM>:setup
Error
44.56
test_vm_life_cycle.py
Trillian test result (tid-354)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36025 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t354-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_password_server.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
69.78
test_kubernetes_clusters.py
test_01_migrate_VM_and_root_volume
Error
69.29
test_vm_life_cycle.py
test_02_migrate_VM_with_two_data_disks
Error
51.05
test_vm_life_cycle.py
test_10_attachAndDetach_iso
Error
12.42
test_vm_life_cycle.py
Trillian test result (tid-353)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 42021 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t353-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 84 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestKubernetesCluster>:teardown
Error
86.04
test_kubernetes_clusters.py
test_01_scale_vm
Failure
12.30
test_scale_vm.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 380
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-402)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32668 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t402-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_scale_vm
Failure
10.27
test_scale_vm.py
Trillian test result (tid-403)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 34375 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t403-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-404)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36173 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t404-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_diagnostics.py
Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Smoke tests completed. 85 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_10_traceroute_in_vr
Failure
61.13
test_diagnostics.py
ContextSuite context=TestVAppsVM>:setup
Error
79.67
test_vm_life_cycle.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 388
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 390
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 401
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-422)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 36092 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t422-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_scale_vm
Failure
9.24
test_scale_vm.py
Trillian test result (tid-423)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 35809 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t423-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-424)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 35981 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t424-vmware-65u2.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 420
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-459)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32999 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t459-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_scale_vm
Failure
11.26
test_scale_vm.py
Trillian test result (tid-461)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 35111 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t461-vmware-65u2.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-460)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 48079 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t460-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_iso.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_loadbalance.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_03_deploy_and_scale_kubernetes_cluster
Failure
3641.88
test_kubernetes_clusters.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 442
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-496)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 33717 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t496-kvm-centos7.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-495)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 42563 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t495-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_multipleips_per_nic.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Smoke tests completed. 85 look OK, 2 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_invalid_upgrade_kubernetes_cluster
Failure
3603.00
test_kubernetes_clusters.py
test_02_deploy_and_upgrade_kubernetes_cluster
Failure
3616.86
test_kubernetes_clusters.py
test_03_deploy_and_scale_kubernetes_cluster
Failure
0.07
test_kubernetes_clusters.py
test_04_basic_lifecycle_kubernetes_cluster
Failure
0.04
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.04
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
112.63
test_kubernetes_clusters.py
test_01_scale_vm
Failure
9.23
test_scale_vm.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 457
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-522)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 41013 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t522-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_list_ids_parameter.py
Intermittent failure detected: /marvin/tests/smoke/test_scale_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_scale_vm
Failure
10.25
test_scale_vm.py
Trillian test result (tid-524)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 41907 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t524-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_snapshots.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_07_deploy_kubernetes_ha_cluster
Failure
3613.91
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.06
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
51.22
test_kubernetes_clusters.py
Trillian test result (tid-523)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 44200 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t523-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_07_deploy_kubernetes_ha_cluster
Failure
3612.37
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
80.94
test_kubernetes_clusters.py
@blueorangutan ui
@davidjumani a Jenkins job has been kicked to build UI QA env. I'll keep you posted as I make progress.
UI build: :heavy_check_mark:
Live QA URL: http://qa.cloudstack.cloud:8080/client/pr/4710 (SL-JID-109)
Trillian test result (tid-582)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32934 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t582-xenserver-71.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-583)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 32915 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t583-kvm-centos7.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_multiplication_x: centos7 :heavy_multiplication_x: centos8 :heavy_check_mark: debian. SL-JID 516
@blueorangutan package
@shwstppr a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 518
@blueorangutan test matrix
@rhtyd a Trillian-Jenkins matrix job (centos7 mgmt + xs71, centos7 mgmt + vmware65, centos7 mgmt + kvmcentos7) has been kicked to run smoke tests
Trillian test result (tid-605)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 33567 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t605-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Intermittent failure detected: /marvin/tests/smoke/test_host_maintenance.py
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-607)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 35381 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t607-vmware-65u2.zip
Smoke tests completed. 87 look OK, 0 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
Trillian test result (tid-606)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 54867 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t606-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_storage_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_templates.py
Intermittent failure detected: /marvin/tests/smoke/test_usage.py
Intermittent failure detected: /marvin/tests/smoke/test_vm_life_cycle.py
Intermittent failure detected: /marvin/tests/smoke/test_volumes.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_vpn.py
Intermittent failure detected: /marvin/tests/smoke/test_hostha_kvm.py
Smoke tests completed. 78 look OK, 9 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
ContextSuite context=TestVMWareStoragePolicies>:setup
Error
0.00
test_storage_policy.py
test_02_create_template_with_checksum_sha1
Error
65.37
test_templates.py
test_03_create_template_with_checksum_sha256
Error
65.36
test_templates.py
test_04_create_template_with_checksum_md5
Error
65.39
test_templates.py
test_05_create_template_with_no_checksum
Error
65.37
test_templates.py
test_04_extract_template
Failure
128.33
test_templates.py
ContextSuite context=TestISOUsage>:setup
Error
0.00
test_usage.py
test_01_volume_usage
Failure
787.32
test_usage.py
test_10_attachAndDetach_iso
Failure
1511.07
test_vm_life_cycle.py
test_06_download_detached_volume
Failure
312.60
test_volumes.py
ContextSuite context=TestVPCRedundancy>:setup
Error
0.00
test_vpc_redundant.py
ContextSuite context=TestVPCNics>:setup
Error
0.00
test_vpc_router_nics.py
ContextSuite context=TestRVPCSite2SiteVpn>:setup
Error
0.00
test_vpc_vpn.py
ContextSuite context=TestVPCSite2SiteVPNMultipleOptions>:setup
Error
0.00
test_vpc_vpn.py
ContextSuite context=TestVpcRemoteAccessVpn>:setup
Error
0.00
test_vpc_vpn.py
ContextSuite context=TestVpcSite2SiteVpn>:setup
Error
0.00
test_vpc_vpn.py
test_disable_oobm_ha_state_ineligible
Error
1511.93
test_hostha_kvm.py
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 524
@blueorangutan package
@rhtyd a Jenkins job has been kicked to build packages. I'll keep you posted as I make progress.
Packaging result: :heavy_check_mark: centos7 :heavy_check_mark: centos8 :heavy_check_mark: debian. SL-JID 531
Trillian test result (tid-666)
Environment: xenserver-71 (x2), Advanced Networking with Mgmt server 7
Total time taken: 51950 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t666-xenserver-71.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_usage.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Smoke tests completed. 86 look OK, 1 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_01_invalid_upgrade_kubernetes_cluster
Failure
3526.47
test_kubernetes_clusters.py
test_02_deploy_and_upgrade_kubernetes_cluster
Failure
3612.57
test_kubernetes_clusters.py
test_03_deploy_and_scale_kubernetes_cluster
Failure
0.05
test_kubernetes_clusters.py
test_04_basic_lifecycle_kubernetes_cluster
Failure
0.05
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.04
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
38.06
test_kubernetes_clusters.py
Trillian test result (tid-667)
Environment: kvm-centos7 (x2), Advanced Networking with Mgmt server 7
Total time taken: 83767 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t667-kvm-centos7.zip
Intermittent failure detected: /marvin/tests/smoke/test_deploy_virtio_scsi_vm.py
Intermittent failure detected: /marvin/tests/smoke/test_internal_lb.py
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_network.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py
Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_routers.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_router_nics.py
Smoke tests completed. 75 look OK, 12 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_04_verify_guest_lspci
Error
810.60
test_deploy_virtio_scsi_vm.py
test_06_verify_guest_lspci_again
Error
810.57
test_deploy_virtio_scsi_vm.py
test_01_invalid_upgrade_kubernetes_cluster
Failure
3610.37
test_kubernetes_clusters.py
test_03_deploy_and_scale_kubernetes_cluster
Failure
36.75
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
43.86
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.04
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
84.64
test_kubernetes_clusters.py
ContextSuite context=TestResetVmOnReboot>:setup
Error
0.00
test_reset_vm_on_reboot.py
ContextSuite context=TestRAMCPUResourceAccounting>:setup
Error
0.00
test_resource_accounting.py
ContextSuite context=TestRouterDHCPHosts>:setup
Error
0.00
test_router_dhcphosts.py
ContextSuite context=TestRouterDHCPOpts>:setup
Error
0.00
test_router_dhcphosts.py
ContextSuite context=TestRouterDns>:setup
Error
0.00
test_router_dns.py
ContextSuite context=TestRouterDnsService>:setup
Error
0.00
test_router_dnsservice.py
ContextSuite context=TestRouterIpTablesPolicies>:setup
Error
0.00
test_routers_iptables_default_policy.py
ContextSuite context=TestVPCIpTablesPolicies>:setup
Error
0.00
test_routers_iptables_default_policy.py
ContextSuite context=TestIsolatedNetworks>:setup
Error
0.00
test_routers_network_ops.py
ContextSuite context=TestRedundantIsolateNetworks>:setup
Error
0.00
test_routers_network_ops.py
ContextSuite context=TestRouterServices>:setup
Error
0.00
test_routers.py
ContextSuite context=TestCpuCapServiceOfferings>:setup
Error
0.00
test_service_offerings.py
ContextSuite context=TestServiceOfferings>:setup
Error
0.16
test_service_offerings.py
ContextSuite context=TestSnapshotRootDisk>:setup
Error
0.00
test_snapshots.py
Trillian test result (tid-668)
Environment: vmware-65u2 (x2), Advanced Networking with Mgmt server 7
Total time taken: 85457 seconds
Marvin logs: https://github.com/blueorangutan/acs-prs/releases/download/trillian/pr4710-t668-vmware-65u2.zip
Intermittent failure detected: /marvin/tests/smoke/test_kubernetes_clusters.py
Intermittent failure detected: /marvin/tests/smoke/test_privategw_acl.py
Intermittent failure detected: /marvin/tests/smoke/test_reset_vm_on_reboot.py
Intermittent failure detected: /marvin/tests/smoke/test_resource_accounting.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dhcphosts.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dns.py
Intermittent failure detected: /marvin/tests/smoke/test_router_dnsservice.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_iptables_default_policy.py
Intermittent failure detected: /marvin/tests/smoke/test_routers_network_ops.py
Intermittent failure detected: /marvin/tests/smoke/test_routers.py
Intermittent failure detected: /marvin/tests/smoke/test_service_offerings.py
Intermittent failure detected: /marvin/tests/smoke/test_snapshots.py
Intermittent failure detected: /marvin/tests/smoke/test_vpc_redundant.py
Smoke tests completed. 74 look OK, 13 have error(s)
Only failed tests results shown below:
Test
Result
Time (s)
Test File
test_03_vpc_privategw_restart_vpc_cleanup
Failure
1046.65
test_privategw_acl.py
test_01_invalid_upgrade_kubernetes_cluster
Failure
3608.66
test_kubernetes_clusters.py
test_02_deploy_and_upgrade_kubernetes_cluster
Failure
3609.05
test_kubernetes_clusters.py
test_03_deploy_and_scale_kubernetes_cluster
Failure
0.05
test_kubernetes_clusters.py
test_04_basic_lifecycle_kubernetes_cluster
Failure
0.05
test_kubernetes_clusters.py
test_05_delete_kubernetes_cluster
Failure
0.05
test_kubernetes_clusters.py
test_07_deploy_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_08_deploy_and_upgrade_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
test_09_delete_kubernetes_ha_cluster
Failure
0.05
test_kubernetes_clusters.py
ContextSuite context=TestKubernetesCluster>:teardown
Error
51.06
test_kubernetes_clusters.py
ContextSuite context=TestResetVmOnReboot>:setup
Error
0.00
test_reset_vm_on_reboot.py
ContextSuite context=TestRAMCPUResourceAccounting>:setup
Error
0.00
test_resource_accounting.py
ContextSuite context=TestRouterDHCPHosts>:setup
Error
0.00
test_router_dhcphosts.py
ContextSuite context=TestRouterDHCPOpts>:setup
Error
0.00
test_router_dhcphosts.py
ContextSuite context=TestRouterDns>:setup
Error
0.00
test_router_dns.py
ContextSuite context=TestRouterDnsService>:setup
Error
0.00
test_router_dnsservice.py
ContextSuite context=TestRouterIpTablesPolicies>:setup
Error
0.00
test_routers_iptables_default_policy.py
ContextSuite context=TestVPCIpTablesPolicies>:setup
Error
0.00
test_routers_iptables_default_policy.py
ContextSuite context=TestIsolatedNetworks>:setup
Error
0.00
test_routers_network_ops.py
ContextSuite context=TestRedundantIsolateNetworks>:setup
Error
0.00
test_routers_network_ops.py
ContextSuite context=TestRouterServices>:setup
Error
0.00
test_routers.py
ContextSuite context=TestServiceOfferings>:setup
Error
1511.59
test_service_offerings.py
ContextSuite context=TestSnapshotRootDisk>:setup
Error
0.00
test_snapshots.py
test_01_create_redundant_VPC_2tiers_4VMs_4IPs_4PF_ACL
Failure
581.90
test_vpc_redundant.py
test_03_create_redundant_VPC_1tier_2VMs_2IPs_2PF_ACL_reboot_routers
Failure
650.51
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Failure
526.25
test_vpc_redundant.py
test_05_rvpc_multi_tiers
Error
562.43
test_vpc_redundant.py
| gharchive/pull-request | 2021-02-19T08:51:57 | 2025-04-01T04:33:28.045494 | {
"authors": [
"DaanHoogland",
"blueorangutan",
"davidjumani",
"rhtyd",
"shwstppr"
],
"repo": "apache/cloudstack",
"url": "https://github.com/apache/cloudstack/pull/4710",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
107130098 | Bug-ID: CLOUDSTACK-8870: Skip external device usage collection if no external devices exist
external network device usage monitor thread that runs every 5mins by default (based on global config external.network.stats.interval) and runs coalesce query to acquire a lock. When there are no external devices exist, there is no need to run usage collection.
Added test case to verify that usage collection task is not run when there are no External LB or External FW
Ping @remibergsma @borisroman
@kishankavala added a unit test to cover his change. I went through the code and haven't found any issue.
The if block if(_hostDao.listByType(Host.Type.ExternalFirewall).isEmpty() && _hostDao.listByType(Host.Type.ExternalLoadBalancer).isEmpty()){ added to ExternalDeviceUsageManagerImpl make sense and are covered.
Please give it a go with the integration tests we have.
Cheers,
Wilder
I have gone through the code. patch LGTM.
I need one more LGTM. I will add this to my CI queue. Thx...
Although I had pointed a typo in a comment line (nothing serious), the code LGTM.
Thank you @GabrielBrascher. @kishankavala would you mind fixing that typo? I will try to get this run though CI soon...
@kishankavala please rebase against latest master and push -f, update on status of your PR
I've left a NPE check, we should merge it once you fix that. Thanks.
tag:easypr
cc @swill
@kishankavala please rebase this PR as we have merge conflicts. Thanks...
@swill rebased with latest master
@GabrielBrascher fixed the typo
LGTM
CI RESULTS
Tests Run: 85
Skipped: 0
Failed: 2
Errors: 1
Duration: 10h 41m 55s
Summary of the problem(s):
ERROR: Test to verify access to loadbalancer haproxy admin stats page
----------------------------------------------------------------------
Traceback (most recent call last):
File "/data/git/cs1/cloudstack/test/integration/smoke/test_internal_lb.py", line 854, in tearDown
raise Exception("Cleanup failed with %s" % e)
Exception: Cleanup failed with Job failed: {jobprocstatus : 0, created : u'2016-05-24T12:24:25+0200', jobresult : {errorcode : 530, errortext : u'Failed to delete network'}, cmd : u'org.apache.cloudstack.api.command.user.network.DeleteNetworkCmd', userid : u'a5538db6-2168-11e6-932f-5254001daa61', jobstatus : 2, jobid : u'8e7ebb93-fb36-4842-8e2d-5cefea2ff80a', jobresultcode : 530, jobresulttype : u'object', jobinstancetype : u'Network', accountid : u'a5537331-2168-11e6-932f-5254001daa61'}
----------------------------------------------------------------------
Additional details in: /tmp/MarvinLogs/test_network_F00C21/results.txt
FAIL: test_02_vpc_privategw_static_routes (integration.smoke.test_privategw_acl.TestPrivateGwACL)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/data/git/cs1/cloudstack/test/integration/smoke/test_privategw_acl.py", line 262, in test_02_vpc_privategw_static_routes
self.performVPCTests(vpc_off)
File "/data/git/cs1/cloudstack/test/integration/smoke/test_privategw_acl.py", line 325, in performVPCTests
privateGw_1 = self.createPvtGw(vpc_1, "10.0.3.100", "10.0.3.101", acl1.id, vlan_1)
File "/data/git/cs1/cloudstack/test/integration/smoke/test_privategw_acl.py", line 595, in createPvtGw
self.fail("Failed to create Private Gateway ==> %s" % e)
AssertionError: Failed to create Private Gateway ==> Execute cmd: createprivategateway failed, due to: errorCode: 431, errorText:Network with vlan vlan://100 already exists in zone 1
----------------------------------------------------------------------
Additional details in: /tmp/MarvinLogs/test_network_F00C21/results.txt
FAIL: Test destroy(expunge) Virtual Machine
----------------------------------------------------------------------
Traceback (most recent call last):
File "/data/git/cs1/cloudstack/test/integration/smoke/test_vm_life_cycle.py", line 646, in test_09_expunge_vm
self.assertEqual(list_vm_response,None,"Check Expunged virtual machine is in listVirtualMachines response")
AssertionError: Check Expunged virtual machine is in listVirtualMachines response
----------------------------------------------------------------------
Additional details in: /tmp/MarvinLogs/test_vpc_routers_J2J0PP/results.txt
Associated Uploads
/tmp/MarvinLogs/DeployDataCenter__May_24_2016_06_36_18_ATGDYK:
dc_entries.obj
failed_plus_exceptions.txt
runinfo.txt
/tmp/MarvinLogs/test_network_F00C21:
failed_plus_exceptions.txt
results.txt
runinfo.txt
/tmp/MarvinLogs/test_vpc_routers_J2J0PP:
failed_plus_exceptions.txt
results.txt
runinfo.txt
Uploads will be available until 2016-07-25 02:00:00 +0200 CEST
Comment created by upr comment.
CI RESULTS
Tests Run: 82
Skipped: 0
Failed: 0
Errors: 3
Duration: 8h 33m 42s
Summary of the problem(s):
ERROR: test suite for <class 'integration.smoke.test_vpc_vpn.TestRVPCSite2SiteVpn'>
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 209, in run
self.setUp()
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 292, in setUp
self.setupContext(ancestor)
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 315, in setupContext
try_run(context, names)
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/util.py", line 471, in try_run
return func()
File "/data/git/cs1/cloudstack/test/integration/smoke/test_vpc_vpn.py", line 835, in setUpClass
cls.template.download(cls.apiclient)
File "/usr/lib/python2.7/site-packages/marvin/lib/base.py", line 1350, in download
elif 'Downloaded' in template.status:
TypeError: argument of type 'NoneType' is not iterable
----------------------------------------------------------------------
Additional details in: /tmp/MarvinLogs/test_network_VKMDK0/results.txt
ERROR: test suite for <class 'integration.smoke.test_vpc_vpn.TestVpcRemoteAccessVpn'>
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 209, in run
self.setUp()
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 292, in setUp
self.setupContext(ancestor)
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 315, in setupContext
try_run(context, names)
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/util.py", line 471, in try_run
return func()
File "/data/git/cs1/cloudstack/test/integration/smoke/test_vpc_vpn.py", line 293, in setUpClass
cls.template.download(cls.apiclient)
File "/usr/lib/python2.7/site-packages/marvin/lib/base.py", line 1350, in download
elif 'Downloaded' in template.status:
TypeError: argument of type 'NoneType' is not iterable
----------------------------------------------------------------------
Additional details in: /tmp/MarvinLogs/test_network_VKMDK0/results.txt
ERROR: test suite for <class 'integration.smoke.test_vpc_vpn.TestVpcSite2SiteVpn'>
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 209, in run
self.setUp()
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 292, in setUp
self.setupContext(ancestor)
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/suite.py", line 315, in setupContext
try_run(context, names)
File "/usr/lib/python2.7/site-packages/nose-1.3.7-py2.7.egg/nose/util.py", line 471, in try_run
return func()
File "/data/git/cs1/cloudstack/test/integration/smoke/test_vpc_vpn.py", line 472, in setUpClass
cls.template.download(cls.apiclient)
File "/usr/lib/python2.7/site-packages/marvin/lib/base.py", line 1350, in download
elif 'Downloaded' in template.status:
TypeError: argument of type 'NoneType' is not iterable
----------------------------------------------------------------------
Additional details in: /tmp/MarvinLogs/test_network_VKMDK0/results.txt
Associated Uploads
/tmp/MarvinLogs/DeployDataCenter__May_25_2016_19_10_57_8DI3FD:
dc_entries.obj
failed_plus_exceptions.txt
runinfo.txt
/tmp/MarvinLogs/test_network_VKMDK0:
failed_plus_exceptions.txt
results.txt
runinfo.txt
/tmp/MarvinLogs/test_vpc_routers_PXO5U1:
failed_plus_exceptions.txt
results.txt
runinfo.txt
Uploads will be available until 2016-07-26 02:00:00 +0200 CEST
Comment created by upr comment.
This is a cleaner run because this issue is a known issue. I think this one is ready now...
code/patch LGTM.
| gharchive/pull-request | 2015-09-18T04:14:17 | 2025-04-01T04:33:28.069207 | {
"authors": [
"GabrielBrascher",
"kishankavala",
"rajesh-battala",
"rhtyd",
"swill",
"wilderrodrigues"
],
"repo": "apache/cloudstack",
"url": "https://github.com/apache/cloudstack/pull/846",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
416567771 | TEXT-152 - new tests and fix for infinite loop
Hi, I have added a test that show shows the infinite loop, and the code that fixes it.
In order to make the test pass, and not only make the infinite loop stop happening, more changes were needed because the code for TEXT-111 did not fix the replacement of 0 length matches when we have multiple matches.
Coverage decreased (-0.02%) to 97.885% when pulling ad07efa10b4091ef3b9d1a9ee9ec0190726e8c55 on CAPS50:TEXT-152 into d880f686338f63b8e0860ad37da190fda792eecf on apache:master.
@CAPS50,
May you please update changes.xml?
Thank you,
Gary
@garydgregory I think the contributor goes under due-to. But sometimes new contributors add it as dev. And that way, it's tricky to know which value to put in dev. So it might be easier for a committer to add it when merging? At least that was my workflow so far
Thank you.
@kinow shall I change the changes.xml ? and if so should be the value of dev and due-to?
@kinow shall I change the changes.xml ? and if so should be the value of dev and due-to?
IMO, you can leave dev empty and put your name in due-to (or you GitHub ID, Jira ID, whatever you prefer).
thanks, I did that.
Thanks @CAPS50 ! Merging now
| gharchive/pull-request | 2019-03-03T21:33:49 | 2025-04-01T04:33:28.075322 | {
"authors": [
"CAPS50",
"coveralls",
"garydgregory",
"kinow"
],
"repo": "apache/commons-text",
"url": "https://github.com/apache/commons-text/pull/107",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
543673347 | Update IndexedDB platform support
General
This proposal supersedes and closes #1011 (PR #1011)
/cc @jgaehring with thanks for the original proposal in #1011
Platforms affected
Android, browser, Electron, iOS, "osx" (macOS)
Motivation and Context
From a second look at PR #1011 I discovered that the documentation of IndexedDB support is outdated for multiple platforms. I would like to get the documentation updated for all affected platforms at once.
Description
Update the description to make it clear that this functionality is supported by the underlying WebView on all platforms, with possible limitations added for the (web) browser platform
The "4.4 and above" condition for Android now gone, since we have already dropped support for Android pre-4.4
"Not supported on iOS" item now completely and unconditionally gone from disadvantages section due to this comment: https://github.com/apache/cordova-docs/pull/1011#issuecomment-567250310
Testing
Visual inspection
Checklist
[ ] [TODO] I've run the tests to see all new and existing tests pass
I added automated test coverage as appropriate for this change
Commit is prefixed with (platform) if this change only applies to one platform (e.g. (android))
[x] If this Pull Request resolves an issue another PR, I linked to the issue PR in the text above (and used the correct keyword to close issues using keywords)
[x] I've updated the documentation if as necessary
FOLLOWUP TODO:
[ ] If merged, we may need to do a manual rebuild (see #1035)
Is 5 MB limitation correct for indexedDB at iOS, Android and browser? I'm new to this topic, but I thought indexedDB was invented to store more. Is it some Cordova plugin's additional limitation?
Good catch, no 5MB is not accurate anymore.
For recent iOS/Safari version AFAIK there are no real sources and Android's limitation are here: https://stackoverflow.com/a/55483549/1902598 (the most up to date source, I've found).
Chrome doesn't really give a limit in a traditional sense, but it tries to not consume all of the disk space. It does this by determining how much remaining space is available and will start implementing quotas if remaining space gets low.
Chrome will either try to keep at least 10% of the total storage or 2GB, whichever is lower, of the disk free, and will start implementing quotas once it reaches this threshold.
The table below comes directly from the chromium sources, and is accurately as of Chrome 76.
// The amount of the device's storage the browser attempts to
--
| // keep free. If there is less than this amount of storage free
| // on the device, Chrome will grant 0 quota to origins.
| //
| // Prior to M66, this was 10% of total storage instead of a fixed value on
| // all devices. Now the minimum of a fixed value (2GB) and 10% is used to
| // limit the reserve on devices with plenty of storage, but scale down for
| // devices with extremely limited storage.
| // * 1TB storage -- min(100GB,2GB) = 2GB
| // * 500GB storage -- min(50GB,2GB) = 2GB
| // * 64GB storage -- min(6GB,2GB) = 2GB
| // * 16GB storage -- min(1.6GB,2GB) = 1.6GB
| // * 8GB storage -- min(800MB,2GB) = 800MB
https://chromium.googlesource.com/chromium/src.git/+/refs/tags/76.0.3775.5/storage/browser/quota/quota_settings.cc
Not sure about iOS/Safari.
@dfahlander - sorry for the mention, but do you happen to know recent iOS and/or Safari quotas?
It's worth noting that the behaviour in web browsers can also differ from the behaviour in web views.
@timbru31 I'm sorry but I don't know the quotas for iOS/Safari.
@dfahlander @brodybits @timbru31 did you have more infos about indexedDb allowed memory size on iOS, and also on Android ? Thanks a lot !
@aarbel
On android using the chrome browser, you can refer to this table https://github.com/apache/cordova-docs/pull/1042#issuecomment-611551394
It isn't clear on the quotas for wkwebview for ios. It isn't documented and afaik wkwebview isn't open sourced.
No, Safari is, at least to me, a blackbox. Last person to ask/tag is @beidson from Apple: Do you know any current size limits of the IndexedDB in iOS?
After PR #1043 is merged in, this PR should be rebased.
The only other minor feedback I have is that we do not need to break up the lines.
E.g.
IndexedDB is supported by the underlying WebView on all platforms,
with known limitations on the following platforms:
This could be written on one line.
IndexedDB is supported by the underlying WebView on all platforms, with known limitations on the following platforms:
IMO, I look at markdown as a text document. When I write a technical specification document in Word, I usual do not do something like this.
But this is just my opinion and not a blocker.
Brody, could you rebase this PR so we can merge this in?
Travis did not like the rebase and rapid commits. Once https://travis-ci.org/github/apache/cordova-docs/builds/741877609 finishes, I'll merge this in.
Thanks Tim for taking on the edit and merging!
| gharchive/pull-request | 2019-12-30T01:56:43 | 2025-04-01T04:33:28.107940 | {
"authors": [
"Aarbel",
"breautek",
"brodybits",
"dfahlander",
"dpogue",
"erisu",
"timbru31",
"vdshb"
],
"repo": "apache/cordova-docs",
"url": "https://github.com/apache/cordova-docs/pull/1042",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
579199457 | How to take a photo or select a file?
How do I make sure that when I activate the camera, I can select whether I want to take a photo or select the image in the gallery?
document.addEventListener("deviceready", onDeviceReady, false);
function onDeviceReady() {
$(document).on("click", ".tirarFoto", tirarFoto);
function tirarFoto(e) {
navigator.camera.getPicture(onSuccess, onFail,{
quality: 35,
saveToPhotoAlbum: true,
destinationType: Camera.DestinationType.DATA_URL,
targetWidth: 800,
targetHeight: 0
});
function onSuccess(imageData) {
img_.src = "data:image/jpeg;base64," + imageData;
input_.value = img_.src;
$(input_).change();
}
function onFail(message){
alert('A camera falhou: ' + message);
}
}
}
you have to create a custom UI for asking (or use dialogs plugin) and based on the user answer configure the camera plugin, this is not supported by the camera plugin
| gharchive/issue | 2020-03-11T11:36:30 | 2025-04-01T04:33:28.109848 | {
"authors": [
"hostcia",
"jcesarmobile"
],
"repo": "apache/cordova-plugin-camera",
"url": "https://github.com/apache/cordova-plugin-camera/issues/570",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
348335998 | Support renaming a database
@davisp:
Distributed systems makes renaming things hard since the dbs db is eventually consistent. For a stop gap, a tool to do the rename offline (ie, stop all nodes in the cluster) could probably be done easily enough.
@janl:
good call, I thought that was going to a be a thing. First reaction was maybe _dbs docs allow for an alias that all nodes can converge towards, dunno if that actually works tho
@davisp:
Also, some of the obvious changes we could make here to do this in an eventually consistent fashion:
Replace the dbname -> shard file name mapping, ie, when we create a database document it stores a UUID that we reference on disk.
Similar to UUID but the "rename" or "alias" doc just lists the original filenames on disk
The downsides to these approaches is that there's currently a large amount of operator knowledge based on the fact that database names map to file system locations. If we break that mapping then we'll have to work on tooling for anyone operating a CouchDB cluster.
Also, with alias docs we'd end up having a single database available under two different names without some other sort of voodoo. So that might be something to think about (though it seems fine on the face of it, just maybe surprising to customers in corner cases).
@wohali:
There actually was a request for DB name aliases 5 years ago in the 2012 summit - so that is very on topic @davisp.
Also, this would help with things like adjusting a db's q (replicate to a new db then change the alias when it's ready). It occurs to me that this would help with compaction, too?
Noting that this one becomes much easier in 4.0. We are already aliasing the DB name to a short key prefix, so renaming a database only requires a (transactional) update of a single KV.
Also gotta bump the db version key so we invalidate all of the cached db handles.
| gharchive/issue | 2018-08-07T14:03:41 | 2025-04-01T04:33:28.118651 | {
"authors": [
"davisp",
"kocolosk",
"wohali"
],
"repo": "apache/couchdb",
"url": "https://github.com/apache/couchdb/issues/1502",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1274816395 | Replicator giving issue , with jobs getting into crashed type
Hi team,
i have a 3 node couch cluster deployed on kubernetes, replication is set in a way like A-B, B-A, B-C, C-B, A-C, C-A using replicator, with continuous replication enabled.
From UI, i can see no issues with replication and everything is replicated properly but when i see the logs, i am seeing following logs,
Replicator GET ****** failed due to error sel_conn_closed
Can someone please help, with this issue? can this logs be ignored?
Also, one more thing would like to share is with scheduler/docs api, i do not see anything in the error count, its 0
It means the _changes feed connection is closed unexpectedly. If the error count is 0, it happens rarely enough that the error count is reset. So it's probably more rare than once a minute or so.
A few things to try:
Check if you're using a restrictive replication filter.
Increase the replicator connection timeout configuration to say 90000(msec).
Check if your load balancer (if you have one), has an idle connection timer then bump that too and then, check that it's not set to buffer connection data.
In general replicator will retry the replications on failure and continue where it left off. It's pretty resilient like that. So, if you don't want to track down the connection issues, and it seems your data replicates as intended you probably don't have to worry about it.
Hi @nickva
Thank you for the quick reply, i would need a further help as i am pretty new to couch
Where do i check this?
restrictive replication filter.
Also can i put this connection timeout under local.d directory in any ini file?
Thanks
Pallavi C
If your replication document doesn't have a filter specified then it's not a concern.
You can put the timeout in any local..d or local.ini file
https://docs.couchdb.org/en/stable/config/replicator.html#replicator/connection_timeout
Closing as there has been no activity in a while. Hopefully the suggested reply helped.
| gharchive/issue | 2022-06-17T09:43:12 | 2025-04-01T04:33:28.124397 | {
"authors": [
"cpallavi",
"nickva"
],
"repo": "apache/couchdb",
"url": "https://github.com/apache/couchdb/issues/4066",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
328604971 | [5/5] Clustered Purge Implementation
Overview
This PR implements clustered purge. Its big.
There are roughly five large chunks for clustered purge:
Single shard APIs
Updates to couch_mrview
Internal replication
Read-repair
Clustered API
At the shard level the major changes are to add two new indexes for purge requests. These indexes store the history of purge requests to a shard (rather than the old method of just storing the most recent purge). This allows for secondary indices and internal replication to manage their eventual consistency. This is the bulk of the PR in that its adding both the implementation and a whole bunch of tests for clustered purge.
One note on naming internally and in commit notes is that a "purge request" is what's sent in a single HTTP request to the purge end point. A "purge info" is a single {DocId, Revs} entry. The new indexes in the shard deal with purge infos as they lose any notion of bundling from which request they came in after they're stored.
The changes to couch_mrview are fairly straightforward after the single node API changes. The only thing that will be interesting in this commit is how we use a _local doc to track where the secondary index has processed in the purge sequence. This is necessary so that compaction knows when it can discard purge infos.
Internal replication is fairly straightforward. This update just ensures that we synchronize our purge infos between each shard so that we don't inadvertently undo a purge request when synchronizing shards.
Read-repair is somewhat tricky. The thing to note here is that when we open a document we track which revisions came from which nodes. Then if read-repair is required we pass that information along with the update request. This way a node can filter out requests from any node that is either a) not up to date with its synchronization or b) we have explicitly purged a revision that has not been sent to the node from whence the revision came back. That last bit of logic may sound kind of odd but the thing to remember is that we could completely remove a revision from a cluster and then re-introduce it later on. This means that we have to reject re-application of a revision for a closed period of time (rather than rejecting it forever).
The clustered API is rather straightforward for anyone familiar any of the existing fabric coordinator and HTTPD handler code.
Testing recommendations
make check
Clustered purge comes with some fairly extensive testing though I'm very much open for suggestions on new tests.
Related Issues or Pull Requests
This PR depends on:
#1366
#1367
#1368
#1369
Checklist
[x] Code is written and works correctly (hopefully ;)
[x] Changes are covered by tests;
[ ] Documentation reflects the changes;
For docs this doesn't change the existing purge APIs but I will need to open a PR against the docs to write up some of the caveats and footgun warnings that we want to add for this.
Also of note, I highly recommend moving through this PR a commit at a time as its quite big and the commits are specifically written to try and separate out all of the major changes to ease review.
Thanks @kocolosk for your review. I tried to address most of your comments on commits 01, 02 and 03 in new commit 7e0dabb. Will work on other comments. @davisp hey Paul, feel free to add your response :-)
Very nice work all.
Historically one could think about external consumers of _changes and internal consumers of the sequence index as being on equal footing, and there's something appealing about that. The latter group has some efficiency advantages but fundamentally an external process could go and do everything that the internal consumers do using the node-local HTTP interface. The lack of HTTP access to the purge index breaks that. An external indexer, for example, has no way of knowing revisions have been purged from the main database short of checking the individual shard's purge sequence and rebuilding the entire index if it changes, whereas the internal consumers can be smarter.
All that said, I wouldn't change anything about this PR, as it contains everything needed to surface an external API to the purge structures if we should conclude there's a need for it at a later date.
I'm +1 pending resolution of my (minor) review comments.
Thanks again for your review @kocolosk
Hey @kocolosk @davisp, I tried to address comments related to code change in commit 9038557 and other previous commits. Any additional comments are highly welcome.
Very nice work. 💯 for a thorough test suite.
I just added commit https://github.com/apache/couchdb/pull/1370/commits/a5c1d1330c8051a9912871049ead60869745c4db to address issue of exposing verify_module and verify_function for unexpected execution. Currently couch_epi:any/5 is used to listen whether there is client reporting a valid index.
In addition, it looks that there is problem with couch_db:get_design_docs/1 and I stripped the suffix (the part after dot) before calling fabric:design_docs/1.
Hey @iilyak I used 1dae6af to adjust returned value from get_purge_seq/1 and get_oldest_purge_seq/1, and use 8a62f68 to add information about module and line. Also added explantation why we use proplists:get_value/2 in fabric_rpc.
I'm +1 pending resolution of my review comments. However I am concerned with deployment aspect of the PR. It is scary that we don't have a rollback. Could we at least have a documentation on what we would do if we would need to rollback to a version which doesn't have additional trees in the header? Ideally I would prefer to have a PR (which we would not be merging) with an implementation of a header downgrade.
thanks again @iilyak I added one response https://github.com/apache/couchdb/pull/1370#discussion_r201238042. Hopefully there is no pending comment from your side except for downgrade database. I am working on downgrade case, but this will be in different PR.
@davisp @jiangphcn With 2.2.0 down to just 4 open issues, and already with PSE in there, let's release that first.
Please hold merging these purge PRs until the 2.2.0 release is cut.
I anticipate that in the next 2 weeks or less.
okay @wohali thanks for letting us know.
Hi Paul, @davisp When using git log --oneline, the order is correct.
localhost:src jiangph$ git log --oneline
bf809a8 [10/10] Clustered Purge: Clustered HTTP API
110dcdc [09/10] Clustered Purge: Fabric API
4dcfb50 [08/10] Clustered Purge: Update read-repair
973a599 [07/10] Clustered Purge: Internal replication
a947b5c [06/10] Clustered Purge: Update mrview indexes
3279c38 [05/10] Clustered Purge: Add upgrade tests
a192aef [04/10] Clustered Purge: Update couch_pse_tests
ac8d0f6 [03/10] Clustered Purge: Update couch_bt_engine
4d695f1 [02/10] Clustered Purge: Update single node APIs
0254305 [01/10] Clustered Purge: Define new purge API
830240b Create new on_compact trigger
4c87d78 Enhance PSE tests with setup/teardown functions
c998aa4 Update to use new couch_pse_tests app
0d85abe Update PSE test definitions for new util module
3c94bc4 Rename PSE test modules
db28f4c Move PSE tests to their own app
f45c036 Rewrite the PSE test suite to use couch_server
da43788 Simplify logic in mem3_rep
aea2833 Update fabric_doc_open eunit tests
bb27d5e Fix race on couch_db:reopen/1
151b9b0 Fix default security object handling
da2e22a Fix bug during purge
3852a00 Fix typos in couch_db_engine.erl
Confirmed locally through not GitHub. Must be ordered on dates or something.
+1 to merge
| gharchive/pull-request | 2018-06-01T17:44:02 | 2025-04-01T04:33:28.139077 | {
"authors": [
"davisp",
"iilyak",
"jiangphcn",
"kocolosk",
"wohali"
],
"repo": "apache/couchdb",
"url": "https://github.com/apache/couchdb/pull/1370",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
233716428 | Don't wrap real error in badmatch exception
Overview
A change in apache/couchdb#576 made any unexpected return from functions called in get_view_index_state/4 to be presented as badmatch and hide the real error.
This unwraps the badmatch and re-throw it as an original response.
Testing recommendations
A test suite make javascript suites=view_errorsjsshould pass.
Related Pull Requests
Bug introduced in apache/couchdb#576
Checklist
[x] Code is written and works correctly;
[ ] Changes are covered by tests;
[ ] Documentation reflects the changes;
You fixed it! +1
| gharchive/pull-request | 2017-06-05T21:29:30 | 2025-04-01T04:33:28.142682 | {
"authors": [
"eiri",
"wohali"
],
"repo": "apache/couchdb",
"url": "https://github.com/apache/couchdb/pull/577",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1021012664 | Restructure Scala code to support JDK 8
Follow up to https://github.com/jw3/example-daffodil-vscode/issues/94:
@arosien said
Daffodil required JDK 8, so we could try to also support it
We currently need JDK 11 to support use of Path.of, which could perhaps be replaced with something available in older JVMs.
Nested object in Scala isn't supported until JDK 9 or greater.
Original issue https://github.com/jw3/example-daffodil-vscode/issues/136
I'm going to reopen this issue. We've solved the compile on Java 8 issue, but not the run on Java 8. Running is much lower priority, but it might want to be fixed some day.
Is Java 8 still something we need to support or are we requiring Java 11 or greater?
Since Daffodil supports Java 8, and there are systems that a still on Java 8, I think there would be value in being able to debug on Java 8 to ensure environments are the same. This doesn't feel high priority to me though.
| gharchive/issue | 2021-10-08T11:54:00 | 2025-04-01T04:33:28.145682 | {
"authors": [
"jw3",
"scholarsmate",
"stevedlawrence"
],
"repo": "apache/daffodil-vscode",
"url": "https://github.com/apache/daffodil-vscode/issues/16",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
824548602 | Update jackson-core to 2.11.4
Updates com.fasterxml.jackson.core:jackson-core from 2.11.3 to 2.11.4.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Files still referring to the old version number
The following files still refer to the old version number (2.11.3).
You might want to review and update them manually.
daffodil-cli/bin.NOTICE
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.fasterxml.jackson.core", artifactId = "jackson-core" } ]
labels: library-update, semver-patch, old-version-remains
I would like us to have a wiki page on our confluence discussing the review criteria for these automated updates. I.e., a checklist of what we should be ckecking on to convince ourselves that one of these upgrades is acceptable.
Agreed, I'll create a page.
I agree we need a checklist. I just realized I overlooked that daffodil-cli/bin.NOTICE still mentions jackson-core's old version number. The review criteria should include:
A determination whether the version update is a patch, minor, or major update. Major updates need more stringent review criteria than the other updates, at least a manual reading of the library's release notes / changelog to check for API changes.
How to update any files like daffodil-cli/bin.NOTICE. Should we push a commit to each and every pull request updating that page AND squash the commits together before merging each pull request? Or update the daffodil-cli/bin.NOTICE file once in one last manually made pull request?
I've created this page as a starting point for documentation:
https://cwiki.apache.org/confluence/display/DAFFODIL/Scala+Steward
Feel free to add anything if I missed anything.
I also added a way I suggest we make changes, which is exactly the same as our normal work flow except any changes are pushed to the appropriate branch on the scala-steward/daffodil fork rather than our own fork. I perfer it done this way (one for every fork) so that the version bump and any related changes (e.g. API updates, license changes) are in the same commit. This might be a bit of a pain for this first go around, but hopefully things haven't changed their licenses. And once we get through this big set, it should be easier to keep up with it.
Another thought, we might want to update our LICENSE/NOTICE files to say something like <VERSION.jar or *.jar so that we don't need to update the LICENSE/NOTICE file for every version bump. That's a bit of a pain, and I'm not sure we gain much extra by including specific jar versions in these files.
Yes, let's replace the explicit version numbers in LICENSE/NOTICE files with foo-.jar to reduce the number of times we need to update our LICENSE/NOTICE files manually to only whenever the library changes its own LICENSE.
I found a problem with the first git command in the Confluence page https://cwiki.apache.org/confluence/display/DAFFODIL/Scala+Steward>. I logged into Confluence (I have an account with the username "interran" and name "John Interrante") but I don't have the ability to edit the page or add a comment to it so I will have to suggest the correction here. The first git command adds the wrong repo URL as the remote URL (right now it adds Scala Stewart's own github repo as the url, not Scala Stewart's github fork of the Daffodil repo). Please change it to:
git remote add scala-steward git@github.com:scala-steward/daffodil.git
Note I'm not sure if we will have permission to push a commit to the Scala Stewart fork, but let's find out.
@tuxji I've given you full permissions to the wiki.
Also, GitHub has a feature that lets owners of a repo to push to fork branches that have been submitted as pull requests as long as the pull request creator allows it (which is the default). Scala steward allows this, so we should be able to push to their daffodil fork.
We don't HAVE to mention the branch name in BOTH the second and third git commands in the Confluence page. We could simply fetch from all our upstream repos and only then say which branch we want to check out:
interran@GH3WPL13E:~/apache/daffodil-asf$ git fetch --all
Fetching origin
Fetching scala-steward
remote: Enumerating objects: 53, done.
remote: Counting objects: 100% (53/53), done.
remote: Compressing objects: 100% (13/13), done.
remote: Total 55 (delta 40), reused 53 (delta 40), pack-reused 2
Unpacking objects: 100% (55/55), 12.81 KiB | 184.00 KiB/s, done.
From github.com:scala-steward/daffodil
* [new branch] master -> scala-steward/master
* [new branch] runtime2-2202 -> scala-steward/runtime2-2202
* [new branch] update/Saxon-HE-9.9.1-8 -> scala-steward/update/Saxon-HE-9.9.1-8
* [new branch] update/icu4j-68.2 -> scala-steward/update/icu4j-68.2
* [new branch] update/jackson-core-2.11.4 -> scala-steward/update/jackson-core-2.11.4
* [new branch] update/jansi-1.18 -> scala-steward/update/jansi-1.18
* [new branch] update/junit-4.13.2 -> scala-steward/update/junit-4.13.2
* [new branch] update/sbt-1.4.8 -> scala-steward/update/sbt-1.4.8
* [new branch] update/sbt-native-packager-1.7.6 -> scala-steward/update/sbt-native-packager-1.7.6
* [new branch] update/scala-library-2.12.13 -> scala-steward/update/scala-library-2.12.13
* [new branch] update/scalacheck-1.15.3 -> scala-steward/update/scalacheck-1.15.3
* [new branch] update/scallop-4.0.2 -> scala-steward/update/scallop-4.0.2
* [new branch] update/typesafe-1.4.1 -> scala-steward/update/typesafe-1.4.1
* [new branch] update/woodstox-core-6.2.4 -> scala-steward/update/woodstox-core-6.2.4
* [new branch] update/xercesImpl-2.12.1 -> scala-steward/update/xercesImpl-2.12.1
interran@GH3WPL13E:~/apache/daffodil-asf$ git checkout update/jackson-core-2.11.4
Branch 'update/jackson-core-2.11.4' set up to track remote branch 'update/jackson-core-2.11.4' from 'scala-steward'.
Switched to a new branch 'update/jackson-core-2.11.4'
interran@GH3WPL13E:~/apache/daffodil-asf$
I figured it would be cleaner to just do it on an individual branch basis so it doesn't pollute everything. Maybe steward will delete these branches once they are merged though, so maybe it's not too bad to just fetch everything. I'll update.
I've downloaded jackson-core-jackson-core-2.11.3.tar.gz, jackson-core-jackson-core-2.11.4.tar.gz, and jackson-core-jackson-core-2.12.2.tar.gz, diff'ed them, found their NOTICE files, and compared their NOTICE file with our bin.NOTICE. I believe we can safely update to 2.11.4 and then 2.12.2 after that since I didn't see any API changes that should affect us. I've incorporated jackson-core's updated NOTICE in our bin.NOTICE and replaced all version numbers in our bin.LICENSE and bin.NOTICE with the literal string <VERSION> to avoid needing to change bin.LICENSE and bin.NOTICE unless dependencies actually change their LICENSE or NOTICE files.
Once I get someone else's +1, I will squash, merge, and move on to the next dependency PR.
+1 looks good to me!
| gharchive/pull-request | 2021-03-08T13:17:10 | 2025-04-01T04:33:28.158813 | {
"authors": [
"mbeckerle",
"scala-steward",
"stevedlawrence",
"tuxji"
],
"repo": "apache/daffodil",
"url": "https://github.com/apache/daffodil/pull/496",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2581975144 | Remove UI
As part of effort outlined in https://github.com/apache/datafusion-ballista/pull/1066 and https://github.com/apache/datafusion-ballista/issues/1067 this PR removes UI package from core ballista.
Relates to: https://github.com/apache/datafusion-ballista/pull/1066 & https://github.com/apache/datafusion-ballista/issues/1067
Which issue does this PR close?
Closes #.
Rationale for this change
What changes are included in this PR?
Are there any user-facing changes?
should be removed now, plus few extra lines i found
done
Looks like there is still a reference in the docker scripts:
+ cd ballista/scheduler/ui
/home/builder/builder-entrypoint.sh: line 27: cd: ballista/scheduler/ui: No such file or directory
| gharchive/pull-request | 2024-10-11T18:40:50 | 2025-04-01T04:33:28.163115 | {
"authors": [
"andygrove",
"milenkovicm"
],
"repo": "apache/datafusion-ballista",
"url": "https://github.com/apache/datafusion-ballista/pull/1072",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2607572167 | [Minor] Update info/list of TPC-DS queries
Which issue does this PR close?
Keep failing queries up to date, refer to new issue.
Rationale for this change
What changes are included in this PR?
Are these changes tested?
Are there any user-facing changes?
thanks for the review @comphead
| gharchive/pull-request | 2024-10-23T07:30:10 | 2025-04-01T04:33:28.165448 | {
"authors": [
"Dandandan"
],
"repo": "apache/datafusion",
"url": "https://github.com/apache/datafusion/pull/13075",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1040858523 | [Bug] [Master] Failed to operate workflow node separately
Search before asking
[X] I had searched in the issues and found no similar issues.
What happened
Failed to operate workflow node alone. No new task instance is generated.
version: dev
There are no problems in the version of 2.0-alpha-release.
What you expected to happen
Operate workflow node separately and generate new task instance.
How to reproduce
Operate workflow node separately.
Anything else
[ERROR] 2021-11-01 15:30:35.452 org.apache.dolphinscheduler.dao.utils.DagHelper:[112] - start node name [a] is not in task node list [[TaskNode{id='null', code=851434614521856, version=1, name='a', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"a\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[]', preTaskNodeList=null, extras='null', depList=[], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}]]
[ERROR] 2021-11-01 15:30:35.452 org.apache.dolphinscheduler.dao.utils.DagHelper:[112] - start node name [a] is not in task node list [[TaskNode{id='null', code=851434614521856, version=1, name='a', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"a\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[]', preTaskNodeList=null, extras='null', depList=[], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}]]
[ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[512] - processDag is null
[ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.dao.utils.DagHelper:[112] - start node name [job_b1] is not in task node list [[TaskNode{id='null', code=851435467538432, version=1, name='job_b1', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"b1\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[851437597130752]', preTaskNodeList=null, extras='null', depList=[851437597130752], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}, TaskNode{id='null', code=851437597130752, version=2, name='depend_a', desc='', type='DEPENDENT', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"waitStartTimeout":{"strategy":"FAILED","interval":null,"checkInterval":null,"enable":false},"switchResult":{}}', preTasks='[]', preTaskNodeList=null, extras='null', depList=[], dependence='{"relation":"AND","dependTaskList":[{"relation":"AND","dependItemList":[{"projectCode":851275541217280,"definitionCode":851435217125376,"depTasks":"a","cycle":"day","dateValue":"today"}]}]}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}, TaskNode{id='null', code=851436591874048, version=1, name='job_b2', desc='', type='SHELL', runFlag='NORMAL', loc='null', maxRetryTimes=0, retryInterval=1, params='{"resourceList":[],"localParams":[],"rawScript":"echo \"b2\"","waitStartTimeout":{},"switchResult":{}}', preTasks='[851437597130752]', preTaskNodeList=null, extras='null', depList=[851437597130752], dependence='{}', conditionResult='{"successNode":[""],"failedNode":[""]}', taskInstancePriority=MEDIUM, workerGroup='default', environmentCode=-1, timeout='{"enable":false,"strategy":null,"interval":0}', delayTime=0}]]
[ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[512] - processDag is null
[ERROR] 2021-11-01 15:30:35.453 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[512] - processDag is null
[ERROR] 2021-11-01 15:30:35.454 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[195] - handler error:
java.lang.NullPointerException: null
at org.apache.dolphinscheduler.dao.utils.DagHelper.parsePostNodes(DagHelper.java:301)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.submitPostNode(WorkflowExecuteThread.java:797)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.startProcess(WorkflowExecuteThread.java:464)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.run(WorkflowExecuteThread.java:192)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
[ERROR] 2021-11-01 15:30:35.454 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[195] - handler error:
java.lang.NullPointerException: null
at org.apache.dolphinscheduler.dao.utils.DagHelper.parsePostNodes(DagHelper.java:301)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.submitPostNode(WorkflowExecuteThread.java:797)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.startProcess(WorkflowExecuteThread.java:464)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.run(WorkflowExecuteThread.java:192)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
[ERROR] 2021-11-01 15:30:35.454 org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread:[195] - handler error:
java.lang.NullPointerException: null
at org.apache.dolphinscheduler.dao.utils.DagHelper.parsePostNodes(DagHelper.java:301)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.submitPostNode(WorkflowExecuteThread.java:797)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.startProcess(WorkflowExecuteThread.java:464)
at org.apache.dolphinscheduler.server.master.runner.WorkflowExecuteThread.run(WorkflowExecuteThread.java:192)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
Are you willing to submit PR?
[ ] Yes I am willing to submit a PR!
Code of Conduct
[X] I agree to follow this project's Code of Conduct
The hi error is the dev code has a judgment logic, I will fix it today @dailidong Can you assign this problem to me
hi, this error is the dev code has a judgment logic, I will fix it today @dailidong Can you assign this problem to me
done
| gharchive/issue | 2021-11-01T07:33:49 | 2025-04-01T04:33:28.175383 | {
"authors": [
"boy-xiaozhang",
"dailidong",
"shangeyao"
],
"repo": "apache/dolphinscheduler",
"url": "https://github.com/apache/dolphinscheduler/issues/6657",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1423661368 | Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.1:compile failed: invalid versionNumber : major.minor
Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.1:compile failed: invalid versionNumber : major.minor
Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.1:compile failed: invalid versionNumber : major.minor
| gharchive/issue | 2022-10-26T08:48:47 | 2025-04-01T04:33:28.177432 | {
"authors": [
"Cyb315966632"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/issues/13687",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1560743371 | refactor bind slot, function and do type coercion in one rule
Proposed changes
Issue Number: close #xxx
Problem summary
Describe your changes.
Checklist(Required)
Does it affect the original behavior:
[ ] Yes
[ ] No
[ ] I don't know
Has unit tests been added:
[ ] Yes
[ ] No
[ ] No Need
Has document been added or modified:
[ ] Yes
[ ] No
[ ] No Need
Does it need to update dependencies:
[ ] Yes
[ ] No
Are there any changes that cannot be rolled back:
[ ] Yes (If Yes, please explain WHY)
[ ] No
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
TeamCity pipeline, clickbench performance test result:
the sum of best hot time: 34.17 seconds
load time: 524 seconds
storage size: 17122260141 Bytes
https://doris-community-test-1308700295.cos.ap-hongkong.myqcloud.com/tmp/20230128113743_clickbench_pr_85630.html
| gharchive/pull-request | 2023-01-28T07:53:06 | 2025-04-01T04:33:28.183597 | {
"authors": [
"hello-stephen",
"morrySnow"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/16156",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1679846682 | Chore add some regression-test case about materialized-view
Proposed changes
add some regression-test case about materialized-view
Problem summary
Describe your changes.
Checklist(Required)
[ ] Does it affect the original behavior
[ ] Has unit tests been added
[ ] Has document been added or modified
[ ] Does it need to update dependencies
[ ] Is this PR support rollback (If NO, please explain WHY)
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
| gharchive/pull-request | 2023-04-23T04:01:51 | 2025-04-01T04:33:28.186654 | {
"authors": [
"BiteTheDDDDt"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/18946",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1827962547 | feature refactor CSV reading process during scanning, and support enclose and escape for stream load
Proposed changes
Refactor thoughts: close #22383
Descriptions about enclose and escape: #22385
Further comments
Cases should be discussed:
When an incomplete enclose appears in the beginning of a large scale data, the line delimiter will be unreachable till the EOF, will the buffer become extremely large?
What if an infinite line occurs in the case? Essentially, 1. is equivalent to this.
Only support stream load as trial in this PR, avoid too many unrelated changes. Docs will be added when enclose and escape is available for all kinds of load.
run buildall
run buildall
run buildall
run buildall
run buildall
run buildall
run buildall
run buildall
run buildall
run p0
run p0
run p0
run p0
run click bench
run clickbench
run p0
Seems weird, p0 has passed as shown. Correctness is verified, now will resolve the performance problem.
run buildall
run buildall
run clickbench
run buildall
run buildall
run buildall
run build all
run buildall
run clickbench
run p0
| gharchive/pull-request | 2023-07-30T14:49:15 | 2025-04-01T04:33:28.194366 | {
"authors": [
"TangSiyang2001",
"Yukang-Lian",
"dataroaring"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/22382",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1851194115 | fixbind slot failed because of "default_cluster"
Proposed changes
slot bind failed for following querys:
select tpch.lineitem.* from lineitem
select tpch.lineitem.l_partkey from lineitem
the unbound slot is lineitem.l_partkey, but the bounded slot is default_cluster:tpch.lineitem.l_partkey. They are not matched.
we need to ignore default_cluster: when compare dbName
Issue Number: close #xxx
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
(From new machine)TeamCity pipeline, clickbench performance test result:
the sum of best hot time: 46.95 seconds
stream load tsv: 510 seconds loaded 74807831229 Bytes, about 139 MB/s
stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s
stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s
stream load parquet: 31 seconds loaded 861443392 Bytes, about 26 MB/s
insert into select: 29.2 seconds inserted 10000000 Rows, about 342K ops/s
storage size: 17162181158 Bytes
run buildall
(From new machine)TeamCity pipeline, clickbench performance test result:
the sum of best hot time: 46.14 seconds
stream load tsv: 509 seconds loaded 74807831229 Bytes, about 140 MB/s
stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s
stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s
stream load parquet: 32 seconds loaded 861443392 Bytes, about 25 MB/s
insert into select: 29.2 seconds inserted 10000000 Rows, about 342K ops/s
storage size: 17162677169 Bytes
run buildall
(From new machine)TeamCity pipeline, clickbench performance test result:
the sum of best hot time: 45.51 seconds
stream load tsv: 512 seconds loaded 74807831229 Bytes, about 139 MB/s
stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s
stream load orc: 65 seconds loaded 1101869774 Bytes, about 16 MB/s
stream load parquet: 30 seconds loaded 861443392 Bytes, about 27 MB/s
insert into select: 29.2 seconds inserted 10000000 Rows, about 342K ops/s
storage size: 17162113308 Bytes
| gharchive/pull-request | 2023-08-15T10:02:21 | 2025-04-01T04:33:28.201342 | {
"authors": [
"englefly",
"hello-stephen"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/23008",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1873279069 | Bug fix load db use analyzer to analyze diffrent metaindex
Proposed changes
fix load db use analyzer to analyze diffrent metaindex
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
(From new machine)TeamCity pipeline, clickbench performance test result:
the sum of best hot time: 48.79 seconds
stream load tsv: 534 seconds loaded 74807831229 Bytes, about 133 MB/s
stream load json: 20 seconds loaded 2358488459 Bytes, about 112 MB/s
stream load orc: 64 seconds loaded 1101869774 Bytes, about 16 MB/s
stream load parquet: 32 seconds loaded 861443392 Bytes, about 25 MB/s
insert into select: 28.9 seconds inserted 10000000 Rows, about 346K ops/s
storage size: 17162010827 Bytes
| gharchive/pull-request | 2023-08-30T09:18:12 | 2025-04-01T04:33:28.204301 | {
"authors": [
"BiteTheDDDDt",
"hello-stephen"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/23673",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1931231909 | FixQuery the transaction status NPE when the task has not yet started scheduling
error log
When the task has not yet started scheduling, there is no transaction status at this time. Therefore the query will cause an NPE exception
show routine load task where JobName="npe"
2023-10-07 13:52:16,883 WARN (mysql-nio-pool-3|1310) [StmtExecutor.executeByLegacy():807] execute Exception. stmt[1368, fc075cfd62aa4459-a8014ca6b5f574fa]
java.lang.NullPointerException: null
at org.apache.doris.load.routineload.RoutineLoadJob.lambda$getTasksShowInfo$6(RoutineLoadJob.java:1452) ~[classes/:?]
at java.util.ArrayList.forEach(ArrayList.java:1259) ~[?:1.8.0_362]
at org.apache.doris.load.routineload.RoutineLoadJob.getTasksShowInfo(RoutineLoadJob.java:1450) ~[classes/:?]
at org.apache.doris.qe.ShowExecutor.handleShowRoutineLoadTask(ShowExecutor.java:1605) ~[classes/:?]
at org.apache.doris.qe.ShowExecutor.execute(ShowExecutor.java:316) ~[classes/:?]
at org.apache.doris.qe.StmtExecutor.handleShow(StmtExecutor.java:2208) ~[classes/:?]
at org.apache.doris.qe.StmtExecutor.executeByLegacy(StmtExecutor.java:775) ~[classes/:?]
at org.apache.doris.qe.StmtExecutor.execute(StmtExecutor.java:468) ~[classes/:?]
at org.apache.doris.qe.StmtExecutor.execute(StmtExecutor.java:438) ~[classes/:?]
at org.apache.doris.qe.ConnectProcessor.handleQuery(ConnectProcessor.java:353) ~[classes/:?]
at org.apache.doris.qe.ConnectProcessor.dispatch(ConnectProcessor.java:501) ~[classes/:?]
at org.apache.doris.qe.ConnectProcessor.processOnce(ConnectProcessor.java:752) ~[classes/:?]
at org.apache.doris.mysql.ReadListener.lambda$handleEvent$0(ReadListener.java:52) ~[classes/:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) ~[?:1.8.0_362]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ~[?:1.8.0_362]
Changes
When the task has not yet started scheduling, that is, when txn_id is -1, the transaction status is set to null
Test
run buildall
Run P0
run p0
| gharchive/pull-request | 2023-10-07T07:15:58 | 2025-04-01T04:33:28.206967 | {
"authors": [
"CalvinKirs"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/25074",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1975480012 | FIXFix cast decimalv2 2 decimalv3
Proposed changes
Issue Number: close #xxx
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
run p0
| gharchive/pull-request | 2023-11-03T05:25:24 | 2025-04-01T04:33:28.208919 | {
"authors": [
"amorynan"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/26368",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2040110132 | chore Add MVN_OPT env variable to enrich building FE with extra options
e.g. just export or add it to custom_env.sh
export MVN_OPT="-o"
will build FE with maven option "-o" (offline), which means maven does not need to download meta from maven repo, it is useful for saving time if the internet is unstable or unusable.
run buidlall
run buildall
run buildall
run buildall
| gharchive/pull-request | 2023-12-13T16:49:29 | 2025-04-01T04:33:28.211442 | {
"authors": [
"gavinchou"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/28375",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2047780550 | refactormake NormalizeAggregate rule more clear and readable
Proposed changes
Issue Number: close #xxx
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
run buildall
run buildall
run buildall
| gharchive/pull-request | 2023-12-19T01:59:58 | 2025-04-01T04:33:28.213639 | {
"authors": [
"starocean999"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/28607",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2143662213 | Fix BE gracefully exit when port conflicts
Proposed changes
Issue Number: close #xxx
almost fixed. still some bugs in rocksdb itself when in Release mode
before:
coredump in somewhere
now log:
Doris BE server did not start correctly, exiting
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
run buildall
| gharchive/pull-request | 2024-02-20T06:53:02 | 2025-04-01T04:33:28.215919 | {
"authors": [
"zclllyybb"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/31143",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2168165399 | fix json_object can't be odd parameters, need even parameters
Proposed changes
Issue Number: close #31766
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
@iwanttobepowerful please add a testcase
| gharchive/pull-request | 2024-03-05T02:46:58 | 2025-04-01T04:33:28.217778 | {
"authors": [
"iwanttobepowerful",
"xiaokang"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/31767",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2224786605 | fix exclude cache ttl
Proposed changes
Issue Number: close #xxx
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
run buildall
| gharchive/pull-request | 2024-04-04T08:10:09 | 2025-04-01T04:33:28.219596 | {
"authors": [
"dataroaring"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/33253",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2267468380 | test add batch size in long time case
Proposed changes
For even PRs, the batch size is set to 50, and this case runs for 50 minutes.
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
run buildall
| gharchive/pull-request | 2024-04-28T10:06:38 | 2025-04-01T04:33:28.221253 | {
"authors": [
"Mryange"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/34249",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2363391117 | Fix Fix core in orc and parquet reader sometimes after low mem exception.
Proposed changes
Backport #36574.
run buildall
| gharchive/pull-request | 2024-06-20T02:43:30 | 2025-04-01T04:33:28.222449 | {
"authors": [
"kaka11chen"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/36575",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2387644716 | chore Add more clear error message when loading timed out
Proposed changes
Issue Number: close #xxx
run buildall
| gharchive/pull-request | 2024-07-03T04:29:16 | 2025-04-01T04:33:28.223476 | {
"authors": [
"zclllyybb"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/37198",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2588059377 | fixOrToIn and SimplifyRange rules dead loop #41689 (3.0)
Proposed changes
pick #41689
Issue Number: close #xxx
run buildall
run buildall
run feut
| gharchive/pull-request | 2024-10-15T08:42:01 | 2025-04-01T04:33:28.224949 | {
"authors": [
"englefly"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/41887",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2631728506 | fix Add synchronized to avoid concurrent modification
Without synchronized protection, when the user manually executes cancel, the backupHandler may also trigger cancel simultaneously. Modifying the snapshotInfo by both at the same time will cause a concurrent modification exception.
The pending state will synchronously wait for the task of creating a replica to end. Therefore, a creating state needs to be added to convert the synchronous wait into asynchronous polling to avoid blocking the user's cancel/list job info requests.
run buildall
run buildall
run buildall
| gharchive/pull-request | 2024-11-04T03:16:04 | 2025-04-01T04:33:28.226599 | {
"authors": [
"w41ter"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/43172",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2682227059 | [DNM]
What problem does this PR solve?
Issue Number: close #xxx
Related PR: #xxx
Problem Summary:
Release note
None
Check List (For Author)
Test
[ ] Regression test
[ ] Unit Test
[ ] Manual test (add detailed scripts or steps below)
[ ] No need to test or manual test. Explain why:
[ ] This is a refactor/code format and no logic has been changed.
[ ] Previous test can cover this change.
[ ] No code files have been changed.
[ ] Other reason
Behavior changed:
[ ] No.
[ ] Yes.
Does this need documentation?
[ ] No.
[ ] Yes.
Check List (For Reviewer who merge this PR)
[ ] Confirm the release note
[ ] Confirm test cases
[ ] Confirm document
[ ] Add branch pick label
run buildall
run cloud_p0
run compile
run cloud_p0
| gharchive/pull-request | 2024-11-22T07:51:35 | 2025-04-01T04:33:28.232975 | {
"authors": [
"TangSiyang2001"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/44473",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2721901086 | branch-2.1: opt Allow to miss auto-increment column and other value columns in partial update #44528
pick https://github.com/apache/doris/pull/44528
run buildall
| gharchive/pull-request | 2024-12-06T02:33:25 | 2025-04-01T04:33:28.234298 | {
"authors": [
"bobhan1"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/45073",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2757535382 | enhanceAdd the default configuration file for ranger under fe/conf
What problem does this PR solve?
Previously, it was necessary to copy the configuration file from the official website
Afterwards, only the necessary configuration items need to be changed
Issue Number: close #xxx
Related PR: #xxx
Problem Summary:
Add the default configuration file for ranger under fe/conf
Release note
None
Check List (For Author)
Test
[ ] Regression test
[ ] Unit Test
[ ] Manual test (add detailed scripts or steps below)
[x] No need to test or manual test. Explain why:
[x] This is a refactor/code format and no logic has been changed.
[ ] Previous test can cover this change.
[ ] No code files have been changed.
[ ] Other reason
Behavior changed:
[x] No.
[ ] Yes.
Does this need documentation?
[ ] No.
[x] Yes.
Check List (For Reviewer who merge this PR)
[ ] Confirm the release note
[ ] Confirm test cases
[ ] Confirm document
[ ] Add branch pick label
Thank you for your contribution to Apache Doris.
Don't know what should be done next? See How to process your PR.
Please clearly describe your PR:
What problem was fixed (it's best to include specific error reporting information). How it was fixed.
Which behaviors were modified. What was the previous behavior, what is it now, why was it modified, and what possible impacts might there be.
What features were added. Why was this function added?
Which code was refactored and why was this part of the code refactored?
Which functions were optimized and what is the difference before and after the optimization?
run buildall
run buildall
| gharchive/pull-request | 2024-12-24T10:05:00 | 2025-04-01T04:33:28.242911 | {
"authors": [
"hello-stephen",
"zddr"
],
"repo": "apache/doris",
"url": "https://github.com/apache/doris/pull/45869",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.