id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
293928891 | DRILL-6130: Fix NPE during physical plan submission for various storage plugins
Fixed ser / de issues for Hive, Kafka, Hbase plugins.
Added physical plan submission unit test for all storage plugins in contrib module.
Refactoring.
@vdiravka please review.
@vdiravka thanks for the code review. Addressed two comments.
| gharchive/pull-request | 2018-02-02T15:39:09 | 2025-04-01T04:33:28.244567 | {
"authors": [
"arina-ielchiieva"
],
"repo": "apache/drill",
"url": "https://github.com/apache/drill/pull/1108",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
744391154 | Doc Fix : modify druid.historical.cache.maxEntrySize property in Unified format
Description
The property druid.historical.cache.maxEntrySize mentioned in Configuration reference -> Historical -> Historical Caching is inconsistent with the existing format, lack of Possible Values.
This PR fills in this column as positive integer
This PR has:
[x] been self-reviewed.
Key changed/added classes in this PR
docs/configuration/index.md
@a2l007 @nishantmonu51 Thanks for your review and merge!
| gharchive/pull-request | 2020-11-17T04:04:09 | 2025-04-01T04:33:28.248375 | {
"authors": [
"zhangyue19921010"
],
"repo": "apache/druid",
"url": "https://github.com/apache/druid/pull/10590",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2555617371 | [Backport] Fix maxRowsInMemory default for streaming (#17028)
Backport apache/druid#17028
Thank you for the backport, @kfaraz
| gharchive/pull-request | 2024-09-30T05:42:45 | 2025-04-01T04:33:28.249268 | {
"authors": [
"AmatyaAvadhanula",
"kfaraz"
],
"repo": "apache/druid",
"url": "https://github.com/apache/druid/pull/17186",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
733657298 | DubboHandlerRegistry#removeService NPE问题
[x] I have searched the issues of this repository and believe that this is not a duplicate.
[x] I have checked the FAQ of this repository and believe that this is not a duplicate.
Environment
Dubbo version: 2.7.7
Operating System version: macOs Mojave 10.14.5
Java version: 1.8
Steps to reproduce this issue
使用dubbo启动一个任意grpc Provider进程
关闭进程
Expected Result
进程正常关闭,资源回收,无任何异常
Actual Result
有NullPointerException报出
java.lang.NullPointerException: null
at org.apache.dubbo.rpc.protocol.grpc.DubboHandlerRegistry.removeService(DubboHandlerRegistry.java:66) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.rpc.protocol.grpc.GrpcProtocol.lambda$doExport$1(GrpcProtocol.java:97) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.rpc.protocol.AbstractProxyProtocol$1.unexport(AbstractProxyProtocol.java:95) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.rpc.listener.ListenerExporterWrapper.unexport(ListenerExporterWrapper.java:71) [dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.config.ServiceConfig.unexport(ServiceConfig.java:170) [dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.config.bootstrap.DubboBootstrap.lambda$unexportServices$17(DubboBootstrap.java:952) [dubbo-2.7.7.jar:2.7.7]
at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_231]
at org.apache.dubbo.config.bootstrap.DubboBootstrap.unexportServices(DubboBootstrap.java:950) [dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.config.bootstrap.DubboBootstrap.destroy(DubboBootstrap.java:1072) [dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.config.bootstrap.DubboBootstrap$1.callback(DubboBootstrap.java:193) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.common.function.ThrowableAction.execute(ThrowableAction.java:46) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.common.lang.ShutdownHookCallbacks.lambda$callback$0(ShutdownHookCallbacks.java:70) ~[dubbo-2.7.7.jar:2.7.7]
at java.lang.Iterable.forEach(Iterable.java:75) ~[na:1.8.0_231]
at org.apache.dubbo.common.lang.ShutdownHookCallbacks.callback(ShutdownHookCallbacks.java:70) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.config.DubboShutdownHook.callback(DubboShutdownHook.java:85) ~[dubbo-2.7.7.jar:2.7.7]
at org.apache.dubbo.config.DubboShutdownHook.run(DubboShutdownHook.java:73) ~[dubbo-2.7.7.jar:2.7.7]
如下两个地方都会去调用 DubboHandlerRegistry#removeService,但这个方法貌似没有重复调用的情况,重复调用会NPE
DubboShutdownHook#destroyProtocols
DubboBootstrap#unexportServices
void removeService(String serviceKey) {
ServerServiceDefinition service = services.remove(serviceKey); // 此处service可能为空
for (ServerMethodDefinition<?, ?> method : service.getMethods()) {
methods.remove(method.getMethodDescriptor().getFullMethodName(), method);
}
}
Can you provide detailed steps to reproduce? I tested it here and it didn’t reproduce the NPE.
| gharchive/issue | 2020-10-31T07:21:04 | 2025-04-01T04:33:28.260090 | {
"authors": [
"xianliguang",
"xiaoheng1"
],
"repo": "apache/dubbo",
"url": "https://github.com/apache/dubbo/issues/6856",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
565841475 | some code optimization
What is the purpose of the change
some code optimization
Brief changelog
XXXXX
Verifying this change
XXXXX
Follow this checklist to help us incorporate your contribution quickly and easily:
[x] Make sure there is a GITHUB_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GITHUB issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue.
[ ] Format the pull request title like [Dubbo-XXX] Fix UnknownException when host config not exist #XXX. Each commit in the pull request should have a meaningful subject line and body.
[ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why.
[ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add sample in dubbo samples project.
[ ] Run mvn clean install -DskipTests=false & mvn clean test-compile failsafe:integration-test to make sure unit-test and integration-test pass.
[ ] If this contribution is large, please follow the Software Donation Guide.
Codecov Report
Merging #5746 into master will decrease coverage by 0.02%.
The diff coverage is 33.33%.
@@ Coverage Diff @@
## master #5746 +/- ##
============================================
- Coverage 61.24% 61.22% -0.03%
- Complexity 422 495 +73
============================================
Files 922 928 +6
Lines 37498 37899 +401
Branches 5431 5447 +16
============================================
+ Hits 22967 23204 +237
- Misses 12055 12176 +121
- Partials 2476 2519 +43
Impacted Files
Coverage Δ
Complexity Δ
...g/apache/dubbo/common/bytecode/ClassGenerator.java
61.53% <0%> (ø)
0 <0> (ø)
:arrow_down:
...a/org/apache/dubbo/monitor/dubbo/DubboMonitor.java
88.57% <100%> (ø)
16 <0> (ø)
:arrow_down:
...g/context/properties/DefaultDubboConfigBinder.java
0% <0%> (-100%)
0% <0%> (ø)
...t/annotation/DubboLifecycleComponentRegistrar.java
0% <0%> (-100%)
0% <0%> (ø)
...ache/dubbo/rpc/cluster/support/ForkingCluster.java
0% <0%> (-100%)
0% <0%> (ø)
.../context/properties/AbstractDubboConfigBinder.java
0% <0%> (-92.31%)
0% <0%> (ø)
...ring/context/config/DubboConfigBeanCustomizer.java
0% <0%> (-66.67%)
0% <0%> (ø)
...che/dubbo/rpc/cluster/support/FailfastCluster.java
50% <0%> (-50%)
0% <0%> (ø)
...va/org/apache/dubbo/remoting/TimeoutException.java
0% <0%> (-22.23%)
0% <0%> (ø)
...dubbo/remoting/exchange/support/DefaultFuture.java
71.96% <0%> (-14.53%)
0% <0%> (ø)
... and 120 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update b288ad7...b110fa4. Read the comment docs.
thanks for your contribution :)
| gharchive/pull-request | 2020-02-16T02:49:02 | 2025-04-01T04:33:28.279528 | {
"authors": [
"ZZMarquis",
"codecov-io",
"htynkn"
],
"repo": "apache/dubbo",
"url": "https://github.com/apache/dubbo/pull/5746",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
925251311 | Contribute to options documentation
Is there a way for me to contribute to the options documentation? For example, in the screenshot below, you can see that the "label.position" for bar charts has the proper "uiControl" setting from the option.series-bar.js file. However, the "uiControl" is missing for pie charts and I'd like to add it. There are a lot of errors and missing items that I come across and would love to contribute.
Sure, this is the repo for Apache ECharts option doc. You can fork it and then make PRs.
| gharchive/issue | 2021-06-18T22:57:27 | 2025-04-01T04:33:28.281727 | {
"authors": [
"ataft",
"susiwen8"
],
"repo": "apache/echarts-doc",
"url": "https://github.com/apache/echarts-doc/issues/188",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
796063781 | 5.0版本resize是否取消掉了缓动动画,为什么执行resize()后缓动效果没有了?
`let myChart = this.$Echarts.init(document.getElementById('myDiv'));
let option = {
xAxis: {
type: 'category',
data: ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun']
},
yAxis: {
type: 'value'
},
series: [{
data: [120, 200, 150, 80, 70, 110, 130],
type: 'bar'
}],
animation: true
};
myChart.setOption(option);
myChart.resize();`
这段代码在4.x下,即使执行了resize,缓动效果还是存在的,可是升级了5.0后,执行了resize之后缓动效果就没有了
resize动画效果确实没了,请问答主加上了吗
| gharchive/issue | 2021-01-28T14:36:24 | 2025-04-01T04:33:28.284152 | {
"authors": [
"JManToGithub",
"zhoufanglu"
],
"repo": "apache/echarts",
"url": "https://github.com/apache/echarts/issues/14146",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
833525888 | about bmap, Incomplete rendering load(关于echart中地图,最近开始加载不全的问题)
Version
5.0.2
Reproduction link
https://echarts.apache.org/examples/zh/editor.html?c=lines-bmap-effect
Steps to reproduce
No code modifiednothing
What is expected?
complete rendering load
What is actually happening?
No code modified
Is it a network problem?
这是百度地图 2.0 自定义地图用到的瓦片地图服务器的问题,可能会遇到白块的地方,ECharts 无法解决此问题。建议尝试使用百度地图 3.0。
这是百度地图 2.0 自定义地图用到的瓦片地图服务器的问题,可能会遇到白块的地方,ECharts 无法解决此问题。建议尝试使用百度地图 3.0。
我用的是 "echarts": "^5.0.0",按理来说应该是3.0了
ECharts 5.0 与百度地图 3.0 没有关系,百度地图 API 脚本是需要开发者自行引入的,所以使用的版本由开发者来决定。
ECharts 5.0 与百度地图 3.0 没有关系,百度地图 API 脚本是需要开发者自行引入的,所以使用的版本由开发者来决定。
因为我写的是vue3.0的框架,所以引用了echarts官方最新的依赖。看了下node_moudles源码包里面bmap的拓展,确实引用3.0版本了
不太清楚你是否有理解,ECharts 并不会帮我们自动引入百度地图 API script,而是需要开发者在代码里手动引入。
<script type="text/javascript" src="https://api.map.baidu.com/api?v=3.0&ak=您的密钥"></script>
不太清楚你是否有理解,ECharts 并不会帮我们自动引入百度地图 API script,而是需要开发者在代码里手动引入。
<script type="text/javascript" src="https://api.map.baidu.com/api?v=3.0&ak=您的密钥"></script>
理解的,看上图
所以 3.0 仍然出现瓦片地图加载不全的问题吗?
所以 3.0 仍然出现瓦片地图加载不全的问题吗?
是的,2.0,3.0,都会出来。
这种情况只能建议向百度地图提工单咨询了。
这种情况只能建议向百度地图提工单咨询了。
感谢回答
| gharchive/issue | 2021-03-17T08:28:06 | 2025-04-01T04:33:28.291014 | {
"authors": [
"joel785454405",
"plainheart"
],
"repo": "apache/echarts",
"url": "https://github.com/apache/echarts/issues/14496",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
916823335 | Export image of the charts using CORS resources
Background
In most cases, one chart may be referring to some external resources, some of them allow requests from any origin with the response header like Access-Control-Allow-Origin: '*' but some of them don't.
For those resources with CORS header, we hope they can be exported. See this simple example,
option = {
xAxis: {
type: 'category',
data: ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun']
},
yAxis: {
type: 'value'
},
toolbox: {
feature: {
saveAsImage: {}
}
},
series: [{
data: [150, 230, 224, 218, 135, 147, 260],
type: 'line',
symbolSize: [95, 22],
symbol: 'image://https://cdn.jsdelivr.net/gh/apache/echarts-website@asf-site/zh/images/logo.png'
}]
};
If we hope to get a screenshot through the current API getDataURL, we would get an unexpected error like this,
Uncaught DOMException: Failed to execute 'toDataURL' on 'HTMLCanvasElement': Tainted canvases may not be exported.
To avoid this, we will have to create some Image objects and specify their own crossOrigin property. Like this,
const logoImg = new Image();
logoImg.src = 'image://https://cdn.jsdelivr.net/gh/apache/echarts-website@asf-site/zh/images/logo.png';
logoImg.crossOrigin = ''; // or 'anonymous'
logoImg.onload = () => { // xxx };
But here is another limit, our series.symbol doesn't support the Image object yet! So we can't directly use the created image objects with the crossOrigin property. But we can try another approach: convert it to a Base64 URL. The code will be,
const symbolSize = [95, 22];
const logoImg = new Image();
logoImg.src = 'https://cdn.jsdelivr.net/gh/apache/echarts-website@asf-site/zh/images/logo.png';
logoImg.crossOrigin = ''; // or 'anonymous'
logoImg.onload = () => {
const tmpCanvas = echarts.zrUtil.createCanvas()
const tmpCtx = tmpCanvas.getContext('2d')
const dpr = myChart.getDevicePixelRatio()
tmpCanvas.width = symbolSize[0] * dpr
tmpCanvas.height = symbolSize[1] * dpr
tmpCanvas.style.width = symbolSize[0] + 'px'
tmpCanvas.style.height = symbolSize[1] + 'px'
tmpCtx.drawImage(logoImg, 0, 0, tmpCanvas.width, tmpCanvas.height)
myChart.setOption({
series: [{
symbol: 'image://' + tmpCanvas.toDataURL()
}]
})
}
option = {
xAxis: {
type: 'category',
data: ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun']
},
yAxis: {
type: 'value'
},
toolbox: {
feature: {
saveAsImage: {}
}
},
series: [{
data: [150, 230, 224, 218, 135, 147, 260],
type: 'line',
symbolSize: symbolSize,
symbol: 'none'
}]
};
Now let's export again, we would found it can be exported and downloaded successfully.
But everything was going to be complex when we hope to set a different image for each symbol item.
It would be simpler and easier if there could be an option to specify whether the image URL is cross-origin.
This feature will involve some required changes in zrender, but that's underlying. Let's leave it alone now and just discuss the things related to ECharts.
Proposed Usage
For symbol with image://, I prefer specifying the crossOrigin in the URL,
either image://crossOrigin/URL or image://URL?crossOrigin.
For the common pattern object,
{
image: imageDom, // Image, HTMLImageElement, and HTMLCanvasElement are supported, while string path is not supported
repeat: 'repeat' // whether to repeat texture, whose value can be repeat-x, repeat-y, or no-repeat
}
The crossOrigin can be set in imageDom in advance. So it should be OK. We can also provide a new option crossOrigin if needed.
Demo
https://www.makeapie.com/editor.html?c=xntd1QC3kI&v=1
Yeah, it's a very useful feature. ECharts GL also has a similar issue. I hardcoded cross-origin to anonymous(or it will not even display because of the strict origin policy in WebGL). But how to make it configurable still bothers me. Specifying it in the URL may be a good idea. But image://URL?crossOrigin may be conflicts with the exists ? in the URL itself.
Another way is to specify this image policy at the beginning of initializing the echarts instead of adding it on every image.
const chart = echarts.init(dom, null, {
crossOrigin: 'anonymous'
})
I'm not sure if using different CORS policies in the same echarts instance is a common sceneraio. Images are not very often used in echarts. Perhaps we can support the global CORS policy currently and see if any developers need more precisely control on each image. In either way, I think it will be convenient for developers if they only need to specify crossOrigin policy once.
Any status on this ? i tried the method suggested by @pissang but i am getting type error for the same here is my code snippet :
const chart = echarts.init(chartRef.current, "100ms-echarts", { crossOrigin: "anonymous", })
Argument of type '{ crossOrigin: string; }' is not assignable to parameter of type 'EChartsInitOpts'.
Object literal may only specify known properties, and 'crossOrigin' does not exist in type 'EChartsInitOpts'.
@hdz-666 This feature is under discussion, it's not implemented yet.
Any info about, when this feature will be available?
| gharchive/issue | 2021-06-10T02:27:01 | 2025-04-01T04:33:28.300674 | {
"authors": [
"AndreyPatseiko",
"hdz-666",
"pissang",
"plainheart"
],
"repo": "apache/echarts",
"url": "https://github.com/apache/echarts/issues/15117",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1163874216 | [Bug] 数据量少的时候起点和终点会有很多留白
Version
5.2.1
Link to Minimal Reproduction
No response
Steps to Reproduce
数据量很少的时候就会这样
Current Behavior
数据量少的时候起点和终点会有很多留白
Expected Behavior
希望能像highchart一样处理,在起点向前移,终点往后移
Environment
- OS:
- Browser:
- Framework:
Any additional comments?
No response
设置 xAxis.boundaryGap 为 false.
设置 xAxis.boundaryGap 为 false.
感谢
| gharchive/issue | 2022-03-09T12:25:15 | 2025-04-01T04:33:28.305939 | {
"authors": [
"aninchiron",
"plainheart"
],
"repo": "apache/echarts",
"url": "https://github.com/apache/echarts/issues/16658",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2457285553 | [Bug] textStyle is invalid when formatter and richText renderMode are being used in the tooltip
Version
5.5.0
Link to Minimal Reproduction
PYBwLglsB2AEC8sDeAoWsA2BTA5l6AJgFzJrqxigkBEARsGJQLbUA0Z6BAhmFyQNrVA8IaBEFWoBdMgF926SsAyQQJVOQoAnCDjzqA8tBrQYWNh1hcMAdy4BPAM4BlABbBLAYRhh8YEmHUBXLFlyEGA7CEgYEgAzf2gAY0joAApwAEpSNXR1LDB_dTh
Steps to Reproduce
use richText renderMode
use formatter
add textStyle
tooltip: {
renderMode: "richText",
formatter: Data: {c},
textStyle: {
fontSize: 60,
fontWeight: "500",
fontFamily: "My-Custom-Font",
},
},
Current Behavior
Expected Behavior
Environment
- OS: macOS Ventura
- Browser:
- Framework: React / React native
Any additional comments?
I have used the same chart option for tooltip in react native framework also, but the result is same as web (as mentioned above).
Do you really need to use renderMode:'richText' ?
formatter and textStyle work fine together without it.
@helgasoft
1 - Yes i need to to use 'richText' for react native frame work, because html option is not working in mobile. If you any suggestion on this please do share, i will definitely give a try.
2 - 'formatter and textStyle work together fine without it' - Yes it is but as mentioned above fontSize, fontWeight and fontFamily are not implementing on the tool tip text while using formatter and textStyle together.
One more request if you have any suggestions regarding the above points, please do share it would be highly appreciated.
https://github.com/apache/echarts/issues/19966#issuecomment-2538259499
@souravpany
| gharchive/issue | 2024-08-09T07:23:02 | 2025-04-01T04:33:28.312342 | {
"authors": [
"XHL-study",
"helgasoft",
"souravpany"
],
"repo": "apache/echarts",
"url": "https://github.com/apache/echarts/issues/20240",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2577718533 | [Feature] 自行车
What problem does this feature solve?
暗室逢灯
What does the proposed API look like?
阿斯顿发到付十分大方撒旦法所发生的发撒说法大萨达发达
S这是打发斯蒂芬
| gharchive/issue | 2024-10-10T06:04:50 | 2025-04-01T04:33:28.313598 | {
"authors": [
"purpletastes"
],
"repo": "apache/echarts",
"url": "https://github.com/apache/echarts/issues/20401",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2496839338 | [FLINK-36188] Fix disable buffer flush lose efficacy
The user lookup joins the hbase table, adds 1 to the col value, and writes it back to hbase
@Test
void testTableSinkDisabledBufferFlush() throws Exception {
StreamExecutionEnvironment execEnv = StreamExecutionEnvironment.getExecutionEnvironment();
StreamTableEnvironment tEnv = StreamTableEnvironment.create(execEnv, streamSettings);
tEnv.executeSql(
"CREATE TABLE hTableForSink ("
+ " rowkey INT PRIMARY KEY NOT ENFORCED,"
+ " family1 ROW<col1 INT>"
+ ") WITH ("
+ " 'connector' = 'hbase-2.2',"
+ " 'sink.buffer-flush.max-size' = '0',"
+ " 'sink.buffer-flush.max-rows' = '0',"
+ " 'table-name' = '"
+ TEST_TABLE_6
+ "',"
+ " 'zookeeper.quorum' = '"
+ getZookeeperQuorum()
+ "'"
+ ")");
String insert = "INSERT INTO hTableForSink VALUES(1, ROW(1))";
tEnv.executeSql(insert).await();
tEnv.executeSql(
"CREATE VIEW user_click AS "
+ " SELECT user_id, proctime() AS proc_time"
+ " FROM ( "
+ " VALUES(1), (1), (1), (1), (1)"
+ " ) AS t (user_id);");
tEnv.executeSql(
"INSERT INTO hTableForSink SELECT "
+ " user_id as rowkey,"
+ " ROW(CAST(family1.col1 + 1 AS INT))"
+ " FROM user_click INNER JOIN hTableForSink"
+ " FOR SYSTEM_TIME AS OF user_click.proc_time"
+ " ON hTableForSink.rowkey = user_click.user_id;");
tEnv.executeSql(
"CREATE TABLE hTableForQuery ("
+ " rowkey INT PRIMARY KEY NOT ENFORCED,"
+ " family1 ROW<col1 INT>"
+ ") WITH ("
+ " 'connector' = 'hbase-2.2',"
+ " 'table-name' = '"
+ TEST_TABLE_6
+ "',"
+ " 'zookeeper.quorum' = '"
+ getZookeeperQuorum()
+ "'"
+ ")");
String query = "SELECT rowkey, family1.col1 FROM hTableForQuery";
TableResult firstResult = tEnv.executeSql(query);
List<Row> firstResults = CollectionUtil.iteratorToList(firstResult.collect());
String firstExpected = "+I[1, 6]";
TestBaseUtils.compareResultAsText(firstResults, firstExpected);
}
test failed
org.junit.ComparisonFailure: Different elements in arrays: expected 1 elements and received 1
expected: [+I[1, 6]]
received: [+I[1, 2]] expected:<+I[1, [6]]> but was:<+I[1, [2]]>
Expected :+I[1, 6]
Actual :+I[1, 2]
Thanks for submitting this fix! Pls. check the test the change broke. And since new changes will be necessary anyways, pls. reword the commit msg as well to contain the jira key to make "Title Validator" happy.
Thanks for submitting this fix! Pls. check the test the change broke. And since new changes will be necessary anyways, pls. reword the commit msg as well to contain the jira key to make "Title Validator" happy.
done,Could you please review it again? thank you @ferenc-csaky
@MartijnVisser can you help to trigger a CI run, thanks~
| gharchive/pull-request | 2024-08-30T10:25:03 | 2025-04-01T04:33:28.317583 | {
"authors": [
"MOBIN-F",
"ferenc-csaky"
],
"repo": "apache/flink-connector-hbase",
"url": "https://github.com/apache/flink-connector-hbase/pull/49",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1962022279 | [FLINK-31859] Update CycloneDX to 2.7.9
What is the purpose of the change
The issue with current cylcone version is that it generates lots of exceptions in logs like
org.codehaus.plexus.util.xml.pull.XmlPullParserException: UTF-8 BOM plus xml decl of ISO-8859-1 is incompatible (position: START_DOCUMENT seen <?xml version="1.0" encoding="ISO-8859-1"... @1:42)
at org.codehaus.plexus.util.xml.pull.MXParser.parseXmlDeclWithVersion (MXParser.java:3423)
at org.codehaus.plexus.util.xml.pull.MXParser.parseXmlDecl (MXParser.java:3345)
at org.codehaus.plexus.util.xml.pull.MXParser.parsePI (MXParser.java:3197)
at org.codehaus.plexus.util.xml.pull.MXParser.parseProlog (MXParser.java:1828)
...
at org.apache.maven.model.io.xpp3.MavenXpp3Reader.read (MavenXpp3Reader.java:627)
at org.cyclonedx.maven.BaseCycloneDxMojo.readPom (BaseCycloneDxMojo.java:759)
at org.cyclonedx.maven.BaseCycloneDxMojo.readPom (BaseCycloneDxMojo.java:746)
at org.cyclonedx.maven.BaseCycloneDxMojo.retrieveParentProject (BaseCycloneDxMojo.java:694)
...
The PR is aiming to fix this
Brief change log
pom.xml
Verifying this change
This change is a trivial rework / code cleanup without any test coverage.
Does this pull request potentially affect one of the following parts:
Dependencies (does it add or upgrade a dependency): (yes )
The public API, i.e., is any changes to the CustomResourceDescriptors: ( no)
Core observer or reconciler logic that is regularly executed: ( no)
Documentation
Does this pull request introduce a new feature? ( no)
If yes, how is the feature documented? (not applicable )
Please add the contents of this PR to https://github.com/apache/flink-kubernetes-operator/pull/691 to avoid fragmentation
@gyfora this is not related to jdk17
this is a separate issue which i just noticed while comiling locally
ok, I am just not a big fan of these tiny PRs/commits. It could have been simply part of the other PR as you noticed and fixed it while you worked on it.
Opening one-liner PRs puts a strain on CI and also takes up everyone's time. Simply not worth it.
ok, i see,
i will put them together
closing since now it is a part of https://github.com/apache/flink-kubernetes-operator/pull/691
| gharchive/pull-request | 2023-10-25T18:47:18 | 2025-04-01T04:33:28.323233 | {
"authors": [
"gyfora",
"snuyanzin"
],
"repo": "apache/flink-kubernetes-operator",
"url": "https://github.com/apache/flink-kubernetes-operator/pull/693",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
725124878 | [FLINK-19624][table-planner-blink] Update deadlock break-up algorithm to cover more cases
What is the purpose of the change
Current deadlock breakup algorithm in Blink planner fails to cover some cases. See FLINK-19624 and the appendix of the design doc for more information.
We're going to introduce a new deadlock breakup algorithm to solve this problem. This algorithm is also used in the upcoming multiple input node creation algorithm.
Brief change log
Introduce a new deadlock breakup algorithm.
Verifying this change
This change added tests and can be verified as follows: Run the added test cases.
Does this pull request potentially affect one of the following parts:
Dependencies (does it add or upgrade a dependency): no
The public API, i.e., is any changed class annotated with @Public(Evolving): no
The serializers: no
The runtime per-record code paths (performance sensitive): no
Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
The S3 file system connector: no
Documentation
Does this pull request introduce a new feature? no
If yes, how is the feature documented? not applicable
@flinkbot run azure
| gharchive/pull-request | 2020-10-20T02:13:22 | 2025-04-01T04:33:28.329407 | {
"authors": [
"TsReaper",
"godfreyhe"
],
"repo": "apache/flink",
"url": "https://github.com/apache/flink/pull/13692",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
850587152 | [hotfix][docs] Fix max parallelism behavior for Reactive Mode
With FLINK-21844, Reactive Mode now has the same behavior
when the same max parallelism as the "default" mode.
This also includes minor typos/ fixes.
Signed-off-by: austin ce austin.cawley@gmail.com
What is the purpose of the change
With FLINK-21844, Reactive Mode now has the same behavior when the same max parallelism as the "default" mode.
This also includes minor typos/ fixes.
Brief change log
Update Elastic Scaling page with description of new behavior
Verifying this change
This change is a trivial rework / code cleanup without any test coverage.
Does this pull request potentially affect one of the following parts:
Dependencies (does it add or upgrade a dependency): no
The public API, i.e., is any changed class annotated with @Public(Evolving): no
The serializers: no
The runtime per-record code paths (performance sensitive): no
Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
The S3 file system connector: no
Documentation
Does this pull request introduce a new feature? no
If yes, how is the feature documented? docs
@rmetzger can you have a look at this when you have a chance?
@flinkbot attention rmetzger
(?)
Thanks for merging Chesnay & Thanks for fixing Austin. I think I ignored these notifications because I thought it is the PR fixing the test instability in ReactiveModeITCase. I'll pay more attention in the future!
No worries! Should've updated these with the original ticket
| gharchive/pull-request | 2021-04-05T19:03:18 | 2025-04-01T04:33:28.334610 | {
"authors": [
"austince",
"rmetzger"
],
"repo": "apache/flink",
"url": "https://github.com/apache/flink/pull/15488",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
338831941 | [FLINK-9755][network] forward exceptions in RemoteInputChannel#notifyBufferAvailable() to the responsible thread
What is the purpose of the change
Exceptions in RemoteInputChannel#notifyBufferAvailable(), e.g. state checks, were swallowed inside LocalBufferPool#recycle() and neither logged nor otherwise processed and may have lead to stalling processes waiting for a notification that never comes.
Please note that this PR builds upon #6271 which also touched the unit tests which we change here. @tillrohrmann, @zhijiangW can you have a look at this cleanup of LocalBufferPool#recycle()?
Brief change log
cleanup of LocalBufferPool#recycle() also clarifying the contract of BufferListener#notifyBufferAvailable() which should recycle the given buffer (one implementation already did that; RemoteInputChannel did not)
forward exceptions in RemoteInputChannel#notifyBufferAvailable() to the responsible channel and recycle the given buffer in that case
Verifying this change
This change added tests and can be verified as follows:
added RemoteInputChannelTest#testFailureInNotifyBufferAvailable()
Does this pull request potentially affect one of the following parts:
Dependencies (does it add or upgrade a dependency): no
The public API, i.e., is any changed class annotated with @Public(Evolving): no
The serializers: no
The runtime per-record code paths (performance sensitive): no
Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: no
The S3 file system connector: no
Documentation
Does this pull request introduce a new feature? no
If yes, how is the feature documented? JavaDocs
Thanks for fixing this potential bug.
It makes sense to handle exception during notifyBufferAvailable on listener side. Just some thoughts on my side above. :)
thanks for the review - I changed the code as requested
actually, I forgot to adapt the unit test which relied on the previous order of the calls and assumed that the buffer got recycled. With the changes from the fixup, the buffer is not recycled by the failure directly but rather during error handling which will eventually release the channel's resources.
-> this is now fixed as well
rebased to solve the merge conflict (auto-solved by git though)
Thanks for the review, merging...
| gharchive/pull-request | 2018-07-06T07:19:44 | 2025-04-01T04:33:28.341987 | {
"authors": [
"NicoK",
"zhijiangW"
],
"repo": "apache/flink",
"url": "https://github.com/apache/flink/pull/6272",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
464849390 | [FLINK-13107][table-planner-blink] Copy TableApi IT and UT to Blink planner
What is the purpose of the change
The issue aims to copy the testcases in the following packages from flink-planner and original blink to Blink-planner:
org.apache.flink.table.api.batch.table
org.apache.flink.table.api.stream.table
org.apache.flink.table.runtime.batch.table
org.apache.flink.table.runtime.stream.table
Brief change log
Copy UT and ITCase of TableApi
Fix some Bugs, commit message contains detail information about those Bugs
Verifying this change
UT, ITCase
Does this pull request potentially affect one of the following parts:
Dependencies (does it add or upgrade a dependency): (no)
The public API, i.e., is any changed class annotated with @Public(Evolving): (no)
The serializers: (no)
The runtime per-record code paths (performance sensitive): (no)
Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
The S3 file system connector: (no)
Documentation
Does this pull request introduce a new feature? (no)
If yes, how is the feature documented? (not applicable)
We need copy all org.apache.flink.table.api.*? I think we have these duplicate plan cases in org.apache.flink.table.table.*? Why not put cases to org.apache.flink.table.table.* to be consistent with previous SQL plan tests?
What are the implications on the test times when we copy so many tests?
According to this: https://travis-ci.org/beyond1920/flink/builds/557233740, the test time increases less than 2 mins.
travis passed here: https://travis-ci.org/beyond1920/flink/builds/557233740
I'm merging this
| gharchive/pull-request | 2019-07-06T12:23:58 | 2025-04-01T04:33:28.348759 | {
"authors": [
"JingsongLi",
"KurtYoung",
"beyond1920"
],
"repo": "apache/flink",
"url": "https://github.com/apache/flink/pull/9006",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
383288365 | GEODE-6079: Attempt to sync filesystem before searching for suspect s…
…trings.
Thank you for submitting a contribution to Apache Geode.
In order to streamline the review of the contribution we ask you
to ensure the following steps have been taken:
For all changes:
[ ] Is there a JIRA ticket associated with this PR? Is it referenced in the commit message?
[ ] Has your PR been rebased against the latest commit within the target branch (typically develop)?
[ ] Is your initial contribution a single, squashed commit?
[ ] Does gradlew build run cleanly?
[ ] Have you written or updated unit tests to verify your changes?
[ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0?
Note:
Please ensure that once the PR is submitted, you check travis-ci for build issues and
submit an update to your PR as soon as possible. If you need help, please send an
email to dev@geode.apache.org.
Because we run tests in parallel, and because they all seem to be writing to the same suspect file, it may well be that either the DUnit run will timeout with the added time of the sync(), and/or the call will throw an IOException after timing out with frequency that makes this change pointless. The precheckin will be interesting to review.
| gharchive/pull-request | 2018-11-21T20:52:46 | 2025-04-01T04:33:28.353427 | {
"authors": [
"PurelyApplied"
],
"repo": "apache/geode",
"url": "https://github.com/apache/geode/pull/2892",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
388536752 | GEODE-6164: CacheClientProxy's closeSocket should be called atomically
Thank you for submitting a contribution to Apache Geode.
@jhuynh1 @Bill
In order to streamline the review of the contribution we ask you
to ensure the following steps have been taken:
For all changes:
[ ] Is there a JIRA ticket associated with this PR? Is it referenced in the commit message?
[ ] Has your PR been rebased against the latest commit within the target branch (typically develop)?
[ ] Is your initial contribution a single, squashed commit?
[ ] Does gradlew build run cleanly?
[ ] Have you written or updated unit tests to verify your changes?
[ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0?
Note:
Please ensure that once the PR is submitted, you check travis-ci for build issues and
submit an update to your PR as soon as possible. If you need help, please send an
email to dev@geode.apache.org.
Looks good to me now.
| gharchive/pull-request | 2018-12-07T07:28:47 | 2025-04-01T04:33:28.357998 | {
"authors": [
"Bill",
"gesterzhou"
],
"repo": "apache/geode",
"url": "https://github.com/apache/geode/pull/2972",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
916700809 | [GOBBLIN-1466] Make meters shared between DagManagerThreads
Dear Gobblin maintainers,
Please accept this PR. I understand that it will not be reviewed until I have checked off all the steps below!
JIRA
[x] My PR addresses the following Gobblin JIRA issues and references them in the PR title. For example, "[GOBBLIN-XXX] My Gobblin PR"
https://issues.apache.org/jira/browse/GOBBLIN-1466
Description
[x] Here are some details about my PR, including screenshots (if applicable):
Each of the DagManagerThreads had its own meter (but with the same names) previously, which causing confusing results when reading them, instead there should only be meters shared across the threads.
For service level meter, created one meter in the DagManager and passed it to each thread instead
For group level meters, created a static map keyed by group name to be shared by all threads
Tests
[x] My PR adds the following unit tests OR does not need testing for this extremely good reason:
Tested locally by printing meter values in each thread
Commits
[x] My commits all reference JIRA issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message":
Subject is separated from body by a blank line
Subject is limited to 50 characters
Subject does not end with a period
Subject uses the imperative mood ("add", not "adding")
Body wraps at 72 characters
Body explains "what" and "why", not "how"
Codecov Report
Merging #3306 (048841f) into master (bf37c76) will decrease coverage by 3.51%.
The diff coverage is 75.00%.
@@ Coverage Diff @@
## master #3306 +/- ##
============================================
- Coverage 46.50% 42.98% -3.52%
+ Complexity 10041 9311 -730
============================================
Files 2041 2041
Lines 79349 79356 +7
Branches 8845 8846 +1
============================================
- Hits 36898 34112 -2786
- Misses 39020 42007 +2987
+ Partials 3431 3237 -194
Impacted Files
Coverage Δ
...blin/service/modules/orchestration/DagManager.java
72.64% <75.00%> (-0.34%)
:arrow_down:
.../org/apache/gobblin/util/filters/HiddenFilter.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...g/apache/gobblin/cluster/HelixMessageSubTypes.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...gobblin/runtime/mapreduce/GobblinOutputFormat.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...obblin/compaction/source/CompactionFailedTask.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...n/cluster/event/ClusterManagerShutdownRequest.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...in/compaction/action/CompactionCompleteAction.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...n/compaction/mapreduce/orc/OrcKeyDedupReducer.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...action/audit/KafkaAuditCountHttpClientFactory.java
0.00% <0.00%> (-100.00%)
:arrow_down:
...askStateCollectorServiceHiveRegHandlerFactory.java
0.00% <0.00%> (-100.00%)
:arrow_down:
... and 138 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update bf37c76...048841f. Read the comment docs.
rebasing to master to fix travis
+1.
| gharchive/pull-request | 2021-06-09T21:52:00 | 2025-04-01T04:33:28.380975 | {
"authors": [
"codecov-commenter",
"jack-moseley",
"sv2000"
],
"repo": "apache/gobblin",
"url": "https://github.com/apache/gobblin/pull/3306",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
964405608 | [GOBBLIN-1509] Announce flow failure on DagManager::addDag error
Dear Gobblin maintainers,
Please accept this PR. I understand that it will not be reviewed until I have checked off all the steps below!
JIRA
[ ] My PR addresses the following Gobblin JIRA issues and references them in the PR title. For example, "[GOBBLIN-XXX] My Gobblin PR"
[ ]
https://issues.apache.org/jira/browse/GOBBLIN-1509
Description
[ ] Here are some details about my PR, including screenshots (if applicable):
Announce flow failure on DagManager::addDag error
Additionally, migrate Orchestrator overall away from deprecated EventSubmitter::getTimingEvent factory method.
Presently, addDag failure leaves the flow marooned in the COMPILED state, as the warranted FLOW_FAILED event is never sent. Particularly insidious is that scheduled flows with their execution stuck in COMPILED miss their next execution, unless flow.allowConcurrentExecutions is set. Thus the scheduled flow is stuck in its entirety, not merely a single execution.
One observed cause of addDag failure is when the DagStateStore is backed by a replicated DB (e.g. MySqlDagStateStore) that just switched leaders. Cached connections in the pool may suddenly point to a read-only follower unable to DagStateStore::writeCheckpoint.
Tests
[ ] My PR adds the following unit tests OR does not need testing for this extremely good reason:
Manual testing while running GaaS locally:
I added a flow to runImmediatly, which I soon after observed COMPLETE
then I locally patched DagManager::addDag to mimic the motivating failure scenario by invariably throwing an IOException.
I again added the same flow as 1.) (adjusted only to bear a unique name and target location) to runImmediately
I observed that second flow as FAILED with the shimmed exception conveyed in the message.
a. patched DagManager:
--- a/gobblin-service/src/main/java/org/apache/gobblin/service/modules/orchestration/DagManager.java
+++ b/gobblin-service/src/main/java/org/apache/gobblin/service/modules/orchestration/DagManager.java
@@ -263,7 +263,8 @@ public class DagManager extends AbstractIdleService {
synchronized void addDag(Dag<JobExecutionPlan> dag, boolean persist, boolean setStatus) throws IOException {
if (persist) {
//Persist the dag
- this.dagStateStore.writeCheckpoint(dag);
+ throw new IOException("No, I won't add the DAG, dawg!");
+ // this.dagStateStore.writeCheckpoint(dag);
}
int queueId = DagManagerUtils.getDagQueueId(dag, this.numThreads);
// Add the dag to the specific queue determined by flowExecutionId
b. submitted flow:
{
"id": {
"flowName": "test005",
"flowGroup": "testKip"
},
"templateUris": "FS:///",
"properties": {
"gobblin.flow.sourceIdentifier": "<<redacted-source>>",
"gobblin.flow.destinationIdentifier": "<<(same) redacted-source>>",
"user.to.proxy": "gobblintest",
"gobblin.flow.input.dataset.descriptor.path": "<<redacted-path>>",
"gobblin.flow.output.dataset.descriptor.path": "/tmp/gaas-testing/kip/test005",
"gobblin.flow.input.dataset.descriptor.partition.type": "none",
"gobblin.flow.output.dataset.descriptor.partition.type": "none",
"gobblin.copy.simulate": "false",
"flow.applyRetention": "false",
"dataset.datetimePattern": "yyyy/MM/dd",
"copy.date.pattern": "yyyy/MM/dd"
},
"schedule": {
"cronSchedule": "0 0 8 * * ? *",
"runImmediately": true
}
}
c. observed (FAILED) status:
curli -k --dv-auth SELF "https://localhost:6956/sharedgobblinservice/flowexecutions?q=latestFlowExecution&flowId=(flowGroup:testKip,flowName:test005)" -X GET -H 'X-RestLi-Protocol-Version: 2.0.0'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 361 0 361 0 0 44 0 --:--:-- 0:00:08 --:--:-- 74
{
"elements": [
{
"id": {
"flowGroup": "testKip",
"flowExecutionId": 1628209427967,
"flowName": "test005"
},
"message": "Failed to add Job Execution Plan due to: No, I won't add the DAG, dawg!",
"executionStatistics": {
"executionEndTime": 1628209428045,
"executionStartTime": 1628209427967
},
"jobStatuses": [],
"executionStatus": "FAILED"
}
],
"paging": {
"count": 10,
"start": 0,
"links": []
}
}
Commits
[ ] My commits all reference JIRA issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message":
Subject is separated from body by a blank line
Subject is limited to 50 characters
Subject does not end with a period
Subject uses the imperative mood ("add", not "adding")
Body wraps at 72 characters
Body explains "what" and "why", not "how"
Codecov Report
Merging #3357 (41ec0a0) into master (de94515) will decrease coverage by 3.44%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #3357 +/- ##
============================================
- Coverage 46.43% 42.99% -3.45%
+ Complexity 10120 1940 -8180
============================================
Files 2052 394 -1658
Lines 79656 16886 -62770
Branches 8896 2080 -6816
============================================
- Hits 36991 7260 -29731
+ Misses 39226 8826 -30400
+ Partials 3439 800 -2639
Impacted Files
Coverage Δ
...in/java/org/apache/gobblin/cluster/HelixUtils.java
32.23% <0.00%> (-2.48%)
:arrow_down:
...a/org/apache/gobblin/cluster/GobblinHelixTask.java
58.06% <0.00%> (-2.16%)
:arrow_down:
.../apache/gobblin/util/test/RetentionTestHelper.java
...gobblin/runtime/mapreduce/GobblinOutputFormat.java
...e/gobblin/runtime/api/GobblinInstanceLauncher.java
...g/apache/gobblin/service/monitoring/JobStatus.java
...va/org/apache/gobblin/publisher/DataPublisher.java
...org/apache/gobblin/source/extractor/Extractor.java
.../apache/gobblin/http/ApacheHttpResponseStatus.java
...blin/data/management/copy/RecursivePathFinder.java
... and 1649 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update de94515...41ec0a0. Read the comment docs.
| gharchive/pull-request | 2021-08-09T21:30:26 | 2025-04-01T04:33:28.406251 | {
"authors": [
"codecov-commenter",
"phet"
],
"repo": "apache/gobblin",
"url": "https://github.com/apache/gobblin/pull/3357",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2482454348 | [#4662] improve(IT): Add ranger authorization Hive E2E test
What changes were proposed in this pull request?
Use an isolated class loader to create an authorization Ranger plugin.
Add authorization for Hive E2E tests using Ranger via Gravitino RESTful access control.
Copy the authorization ranger jar to the distribution/package/authorizations/ranger/libs.
Add Apache Ranger to license.bin
Why are the changes needed?
#4662
Does this PR introduce any user-facing change?
N/A
How was this patch tested?
CI
@jerryshao Please help me review this PR, thanks.
| gharchive/pull-request | 2024-08-23T06:58:11 | 2025-04-01T04:33:28.409811 | {
"authors": [
"xunliu"
],
"repo": "apache/gravitino",
"url": "https://github.com/apache/gravitino/pull/4651",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2485068850 | GUACAMOLE-1866: Clean up style of recent connection removal button.
This change restyles the recent connection removal button such that:
It's a gray-colored "X" that becomes black when hovered.
The button itself is not visible unless the associated recent connection is being hovered.
Closing in favor of broader UI cleanup around the newly configurable "Recent Connections" section.
| gharchive/pull-request | 2024-08-25T05:33:50 | 2025-04-01T04:33:28.411282 | {
"authors": [
"mike-jumper"
],
"repo": "apache/guacamole-client",
"url": "https://github.com/apache/guacamole-client/pull/1008",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
506966656 | HDDS-2194. Replication of Container fails with "Only closed containers could be exported"
https://issues.apache.org/jira/browse/HDDS-2194
The issue is because when the Replication Manager is considering to replicate containers which are under replicated, it considers replicas in QuasiClosed/Closed. Whereas in Datanode we have a check of Closed. This has caused the issue. So, that is why we see IllegealStateException in the logs.
Test failures are not related.
Thanks @bharatviswa504 for the contribution and @adoroszlai for review.
| gharchive/pull-request | 2019-10-15T02:29:57 | 2025-04-01T04:33:28.413002 | {
"authors": [
"bharatviswa504",
"nandakumar131"
],
"repo": "apache/hadoop-ozone",
"url": "https://github.com/apache/hadoop-ozone/pull/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
504997511 | YARN-7243. Moving logging APIs over to slf4j in hadoop-yarn-server-resourcemanager.
Backport https://issues.apache.org/jira/browse/YARN-7243 to branch-3.2 to fix compile failure caused by https://issues.apache.org/jira/browse/YARN-9873
All the conflicts are import sections.
Ran {{mvn install -DskipTests}} successfully.
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Comment
0
reexec
531
Docker mode activated.
_ Prechecks _
+1
dupname
6
No case conflicting files found.
+1
@author
0
The patch does not contain any @author tags.
+1
test4tests
0
The patch appears to include 74 new or modified test files.
_ branch-3.2 Compile Tests _
0
mvndep
17
Maven dependency ordering for branch
-1
mvninstall
399
root in branch-3.2 failed.
-1
compile
367
root in branch-3.2 failed.
-0
checkstyle
207
The patch fails to run checkstyle in root
-1
mvnsite
29
hadoop-yarn-server-resourcemanager in branch-3.2 failed.
-1
shadedclient
550
branch has errors when building and testing our client artifacts.
+1
javadoc
83
branch-3.2 passed
0
spotbugs
103
Used deprecated FindBugs config; considering switching to SpotBugs.
-1
findbugs
27
hadoop-yarn-server-resourcemanager in branch-3.2 failed.
_ Patch Compile Tests _
0
mvndep
18
Maven dependency ordering for patch
+1
mvninstall
91
the patch passed
+1
compile
906
the patch passed
-1
javac
906
root generated 517 new + 804 unchanged - 8 fixed = 1321 total (was 812)
-0
checkstyle
219
The patch fails to run checkstyle in root
+1
mvnsite
134
the patch passed
+1
whitespace
0
The patch has no whitespace issues.
+1
shadedclient
707
patch has no errors when building and testing our client artifacts.
+1
javadoc
114
the patch passed
+1
findbugs
227
the patch passed
_ Other Tests _
+1
unit
536
hadoop-common in the patch passed.
-1
unit
4292
hadoop-yarn-server-resourcemanager in the patch failed.
+1
asflicense
49
The patch does not generate ASF License warnings.
9655
Reason
Tests
Failed junit tests
hadoop.yarn.server.resourcemanager.metrics.TestSystemMetricsPublisherForV2
hadoop.yarn.server.resourcemanager.metrics.TestCombinedSystemMetricsPublisher
Subsystem
Report/Notes
Docker
Client=19.03.3 Server=19.03.3 base: https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/1634
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle
uname
Linux c9a61fad2458 4.15.0-58-generic #64-Ubuntu SMP Tue Aug 6 11:12:41 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
personality/hadoop.sh
git revision
branch-3.2 / ae8ca70
Default Java
1.8.0_222
mvninstall
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-mvninstall-root.txt
compile
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-compile-root.txt
checkstyle
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out//home/jenkins/jenkins-slave/workspace/hadoop-multibranch_PR-1634/out/maven-branch-checkstyle-root.txt
mvnsite
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-mvnsite-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
findbugs
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
javac
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/diff-compile-javac-root.txt
checkstyle
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out//home/jenkins/jenkins-slave/workspace/hadoop-multibranch_PR-1634/out/maven-patch-checkstyle-root.txt
unit
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
Test Results
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/testReport/
Max. process+thread count
1532 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: .
Console output
https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/console
versions
git=2.7.4 maven=3.3.9 findbugs=3.1.0-RC1
Powered by
Apache Yetus 0.10.0 http://yetus.apache.org
This message was automatically generated.
The test failures are not related to the patch. Merging this into branch-3.2.
| gharchive/pull-request | 2019-10-10T02:39:59 | 2025-04-01T04:33:28.443369 | {
"authors": [
"aajisaka",
"hadoop-yetus"
],
"repo": "apache/hadoop",
"url": "https://github.com/apache/hadoop/pull/1634",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
655719130 | MAPREDUCE-7284. TestCombineFileInputFormat#testMissingBlocks fails
JIRA: https://issues.apache.org/jira/browse/MAPREDUCE-7284
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Comment
+0 :ok:
reexec
0m 31s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 1 new or modified test files.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
22m 2s
trunk passed
+1 :green_heart:
compile
0m 36s
trunk passed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04
+1 :green_heart:
compile
0m 33s
trunk passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09
+1 :green_heart:
checkstyle
0m 30s
trunk passed
+1 :green_heart:
mvnsite
0m 32s
trunk passed
+1 :green_heart:
shadedclient
15m 11s
branch has no errors when building and testing our client artifacts.
-1 :x:
javadoc
0m 26s
hadoop-mapreduce-client-jobclient in trunk failed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04.
+1 :green_heart:
javadoc
0m 24s
trunk passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09
+0 :ok:
spotbugs
0m 46s
Used deprecated FindBugs config; considering switching to SpotBugs.
+1 :green_heart:
findbugs
0m 44s
trunk passed
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
0m 31s
the patch passed
+1 :green_heart:
compile
0m 30s
the patch passed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04
+1 :green_heart:
javac
0m 30s
the patch passed
+1 :green_heart:
compile
0m 27s
the patch passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09
+1 :green_heart:
javac
0m 27s
the patch passed
+1 :green_heart:
checkstyle
0m 20s
the patch passed
+1 :green_heart:
mvnsite
0m 29s
the patch passed
+1 :green_heart:
whitespace
0m 0s
The patch has no whitespace issues.
+1 :green_heart:
shadedclient
14m 9s
patch has no errors when building and testing our client artifacts.
-1 :x:
javadoc
0m 21s
hadoop-mapreduce-client-jobclient in the patch failed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04.
+1 :green_heart:
javadoc
0m 17s
the patch passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09
+1 :green_heart:
findbugs
0m 45s
the patch passed
_ Other Tests _
+1 :green_heart:
unit
123m 21s
hadoop-mapreduce-client-jobclient in the patch passed.
-1 :x:
asflicense
0m 42s
The patch generated 1 ASF License warnings.
185m 20s
Subsystem
Report/Notes
Docker
ClientAPI=1.40 ServerAPI=1.40 base: https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/2136
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle
uname
Linux 65816539f21c 4.15.0-60-generic #67-Ubuntu SMP Thu Aug 22 16:55:30 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
personality/hadoop.sh
git revision
trunk / b9fa5e0182c
Default Java
Private Build-1.8.0_252-8u252-b09-1~18.04-b09
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_252-8u252-b09-1~18.04-b09
javadoc
https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/branch-javadoc-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient-jdkUbuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04.txt
javadoc
https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/patch-javadoc-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient-jdkUbuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04.txt
Test Results
https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/testReport/
asflicense
https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/patch-asflicense-problems.txt
Max. process+thread count
1330 (vs. ulimit of 5500)
modules
C: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient U: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient
Console output
https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/console
versions
git=2.17.1 maven=3.6.0 findbugs=3.1.0-RC1
Powered by
Apache Yetus 0.12.0 https://yetus.apache.org
This message was automatically generated.
Thank you @iwasakims
| gharchive/pull-request | 2020-07-13T09:58:44 | 2025-04-01T04:33:28.474157 | {
"authors": [
"aajisaka",
"hadoop-yetus"
],
"repo": "apache/hadoop",
"url": "https://github.com/apache/hadoop/pull/2136",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1329141741 | HADOOP-18392. Propagate vectored s3a input stream stats to file system stats.
Description of PR
Propagate vectored s3a input stream stats to file system stats.
How was this patch tested?
Re-ran the vectored io tests using us-east-1 endpoint.
For code changes:
[x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')?
[x] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation?
[ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0?
[ ] If applicable, have you updated the LICENSE, LICENSE-binary, NOTICE-binary files?
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 59s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 2 new or modified test files.
_ trunk Compile Tests _
+0 :ok:
mvndep
15m 14s
Maven dependency ordering for branch
+1 :green_heart:
mvninstall
25m 24s
trunk passed
+1 :green_heart:
compile
23m 14s
trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
compile
21m 2s
trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
checkstyle
4m 24s
trunk passed
+1 :green_heart:
mvnsite
3m 47s
trunk passed
+1 :green_heart:
javadoc
2m 54s
trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javadoc
2m 34s
trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
spotbugs
5m 3s
trunk passed
+1 :green_heart:
shadedclient
22m 12s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+0 :ok:
mvndep
0m 33s
Maven dependency ordering for patch
+1 :green_heart:
mvninstall
1m 45s
the patch passed
+1 :green_heart:
compile
22m 32s
the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javac
22m 32s
the patch passed
+1 :green_heart:
compile
21m 1s
the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
javac
21m 1s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
4m 22s
the patch passed
+1 :green_heart:
mvnsite
3m 37s
the patch passed
+1 :green_heart:
javadoc
2m 50s
the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javadoc
2m 39s
the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
spotbugs
5m 12s
the patch passed
+1 :green_heart:
shadedclient
22m 33s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
18m 50s
hadoop-common in the patch passed.
+1 :green_heart:
unit
3m 34s
hadoop-aws in the patch passed.
+1 :green_heart:
asflicense
1m 35s
The patch does not generate ASF License warnings.
243m 3s
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/4704
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux 7114ac705ada 4.15.0-65-generic #74-Ubuntu SMP Tue Sep 17 17:06:04 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / fef7f5f6e5c79af4f16cd5a52af56317cf260a30
Default Java
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/1/testReport/
Max. process+thread count
1330 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws U: .
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/1/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 57s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 0s
codespell was not available.
+0 :ok:
detsecrets
0m 0s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 2 new or modified test files.
_ trunk Compile Tests _
+0 :ok:
mvndep
15m 30s
Maven dependency ordering for branch
+1 :green_heart:
mvninstall
25m 36s
trunk passed
+1 :green_heart:
compile
23m 20s
trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
compile
20m 45s
trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
checkstyle
4m 22s
trunk passed
+1 :green_heart:
mvnsite
3m 35s
trunk passed
+1 :green_heart:
javadoc
2m 57s
trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javadoc
2m 41s
trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
spotbugs
5m 5s
trunk passed
+1 :green_heart:
shadedclient
22m 17s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+0 :ok:
mvndep
0m 30s
Maven dependency ordering for patch
+1 :green_heart:
mvninstall
1m 47s
the patch passed
+1 :green_heart:
compile
22m 28s
the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javac
22m 28s
the patch passed
+1 :green_heart:
compile
20m 43s
the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
javac
20m 43s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
-0 :warning:
checkstyle
4m 9s
/results-checkstyle-root.txt
root: The patch generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0)
+1 :green_heart:
mvnsite
3m 44s
the patch passed
+1 :green_heart:
javadoc
2m 48s
the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javadoc
2m 41s
the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
spotbugs
5m 11s
the patch passed
+1 :green_heart:
shadedclient
22m 28s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
-1 :x:
unit
18m 55s
/patch-unit-hadoop-common-project_hadoop-common.txt
hadoop-common in the patch passed.
+1 :green_heart:
unit
3m 25s
hadoop-aws in the patch passed.
+1 :green_heart:
asflicense
1m 37s
The patch does not generate ASF License warnings.
242m 36s
Reason
Tests
Failed junit tests
hadoop.fs.contract.localfs.TestLocalFSContractVectoredRead
hadoop.fs.contract.rawlocal.TestRawLocalContractVectoredRead
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/2/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/4704
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux e9a142c077c3 4.15.0-65-generic #74-Ubuntu SMP Tue Sep 17 17:06:04 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / a65a68dda3b020b35b116cb97a59e5e1b09d98d3
Default Java
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/2/testReport/
Max. process+thread count
1859 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws U: .
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/2/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 56s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 2 new or modified test files.
_ trunk Compile Tests _
+0 :ok:
mvndep
15m 15s
Maven dependency ordering for branch
+1 :green_heart:
mvninstall
25m 25s
trunk passed
+1 :green_heart:
compile
23m 13s
trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
compile
20m 50s
trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
checkstyle
4m 23s
trunk passed
+1 :green_heart:
mvnsite
3m 47s
trunk passed
+1 :green_heart:
javadoc
2m 50s
trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javadoc
2m 36s
trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
spotbugs
5m 1s
trunk passed
+1 :green_heart:
shadedclient
22m 25s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+0 :ok:
mvndep
0m 35s
Maven dependency ordering for patch
+1 :green_heart:
mvninstall
1m 42s
the patch passed
+1 :green_heart:
compile
22m 29s
the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javac
22m 29s
the patch passed
+1 :green_heart:
compile
20m 54s
the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
javac
20m 54s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
4m 11s
the patch passed
+1 :green_heart:
mvnsite
3m 36s
the patch passed
+1 :green_heart:
javadoc
2m 51s
the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1
+1 :green_heart:
javadoc
2m 40s
the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
+1 :green_heart:
spotbugs
5m 16s
the patch passed
+1 :green_heart:
shadedclient
22m 32s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
18m 48s
hadoop-common in the patch passed.
+1 :green_heart:
unit
3m 22s
hadoop-aws in the patch passed.
+1 :green_heart:
asflicense
1m 36s
The patch does not generate ASF License warnings.
242m 33s
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/3/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/4704
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux c9c07ee729e1 4.15.0-65-generic #74-Ubuntu SMP Tue Sep 17 17:06:04 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / cad85597b88ae52871472fed31e0f30c3cc49c68
Default Java
Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/3/testReport/
Max. process+thread count
3159 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws U: .
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/3/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
| gharchive/pull-request | 2022-08-04T20:27:17 | 2025-04-01T04:33:28.576851 | {
"authors": [
"hadoop-yetus",
"mukund-thakur"
],
"repo": "apache/hadoop",
"url": "https://github.com/apache/hadoop/pull/4704",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1537931087 | YARN-11419. Simplify code and improve readability
Description of PR
Remove YarnException and IOException, because they are all subclasses of Throwable.
Use try-resource.
A non-empty string should be on the left side of the equal.
How was this patch tested?
Origin uts.
For code changes:
[x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')?
[ ] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation?
[ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0?
[ ] If applicable, have you updated the LICENSE, LICENSE-binary, NOTICE-binary files?
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 49s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
43m 29s
trunk passed
+1 :green_heart:
compile
0m 37s
trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
compile
0m 35s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
checkstyle
0m 35s
trunk passed
+1 :green_heart:
mvnsite
0m 39s
trunk passed
-1 :x:
javadoc
0m 40s
/branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 29s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
1m 7s
trunk passed
+1 :green_heart:
shadedclient
23m 19s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
0m 34s
the patch passed
+1 :green_heart:
compile
0m 28s
the patch passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
javac
0m 28s
the patch passed
+1 :green_heart:
compile
0m 24s
the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
javac
0m 24s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
-0 :warning:
checkstyle
0m 17s
/results-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client: The patch generated 2 new + 14 unchanged - 0 fixed = 16 total (was 14)
+1 :green_heart:
mvnsite
0m 28s
the patch passed
-1 :x:
javadoc
0m 22s
/patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 21s
the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
0m 56s
the patch passed
+1 :green_heart:
shadedclient
23m 8s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
28m 10s
hadoop-yarn-client in the patch passed.
+1 :green_heart:
asflicense
0m 41s
The patch does not generate ASF License warnings.
129m 37s
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5309
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux 29d2cd1531f3 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 15d3246fa4e299adf24dd9335998d886b117a0e9
Default Java
Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/1/testReport/
Max. process+thread count
577 (vs. ulimit of 5500)
modules
C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/1/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
Hi @smallzhongfeng ,
Could you be more specific with the title of PR / jira about what are you improving / cleaning up? Class name should be in it at least.
Hi @smallzhongfeng ,
Could you be more specific with the title of PR / jira about what are you improving / cleaning up? Class name should be in it at least.
Thank you for your reminder, could you help me review this? @szilard-nemeth @slfan1989 Thx !
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 58s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
48m 51s
trunk passed
+1 :green_heart:
compile
0m 44s
trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
compile
0m 40s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
checkstyle
0m 34s
trunk passed
+1 :green_heart:
mvnsite
0m 46s
trunk passed
-1 :x:
javadoc
0m 46s
/branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 32s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
1m 26s
trunk passed
-1 :x:
shadedclient
28m 36s
branch has errors when building and testing our client artifacts.
_ Patch Compile Tests _
-1 :x:
mvninstall
0m 23s
/patch-mvninstall-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
hadoop-yarn-client in the patch failed.
-1 :x:
compile
0m 22s
/patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
-1 :x:
javac
0m 22s
/patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
-1 :x:
compile
0m 22s
/patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkPrivateBuild-1.8.0_352-8u352-ga-1~20.04-b08.txt
hadoop-yarn-client in the patch failed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08.
-1 :x:
javac
0m 22s
/patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkPrivateBuild-1.8.0_352-8u352-ga-1~20.04-b08.txt
hadoop-yarn-client in the patch failed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08.
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
-0 :warning:
checkstyle
0m 20s
/buildtool-patch-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
The patch fails to run checkstyle in hadoop-yarn-client
-1 :x:
mvnsite
0m 23s
/patch-mvnsite-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
hadoop-yarn-client in the patch failed.
-1 :x:
javadoc
0m 22s
/patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
-1 :x:
javadoc
0m 22s
/patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkPrivateBuild-1.8.0_352-8u352-ga-1~20.04-b08.txt
hadoop-yarn-client in the patch failed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08.
-1 :x:
spotbugs
0m 22s
/patch-spotbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
hadoop-yarn-client in the patch failed.
+1 :green_heart:
shadedclient
31m 9s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
28m 16s
hadoop-yarn-client in the patch passed.
+1 :green_heart:
asflicense
0m 41s
The patch does not generate ASF License warnings.
145m 33s
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/2/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5309
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux faa74cdeb796 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / e6cf7b764f1feb0dee2a9ddbe5d1de516f65a254
Default Java
Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/2/testReport/
Max. process+thread count
576 (vs. ulimit of 5500)
modules
C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/2/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
@smallzhongfeng Thank you very much for your contribution, We need to fix checkstyle
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 50s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
44m 4s
trunk passed
+1 :green_heart:
compile
0m 38s
trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
compile
0m 34s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
checkstyle
0m 34s
trunk passed
+1 :green_heart:
mvnsite
0m 39s
trunk passed
-1 :x:
javadoc
0m 39s
/branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 27s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
1m 7s
trunk passed
+1 :green_heart:
shadedclient
23m 4s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
0m 33s
the patch passed
+1 :green_heart:
compile
0m 29s
the patch passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
javac
0m 29s
the patch passed
+1 :green_heart:
compile
0m 25s
the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
javac
0m 25s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
0m 17s
the patch passed
+1 :green_heart:
mvnsite
0m 28s
the patch passed
-1 :x:
javadoc
0m 22s
/patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 21s
the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
0m 58s
the patch passed
+1 :green_heart:
shadedclient
23m 10s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
28m 14s
hadoop-yarn-client in the patch passed.
+1 :green_heart:
asflicense
0m 38s
The patch does not generate ASF License warnings.
130m 3s
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/3/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5309
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux 08b67f6886cb 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / f59718b431a01b9652003d4cadf54192e4fedf33
Default Java
Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/3/testReport/
Max. process+thread count
576 (vs. ulimit of 5500)
modules
C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/3/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 52s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 0s
codespell was not available.
+0 :ok:
detsecrets
0m 0s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
43m 20s
trunk passed
+1 :green_heart:
compile
0m 38s
trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
compile
0m 35s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
checkstyle
0m 35s
trunk passed
+1 :green_heart:
mvnsite
0m 39s
trunk passed
-1 :x:
javadoc
0m 41s
/branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 29s
trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
1m 7s
trunk passed
+1 :green_heart:
shadedclient
23m 28s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
0m 34s
the patch passed
+1 :green_heart:
compile
0m 29s
the patch passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04
+1 :green_heart:
javac
0m 29s
the patch passed
+1 :green_heart:
compile
0m 26s
the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
javac
0m 25s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
0m 18s
the patch passed
+1 :green_heart:
mvnsite
0m 28s
the patch passed
-1 :x:
javadoc
0m 22s
/patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt
hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.
+1 :green_heart:
javadoc
0m 21s
the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08
+1 :green_heart:
spotbugs
0m 56s
the patch passed
+1 :green_heart:
shadedclient
23m 6s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
28m 14s
hadoop-yarn-client in the patch passed.
+1 :green_heart:
asflicense
0m 40s
The patch does not generate ASF License warnings.
129m 47s
Subsystem
Report/Notes
Docker
ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/4/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5309
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux 06bd08ac6e2e 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 05b14c1f60b61487ed06fabbba84b6c1ac373f35
Default Java
Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/4/testReport/
Max. process+thread count
576 (vs. ulimit of 5500)
modules
C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/4/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
merged trunk, @smallzhongfeng Thanks for your contribution!
Thanks for your review. @slfan1989 👍
| gharchive/pull-request | 2023-01-18T12:37:33 | 2025-04-01T04:33:28.719020 | {
"authors": [
"hadoop-yetus",
"slfan1989",
"smallzhongfeng",
"szilard-nemeth"
],
"repo": "apache/hadoop",
"url": "https://github.com/apache/hadoop/pull/5309",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1660810905 | HADOOP-18691. Client.Connection#updateAddress needs to ensure that address is resolved before updating
Description of PR
When Client.Connection#setupConnection encounters an IOException, it will try to update the server address. (HADOOP-18365)
When the address is re-parsed, it may be an unresolved address (UnknownHostException), which causes Client.Connection#setupConnection to fail to reconnect.
while (true) {
try {
if (server.isUnresolved()) {
Especially when DN is connected to NN, BPServiceActor#bpNamenode is only initialized once, which causes DN to never connect to NN before restarting.
How was this patch tested?
For code changes:
[ ] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')?
[ ] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation?
[ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0?
[ ] If applicable, have you updated the LICENSE, LICENSE-binary, NOTICE-binary files?
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 52s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 2s
codespell was not available.
+0 :ok:
detsecrets
0m 2s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
58m 43s
trunk passed
-1 :x:
compile
31m 46s
/branch-compile-root-jdkUbuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.txt
root in trunk failed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.
-1 :x:
compile
0m 54s
/branch-compile-root-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt
root in trunk failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.
+1 :green_heart:
checkstyle
2m 8s
trunk passed
-1 :x:
mvnsite
1m 30s
/branch-mvnsite-hadoop-common-project_hadoop-common.txt
hadoop-common in trunk failed.
+1 :green_heart:
javadoc
1m 38s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 57s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
3m 51s
trunk passed
-1 :x:
shadedclient
25m 44s
branch has errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
1m 29s
the patch passed
-1 :x:
compile
10m 14s
/patch-compile-root-jdkUbuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.txt
root in the patch failed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.
-1 :x:
javac
10m 14s
/patch-compile-root-jdkUbuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.txt
root in the patch failed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.
-1 :x:
compile
0m 30s
/patch-compile-root-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt
root in the patch failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.
-1 :x:
javac
0m 30s
/patch-compile-root-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt
root in the patch failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
1m 15s
the patch passed
+1 :green_heart:
mvnsite
1m 55s
the patch passed
+1 :green_heart:
javadoc
1m 4s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
-1 :x:
javadoc
0m 25s
/patch-javadoc-hadoop-common-project_hadoop-common-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt
hadoop-common in the patch failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.
+1 :green_heart:
spotbugs
3m 51s
the patch passed
-1 :x:
shadedclient
8m 39s
patch has errors when building and testing our client artifacts.
_ Other Tests _
-1 :x:
unit
0m 25s
/patch-unit-hadoop-common-project_hadoop-common.txt
hadoop-common in the patch failed.
+0 :ok:
asflicense
0m 25s
ASF License check generated no output?
155m 59s
Subsystem
Report/Notes
Docker
ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5542
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux 7f78a852880b 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 29ac1d27728be1d3e68a4ca158ed21fd673d5e02
Default Java
Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/1/testReport/
Max. process+thread count
186 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/1/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 48s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
43m 50s
trunk passed
+1 :green_heart:
compile
25m 5s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
compile
21m 42s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
checkstyle
1m 5s
trunk passed
+1 :green_heart:
mvnsite
1m 39s
trunk passed
+1 :green_heart:
javadoc
1m 9s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 42s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
2m 43s
trunk passed
+1 :green_heart:
shadedclient
25m 37s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
1m 0s
the patch passed
+1 :green_heart:
compile
24m 37s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javac
24m 37s
the patch passed
+1 :green_heart:
compile
21m 43s
the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
javac
21m 43s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
0m 59s
the patch passed
+1 :green_heart:
mvnsite
1m 35s
the patch passed
+1 :green_heart:
javadoc
1m 1s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 41s
the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
2m 40s
the patch passed
+1 :green_heart:
shadedclient
25m 14s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
18m 9s
hadoop-common in the patch passed.
+1 :green_heart:
asflicense
0m 53s
The patch does not generate ASF License warnings.
223m 6s
Subsystem
Report/Notes
Docker
ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/2/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5542
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux b1511504b561 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 596bb839cdbdf52a34f324c8313e5777aa604894
Default Java
Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/2/testReport/
Max. process+thread count
1375 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/2/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
17m 31s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 0s
codespell was not available.
+0 :ok:
detsecrets
0m 0s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 1 new or modified test files.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
42m 46s
trunk passed
+1 :green_heart:
compile
25m 14s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
compile
21m 43s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
checkstyle
1m 6s
trunk passed
+1 :green_heart:
mvnsite
1m 38s
trunk passed
+1 :green_heart:
javadoc
1m 9s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 43s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
2m 47s
trunk passed
+1 :green_heart:
shadedclient
25m 24s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
1m 1s
the patch passed
+1 :green_heart:
compile
27m 2s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javac
27m 2s
the patch passed
+1 :green_heart:
compile
22m 55s
the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
javac
22m 55s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
1m 0s
the patch passed
+1 :green_heart:
mvnsite
1m 42s
the patch passed
+1 :green_heart:
javadoc
1m 0s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 40s
the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
2m 46s
the patch passed
+1 :green_heart:
shadedclient
25m 12s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
18m 25s
hadoop-common in the patch passed.
+1 :green_heart:
asflicense
0m 54s
The patch does not generate ASF License warnings.
242m 38s
Subsystem
Report/Notes
Docker
ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/3/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5542
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux c172ac28a844 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 9620ed8db83b5f4c5bc73ecf86b6efc3b5f05275
Default Java
Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/3/testReport/
Max. process+thread count
1375 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/3/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
0m 47s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 1s
codespell was not available.
+0 :ok:
detsecrets
0m 1s
detect-secrets was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 1 new or modified test files.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
42m 41s
trunk passed
+1 :green_heart:
compile
25m 20s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
compile
21m 48s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
checkstyle
1m 7s
trunk passed
+1 :green_heart:
mvnsite
1m 37s
trunk passed
+1 :green_heart:
javadoc
1m 11s
trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 41s
trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
2m 43s
trunk passed
+1 :green_heart:
shadedclient
25m 34s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
1m 0s
the patch passed
+1 :green_heart:
compile
24m 33s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javac
24m 33s
the patch passed
+1 :green_heart:
compile
21m 36s
the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
javac
21m 36s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
checkstyle
1m 0s
the patch passed
+1 :green_heart:
mvnsite
1m 36s
the patch passed
+1 :green_heart:
javadoc
0m 59s
the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1
+1 :green_heart:
javadoc
0m 42s
the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
+1 :green_heart:
spotbugs
2m 43s
the patch passed
+1 :green_heart:
shadedclient
25m 11s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
+1 :green_heart:
unit
18m 31s
hadoop-common in the patch passed.
+1 :green_heart:
asflicense
0m 52s
The patch does not generate ASF License warnings.
222m 12s
Subsystem
Report/Notes
Docker
ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/4/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/5542
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets
uname
Linux 8d8fd1b002d9 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 4c1c871292b40c5d38ce0e232bc19ef8949d81ef
Default Java
Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/4/testReport/
Max. process+thread count
2253 (vs. ulimit of 5500)
modules
C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/4/console
versions
git=2.25.1 maven=3.6.3 spotbugs=4.2.2
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
| gharchive/pull-request | 2023-04-10T13:50:01 | 2025-04-01T04:33:28.852365 | {
"authors": [
"cxzl25",
"hadoop-yetus"
],
"repo": "apache/hadoop",
"url": "https://github.com/apache/hadoop/pull/5542",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2205076172 | HDFS-17441. Fix junit dependency by adding missing library in hadoop-hdfs-rbf.
Description of PR
We need to add some missing junit libraries in hadoop-hdfs-rbf.
Based on Ayush's advice, we need to add junit-platform-launcher. We might also need junit-jupiter-api, similar to other modules.
How was this patch tested?
For code changes:
[x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')?
@ayushtkn Could you please review it if you have bandwidth?
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Logfile
Comment
+0 :ok:
reexec
6m 41s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
codespell
0m 0s
codespell was not available.
+0 :ok:
detsecrets
0m 0s
detect-secrets was not available.
+0 :ok:
xmllint
0m 0s
xmllint was not available.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
-1 :x:
test4tests
0m 0s
The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch.
_ trunk Compile Tests _
+1 :green_heart:
mvninstall
31m 42s
trunk passed
+1 :green_heart:
compile
0m 25s
trunk passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1
+1 :green_heart:
compile
0m 23s
trunk passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06
+1 :green_heart:
mvnsite
0m 26s
trunk passed
+1 :green_heart:
javadoc
0m 30s
trunk passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1
+1 :green_heart:
javadoc
0m 22s
trunk passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06
+1 :green_heart:
shadedclient
53m 29s
branch has no errors when building and testing our client artifacts.
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
0m 20s
the patch passed
+1 :green_heart:
compile
0m 19s
the patch passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1
+1 :green_heart:
javac
0m 19s
the patch passed
+1 :green_heart:
compile
0m 18s
the patch passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06
+1 :green_heart:
javac
0m 18s
the patch passed
+1 :green_heart:
blanks
0m 0s
The patch has no blanks issues.
+1 :green_heart:
mvnsite
0m 22s
the patch passed
+1 :green_heart:
javadoc
0m 19s
the patch passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1
+1 :green_heart:
javadoc
0m 18s
the patch passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06
+1 :green_heart:
shadedclient
20m 44s
patch has no errors when building and testing our client artifacts.
_ Other Tests _
-1 :x:
unit
26m 7s
/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt
hadoop-hdfs-rbf in the patch passed.
+1 :green_heart:
asflicense
0m 26s
The patch does not generate ASF License warnings.
112m 4s
Reason
Tests
Failed junit tests
hadoop.hdfs.server.federation.router.TestRouterRpc
hadoop.hdfs.server.federation.router.TestRouterRpcMultiDestination
Subsystem
Report/Notes
Docker
ClientAPI=1.45 ServerAPI=1.45 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6669/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hadoop/pull/6669
Optional Tests
dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient codespell detsecrets xmllint
uname
Linux c1ea8b931ca6 5.15.0-94-generic #104-Ubuntu SMP Tue Jan 9 15:25:40 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/bin/hadoop.sh
git revision
trunk / 41759a731f419a773c73fe0f62ab5586be4ffcfa
Default Java
Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06
Multi-JDK versions
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06
Test Results
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6669/1/testReport/
Max. process+thread count
3981 (vs. ulimit of 5500)
modules
C: hadoop-hdfs-project/hadoop-hdfs-rbf U: hadoop-hdfs-project/hadoop-hdfs-rbf
Console output
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6669/1/console
versions
git=2.25.1 maven=3.6.3
Powered by
Apache Yetus 0.14.0 https://yetus.apache.org
This message was automatically generated.
The failed tests are filed by HDFS-17435.
@ayushtkn Thanks for your review!
| gharchive/pull-request | 2024-03-25T07:24:50 | 2025-04-01T04:33:28.886378 | {
"authors": [
"hadoop-yetus",
"tasanuma"
],
"repo": "apache/hadoop",
"url": "https://github.com/apache/hadoop/pull/6669",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1106441239 | Backport HBASE-26579 Set storage policy of recovered edits when wal storage type is configured
https://issues.apache.org/jira/browse/HBASE-26678
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Comment
+0 :ok:
reexec
4m 42s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+1 :green_heart:
hbaseanti
0m 0s
Patch does not have any anti-patterns.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
+1 :green_heart:
test4tests
0m 0s
The patch appears to include 1 new or modified test files.
_ branch-1 Compile Tests _
+1 :green_heart:
mvninstall
10m 36s
branch-1 passed
+1 :green_heart:
compile
0m 44s
branch-1 passed with JDK Azul Systems, Inc.-1.8.0_262-b19
+1 :green_heart:
compile
0m 49s
branch-1 passed with JDK Azul Systems, Inc.-1.7.0_272-b10
+1 :green_heart:
checkstyle
1m 43s
branch-1 passed
+1 :green_heart:
shadedjars
3m 15s
branch has no errors when building our shaded downstream artifacts.
+1 :green_heart:
javadoc
0m 48s
branch-1 passed with JDK Azul Systems, Inc.-1.8.0_262-b19
+1 :green_heart:
javadoc
0m 42s
branch-1 passed with JDK Azul Systems, Inc.-1.7.0_272-b10
+0 :ok:
spotbugs
3m 9s
Used deprecated FindBugs config; considering switching to SpotBugs.
+1 :green_heart:
findbugs
3m 7s
branch-1 passed
_ Patch Compile Tests _
+1 :green_heart:
mvninstall
2m 7s
the patch passed
+1 :green_heart:
compile
0m 45s
the patch passed with JDK Azul Systems, Inc.-1.8.0_262-b19
+1 :green_heart:
javac
0m 45s
the patch passed
+1 :green_heart:
compile
0m 48s
the patch passed with JDK Azul Systems, Inc.-1.7.0_272-b10
+1 :green_heart:
javac
0m 48s
the patch passed
+1 :green_heart:
checkstyle
1m 31s
the patch passed
+1 :green_heart:
whitespace
0m 0s
The patch has no whitespace issues.
+1 :green_heart:
shadedjars
3m 4s
patch has no errors when building our shaded downstream artifacts.
+1 :green_heart:
hadoopcheck
5m 10s
Patch does not cause any errors with Hadoop 2.8.5 2.9.2.
+1 :green_heart:
javadoc
0m 32s
the patch passed with JDK Azul Systems, Inc.-1.8.0_262-b19
+1 :green_heart:
javadoc
0m 41s
the patch passed with JDK Azul Systems, Inc.-1.7.0_272-b10
+1 :green_heart:
findbugs
2m 59s
the patch passed
_ Other Tests _
+1 :green_heart:
unit
132m 14s
hbase-server in the patch passed.
+1 :green_heart:
asflicense
0m 41s
The patch does not generate ASF License warnings.
180m 55s
Subsystem
Report/Notes
Docker
ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/HBase/job/HBase-PreCommit-GitHub-PR/job/PR-4037/1/artifact/out/Dockerfile
GITHUB PR
https://github.com/apache/hbase/pull/4037
Optional Tests
dupname asflicense javac javadoc unit spotbugs findbugs shadedjars hadoopcheck hbaseanti checkstyle compile
uname
Linux ef48a0add413 4.15.0-112-generic #113-Ubuntu SMP Thu Jul 9 23:41:39 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
/home/jenkins/jenkins-home/workspace/Base-PreCommit-GitHub-PR_PR-4037/out/precommit/personality/provided.sh
git revision
branch-1 / 598b453a41
Default Java
Azul Systems, Inc.-1.7.0_272-b10
Multi-JDK versions
/usr/lib/jvm/zulu-8-amd64:Azul Systems, Inc.-1.8.0_262-b19 /usr/lib/jvm/zulu-7-amd64:Azul Systems, Inc.-1.7.0_272-b10
Test Results
https://ci-hadoop.apache.org/job/HBase/job/HBase-PreCommit-GitHub-PR/job/PR-4037/1/testReport/
Max. process+thread count
4746 (vs. ulimit of 10000)
modules
C: hbase-server U: hbase-server
Console output
https://ci-hadoop.apache.org/job/HBase/job/HBase-PreCommit-GitHub-PR/job/PR-4037/1/console
versions
git=2.17.1 maven=3.6.0 findbugs=3.0.1
Powered by
Apache Yetus 0.12.0 https://yetus.apache.org
This message was automatically generated.
| gharchive/pull-request | 2022-01-18T03:36:40 | 2025-04-01T04:33:28.915661 | {
"authors": [
"Apache-HBase",
"YutSean"
],
"repo": "apache/hbase",
"url": "https://github.com/apache/hbase/pull/4037",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2015726686 | HBASE-28212 Do not need to maintain rollback step when root procedure…
… does not support rollback (#5538)
Signed-off-by: GeorryHuang huangzhuoyue@apache.org
(cherry picked from commit 4b015e6a5486394d70bbf5fc0197e469c0987913) (cherry picked from commit 7d6de8b948a088d316124d6587e6a540bc5b38f1)
:broken_heart: -1 overall
Vote
Subsystem
Runtime
Comment
+0 :ok:
reexec
0m 44s
Docker mode activated.
_ Prechecks _
+1 :green_heart:
dupname
0m 0s
No case conflicting files found.
+0 :ok:
prototool
0m 0s
prototool was not available.
+1 :green_heart:
hbaseanti
0m 0s
Patch does not have any anti-patterns.
+1 :green_heart:
@author
0m 0s
The patch does not contain any @author tags.
_ branch-2 Compile Tests _
+0 :ok:
mvndep
0m 13s
Maven dependency ordering for branch
+1 :green_heart:
mvninstall
3m 14s
branch-2 passed
+1 :green_heart:
compile
3m 24s
branch-2 passed
+1 :green_heart:
checkstyle
0m 50s
branch-2 passed
-1 :x:
spotless
0m 44s
branch has 1 errors when running spotless:check, run spotless:apply to fix.
+1 :green_heart:
spotbugs
3m 50s
branch-2 passed
_ Patch Compile Tests _
+0 :ok:
mvndep
0m 15s
Maven dependency ordering for patch
+1 :green_heart:
mvninstall
2m 45s
the patch passed
+1 :green_heart:
compile
3m 23s
the patch passed
+1 :green_heart:
cc
3m 23s
the patch passed
+1 :green_heart:
javac
3m 23s
the patch passed
+1 :green_heart:
checkstyle
0m 48s
the patch passed
+1 :green_heart:
whitespace
0m 0s
The patch has no whitespace issues.
+1 :green_heart:
hadoopcheck
15m 10s
Patch does not cause any errors with Hadoop 2.10.2 or 3.2.4 3.3.6.
+1 :green_heart:
hbaseprotoc
1m 13s
the patch passed
+1 :green_heart:
spotless
0m 46s
patch has no errors when running spotless:check.
+1 :green_heart:
spotbugs
4m 14s
the patch passed
_ Other Tests _
+1 :green_heart:
asflicense
0m 23s
The patch does not generate ASF License warnings.
43m 51s
Subsystem
Report/Notes
Docker
ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-general-check/output/Dockerfile
GITHUB PR
https://github.com/apache/hbase/pull/5547
Optional Tests
dupname asflicense javac spotbugs hadoopcheck hbaseanti spotless checkstyle compile cc hbaseprotoc prototool
uname
Linux 7a6175b64081 5.4.0-1103-aws #111~18.04.1-Ubuntu SMP Tue May 23 20:04:10 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/hbase-personality.sh
git revision
branch-2 / ea11923239
Default Java
Eclipse Adoptium-11.0.17+8
spotless
https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-general-check/output/branch-spotless.txt
Max. process+thread count
81 (vs. ulimit of 30000)
modules
C: hbase-protocol-shaded hbase-procedure hbase-server U: .
Console output
https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/console
versions
git=2.34.1 maven=3.8.6 spotbugs=4.7.3
Powered by
Apache Yetus 0.12.0 https://yetus.apache.org
This message was automatically generated.
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Comment
+0 :ok:
reexec
0m 44s
Docker mode activated.
-0 :warning:
yetus
0m 4s
Unprocessed flag(s): --brief-report-file --spotbugs-strict-precheck --whitespace-eol-ignore-list --whitespace-tabs-ignore-list --quick-hadoopcheck
_ Prechecks _
_ branch-2 Compile Tests _
+0 :ok:
mvndep
0m 12s
Maven dependency ordering for branch
+1 :green_heart:
mvninstall
3m 14s
branch-2 passed
+1 :green_heart:
compile
1m 26s
branch-2 passed
+1 :green_heart:
shadedjars
5m 33s
branch has no errors when building our shaded downstream artifacts.
+1 :green_heart:
javadoc
0m 43s
branch-2 passed
_ Patch Compile Tests _
+0 :ok:
mvndep
0m 18s
Maven dependency ordering for patch
+1 :green_heart:
mvninstall
2m 45s
the patch passed
+1 :green_heart:
compile
1m 28s
the patch passed
+1 :green_heart:
javac
1m 28s
the patch passed
+1 :green_heart:
shadedjars
5m 32s
patch has no errors when building our shaded downstream artifacts.
+1 :green_heart:
javadoc
0m 41s
the patch passed
_ Other Tests _
+1 :green_heart:
unit
0m 30s
hbase-protocol-shaded in the patch passed.
+1 :green_heart:
unit
1m 27s
hbase-procedure in the patch passed.
+1 :green_heart:
unit
207m 47s
hbase-server in the patch passed.
236m 34s
Subsystem
Report/Notes
Docker
ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-jdk11-hadoop3-check/output/Dockerfile
GITHUB PR
https://github.com/apache/hbase/pull/5547
Optional Tests
javac javadoc unit shadedjars compile
uname
Linux 12789fc35368 5.4.0-1103-aws #111~18.04.1-Ubuntu SMP Tue May 23 20:04:10 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/hbase-personality.sh
git revision
branch-2 / ea11923239
Default Java
Eclipse Adoptium-11.0.17+8
Test Results
https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/testReport/
Max. process+thread count
4759 (vs. ulimit of 30000)
modules
C: hbase-protocol-shaded hbase-procedure hbase-server U: .
Console output
https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/console
versions
git=2.34.1 maven=3.8.6
Powered by
Apache Yetus 0.12.0 https://yetus.apache.org
This message was automatically generated.
:confetti_ball: +1 overall
Vote
Subsystem
Runtime
Comment
+0 :ok:
reexec
0m 52s
Docker mode activated.
-0 :warning:
yetus
0m 7s
Unprocessed flag(s): --brief-report-file --spotbugs-strict-precheck --whitespace-eol-ignore-list --whitespace-tabs-ignore-list --quick-hadoopcheck
_ Prechecks _
_ branch-2 Compile Tests _
+0 :ok:
mvndep
0m 15s
Maven dependency ordering for branch
+1 :green_heart:
mvninstall
3m 49s
branch-2 passed
+1 :green_heart:
compile
1m 46s
branch-2 passed
+1 :green_heart:
shadedjars
6m 34s
branch has no errors when building our shaded downstream artifacts.
+1 :green_heart:
javadoc
0m 53s
branch-2 passed
_ Patch Compile Tests _
+0 :ok:
mvndep
0m 10s
Maven dependency ordering for patch
+1 :green_heart:
mvninstall
3m 22s
the patch passed
+1 :green_heart:
compile
1m 48s
the patch passed
+1 :green_heart:
javac
1m 48s
the patch passed
+1 :green_heart:
shadedjars
6m 27s
patch has no errors when building our shaded downstream artifacts.
+1 :green_heart:
javadoc
0m 53s
the patch passed
_ Other Tests _
+1 :green_heart:
unit
0m 38s
hbase-protocol-shaded in the patch passed.
+1 :green_heart:
unit
1m 57s
hbase-procedure in the patch passed.
+1 :green_heart:
unit
264m 25s
hbase-server in the patch passed.
299m 1s
Subsystem
Report/Notes
Docker
ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-jdk8-hadoop2-check/output/Dockerfile
GITHUB PR
https://github.com/apache/hbase/pull/5547
Optional Tests
javac javadoc unit shadedjars compile
uname
Linux f8d79484a930 5.4.0-163-generic #180-Ubuntu SMP Tue Sep 5 13:21:23 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Build tool
maven
Personality
dev-support/hbase-personality.sh
git revision
branch-2 / ea11923239
Default Java
Temurin-1.8.0_352-b08
Test Results
https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/testReport/
Max. process+thread count
4169 (vs. ulimit of 30000)
modules
C: hbase-protocol-shaded hbase-procedure hbase-server U: .
Console output
https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/console
versions
git=2.34.1 maven=3.8.6
Powered by
Apache Yetus 0.12.0 https://yetus.apache.org
This message was automatically generated.
The spotless error is not a real style problem...
[ERROR] An internal error occurred during: "Periodic workspace save.".
java.lang.IllegalStateException: Job manager has been shut down.
at org.eclipse.core.internal.jobs.JobManager.schedule (JobManager.java:1295)
at org.eclipse.core.internal.jobs.InternalJob.schedule (InternalJob.java:385)
at org.eclipse.core.runtime.jobs.Job.schedule (Job.java:684)
at org.eclipse.core.internal.events.AutoBuildJob.build (AutoBuildJob.java:110)
at org.eclipse.core.internal.events.BuildManager.endTopLevel (BuildManager.java:604)
at org.eclipse.core.internal.resources.Workspace.endOperation (Workspace.java:1518)
at org.eclipse.core.internal.resources.SaveManager.save (SaveManager.java:1246)
at org.eclipse.core.internal.resources.SaveManager.save (SaveManager.java:1143)
at org.eclipse.core.internal.resources.DelayedSnapshotJob.run (DelayedSnapshotJob.java:55)
at org.eclipse.core.internal.jobs.Worker.run (Worker.java:63)
Let me merge.
| gharchive/pull-request | 2023-11-29T03:26:18 | 2025-04-01T04:33:28.978706 | {
"authors": [
"Apache-HBase",
"Apache9"
],
"repo": "apache/hbase",
"url": "https://github.com/apache/hbase/pull/5547",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2502892236 | [BUG] [customize dashboard]Multiple rows of data are not displayed correctly
Is there an existing issue for this?
[X] I have searched the existing issues
Current Behavior
Multiple rows of data are not displayed correctly
Expected Behavior
No response
Steps To Reproduce
No response
Environment
HertzBeat version(s):
Debug logs
No response
Anything else?
No response
How to reproduce this bug
I have repeatedly tested the code from the latest branch several times but failed to reproduce the issue.
| gharchive/issue | 2024-09-03T13:29:40 | 2025-04-01T04:33:28.990631 | {
"authors": [
"kerwin612",
"zhangshenghang",
"zqr10159"
],
"repo": "apache/hertzbeat",
"url": "https://github.com/apache/hertzbeat/issues/2663",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1458755560 | HIVE-23559: Optimise Hive::moveAcidFiles for cloud storage.
What changes were proposed in this pull request?
Transforming Hive::moveAcidFiles for cloud storage to work in parallel instead of sequentially.
Why are the changes needed?
Currently this function ends up transferring DELTA, DELETE_DELTA, BASE prefixes sequentially from staging to final location. This causes delays even with simple updates statements, which updates smaller number of records in cloud storage. The proposed change fixes that.
Does this PR introduce any user-facing change?
No.
How was this patch tested?
Hive pre-commit testing.
Hi @ramesh0201, can you please review?
LGTM. +1
| gharchive/pull-request | 2022-11-21T23:38:46 | 2025-04-01T04:33:28.994000 | {
"authors": [
"difin",
"ramesh0201"
],
"repo": "apache/hive",
"url": "https://github.com/apache/hive/pull/3795",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2526485386 | feat: support storage_options param when reading from table
To integrate hudi-rs with AWS SDK for Pandas (aws wrangler), we must be able to pass boto_session related aws authentication params (mostly AWS_* params) directly and not only rely on env variable inference.
I want to propose adding an option to handle this:
storage_options = {"AWS_ACCESS_KEY_ID": "xxxx", "AWS_SECRET_ACCESS_KEY":"xxxx", "AWS_SECRET_ACCESS_TOKEN":"xxxx"}
hudi_table = HudiTable("/tmp/trips_table", storage_options=storage_options)
records = hudi_table.read_snapshot()
Although I want to add this for S3, it should work for other storage backends.
I'm happy to contribute and add this.
@kazdy sounds good. feel free to take this up and send a pr
I'll wait until #72 gets merged.
I did the first strawman impl and it requires some refactoring in the Table itself.
@xushiyan I also have some questions about this, maybe you can give me your opinion on these:
Should we rename Table to HudiTable?
I don't know why Timeline and FileSystemView both use separate storage instances, can't they share it, maybe there's a reason why it's done this way I can't see atm?
Does it make sense to introduce something that will hold both Timeline and FileSystemView (basically table state) and expose coherent API?
thanks
hey @kazdy
we keep name Table within hudi-core to avoid redundant prefix; everything in hudi-core is about Hudi. When import to other crates, we can give it an alias like HudiTable. We can also add an alias in hudi crate for external facing API when needed. As of now, no strong need for this.
Timeline is responsible for data stored in timeline files under .hoodie/, and FileSystemView is responsible for the data stored under the table excluding .hoodie/. It's good to keep things less coupled, unless there is a need for sharing - it's a stateless client performing IO anyway. Maybe you can make a case about why sharing it?
Currently Table holds Timeline and FileSystemView. You want to elaborate on what you meant by coherent API?
| gharchive/issue | 2024-09-14T15:50:55 | 2025-04-01T04:33:28.999087 | {
"authors": [
"kazdy",
"xushiyan"
],
"repo": "apache/hudi-rs",
"url": "https://github.com/apache/hudi-rs/issues/139",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1097088676 | [HUDI-3157] Remove aws jars from hudi bundles
Tips
Thank you very much for contributing to Apache Hudi.
Please review https://hudi.apache.org/contribute/how-to-contribute before opening a pull request.
What is the purpose of the pull request
Remove aws jars from hudi bundles to align with flink-bundle
ref: https://github.com/apache/hudi/issues/4474
Brief change log
(for example:)
Modify AnnotationLocation checkstyle rule in checkstyle.xml
Verify this pull request
(Please pick either of the following options)
This pull request is a trivial rework / code cleanup without any test coverage.
(or)
This pull request is already covered by existing tests, such as (please describe tests).
(or)
This change added tests and can be verified as follows:
(example:)
Added integration tests for end-to-end.
Added HoodieClientWriteTest to verify the change.
Manually verified the change by running a job locally.
Committer checklist
[ ] Has a corresponding JIRA in PR title & commit
[ ] Commit message is descriptive of the change
[ ] CI is green
[ ] Necessary doc changes done or have another open PR
[ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA.
@xushiyan @nsivabalan pls take a took.
@boneanxs what about also relocating the aws dependency in hudi-aws ? because this alone don't help for aws sdk version mismatch in case they use the dynamodb-lock . they still need to have there own hudi build with shadow relocation ?
@xushiyan I think a better approach would be to provide a hudi-aws-bundle with all the different dependencies for aws separately? As @parisni points out as well, there are more issues here? Can we summarize how we are on master and if anything needs to be done before the release.
@xushiyan I think a better approach would be to provide a hudi-aws-bundle with all the different dependencies for aws separately? As @parisni points out as well, there are more issues here? Can we summarize how we are on master and if anything needs to be done before the release.
as discussed, we should have the aws bundle with some dep issue resolved https://issues.apache.org/jira/browse/HUDI-3878
| gharchive/pull-request | 2022-01-09T02:40:02 | 2025-04-01T04:33:29.008006 | {
"authors": [
"boneanxs",
"parisni",
"vinothchandar",
"xushiyan"
],
"repo": "apache/hudi",
"url": "https://github.com/apache/hudi/pull/4542",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1587402330 | [HUDI-5812] Optimize the data size check in HoodieBaseParquetWriter
Change Logs
Describe context and summary for this change. Highlight if any code was copied.
Finish TODO, we can actually do evaluation more accurately:
if we cache last data size check, since we account for how many records were written we can accurately project avg record size, and therefore estimate how many more records we can write before cut off
Impact
Describe any public API or user-facing feature change or any performance impact.
None
Risk level (write none, low medium or high below)
If medium or high, explain what verification was done to mitigate the risks.
none
Documentation Update
Describe any necessary documentation update if there is any new feature, config, or user-facing change
The config description must be updated if new configs are added or the default value of the configs are changed
Any new feature or user-facing change requires updating the Hudi website. Please create a Jira ticket, attach the
ticket number here and follow the instruction to make
changes to the website.
Contributor's checklist
[ ] Read through contributor's guide
[ ] Change Logs and Impact were stated clearly
[ ] Adequate tests were added if applicable
[ ] CI passed
@alexeykudinkin @danny0405 could you please help to review this?
Gentle ping @danny0405, all comments are addressed, is it good to merge now?
| gharchive/pull-request | 2023-02-16T10:13:31 | 2025-04-01T04:33:29.013566 | {
"authors": [
"boneanxs"
],
"repo": "apache/hudi",
"url": "https://github.com/apache/hudi/pull/7978",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
791908386 | IGNITE-14035: Table access API. Draft.
Simple examples added.
Tip: use draft PRs
https://github.blog/2019-02-14-introducing-draft-pull-requests/
| gharchive/pull-request | 2021-01-22T10:57:39 | 2025-04-01T04:33:29.040934 | {
"authors": [
"AMashenkov",
"ptupitsyn"
],
"repo": "apache/ignite-3",
"url": "https://github.com/apache/ignite-3/pull/33",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
278589215 | IGNITE-6565: Use long type for size and keySize in cache metrics
https://issues.apache.org/jira/browse/IGNITE-6565
@akuznetsov-gridgain Sorry, but I don't understand how it connected with the task?
@akuznetsov-gridgain done. I also fixed some conflicts which appeared since PR was created.
Hi!
Thanks for changes, they needed for compatibility with previous versions.
Code now almost looks good for me, except minor note.
Could you please change VisorCacheMetrics#writeExternalData in this way:
@Override protected void writeExternalData(ObjectOutput out) throws IOException { ... write. ALL old metrics... out.writeLong(sizeLong); out.writeLong(keySizeLong); }
and
`
@Override protected void readExternalData(byte protoVer, ObjectInput in) throws IOException, ClassNotFoundException {
.... read ALL old metrics ....
if (protoVer > V1) {
sizeLong = in.readLong();
keySizeLong = in.readLong();
}
`
@akuznetsov-gridgain done.
merged
| gharchive/pull-request | 2017-12-01T20:37:43 | 2025-04-01T04:33:29.044586 | {
"authors": [
"SharplEr",
"akuznetsov-gridgain"
],
"repo": "apache/ignite",
"url": "https://github.com/apache/ignite/pull/3130",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
317306409 | IGNITE-7896 FilePageStore truncate now actually remove redundant
partition page file.
The ticket is resolved. Closing PR
| gharchive/pull-request | 2018-04-24T16:37:37 | 2025-04-01T04:33:29.045453 | {
"authors": [
"dspavlov",
"ivandasch"
],
"repo": "apache/ignite",
"url": "https://github.com/apache/ignite/pull/3910",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
387942049 | [AIRFLOW-3450] Remove unnecessary sigint handler
This obfuscate the actual exit signal.
Make sure you have checked all steps below.
Jira
[x] My PR addresses the following Airflow Jira issues and references them in the PR title. For example, "[AIRFLOW-3450] My Airflow PR"
https://issues.apache.org/jira/browse/AIRFLOW-3450
In case you are fixing a typo in the documentation you can prepend your commit with [AIRFLOW-3450], code changes always need a Jira issue.
Description
[x] Here are some details about my PR, including screenshots of any UI changes:
The signal handler obfuscates the actual exit signal.
Tests
[x] My PR adds the following unit tests OR does not need testing for this extremely good reason:
Commits
[x] My commits all reference Jira issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message":
Subject is separated from body by a blank line
Subject is limited to 50 characters (not including Jira issue reference)
Subject does not end with a period
Subject uses the imperative mood ("add", not "adding")
Body wraps at 72 characters
Body explains "what" and "why", not "how"
Documentation
[x] In case of new functionality, my PR adds documentation that describes how to use it.
When adding new operators/hooks/sensors, the autoclass documentation generation needs to be added.
Code Quality
[x] Passes flake8
@bolkedebruin Any idea why this empty sighandle is there?
Codecov Report
Merging #4283 into master will increase coverage by 0.02%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #4283 +/- ##
=========================================
+ Coverage 78.08% 78.1% +0.02%
=========================================
Files 201 201
Lines 16458 16450 -8
=========================================
- Hits 12851 12849 -2
+ Misses 3607 3601 -6
Impacted Files
Coverage Δ
airflow/bin/cli.py
65.09% <ø> (+0.5%)
:arrow_up:
airflow/models.py
92.29% <0%> (-0.05%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 9dce1f0...66f2e46. Read the comment docs.
I'm running Docker (Cloud composer), but I don't get any explicit reason:
[2018-12-05 20:20:57,138] {jobs.py:401} ERROR - Got an exception! Propagating...
Traceback (most recent call last):
File "/usr/local/lib/airflow/airflow/jobs.py", line 376, in helper
settings.configure_orm()
File "/usr/local/lib/airflow/airflow/settings.py", line 180, in configure_orm
engine = create_engine(SQL_ALCHEMY_CONN, **engine_args)
File "/usr/local/lib/python3.6/site-packages/sqlalchemy/engine/__init__.py", line 391, in create_engine
return strategy.create(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/sqlalchemy/engine/strategies.py", line 162, in create
engine = engineclass(pool, dialect, u, **engine_args)
File "/usr/local/lib/python3.6/site-packages/sqlalchemy/engine/base.py", line 1772, in __init__
self.echo = echo
File "/usr/local/lib/python3.6/site-packages/sqlalchemy/log.py", line 217, in __set__
instance_logger(instance, echoflag=value)
File "/usr/local/lib/python3.6/site-packages/sqlalchemy/log.py", line 173, in instance_logger
def instance_logger(instance, echoflag=None):
File "/usr/local/lib/airflow/airflow/bin/cli.py", line 85, in sigint_handler
sys.exit(0)
So my initial thought was to just let the exception propagate.
I think he behaviour without those signal handlers is that the “graceful” kill docker does won’t work and it will wait the 10s before sending a hard kill. Assuming airflow is pid 1 in the container.
Might a different fix be to not catch SystemExit exception?
@fokko in the case of Cloud Composer it is indeed the health check failing and hence a docker sends a sigterm. The actual exception which is in this case being propagated is the SystemExit exception being thrown by the sys.exit(0).
It is a bit confusing, as the sigint_handler will be called on the main_thread and intermixed with other code. So the stacktrace isn't really a logical sequence of events anymore.
Catching the sigterm and attempting a nice shutdown seems like a good idea, but I don't know if the current sys.exit(0) is the best implementation. Maybe setting a flag somewhere is a better option.
Catching the sigint (which could be a ctrl+c) is also good idea, as otherwise the scheduler will probably not respond to ctrl+c anymore.
Very good point @NielsZeilemaker Thanks for the elaboration. Setting a flag which will cause the scheduler to shutdown sounds like a better plan. In this case we should catch the SIGTERM, set a flag which will stop the scheduler loop, and shutdown the application in 30 seconds. Ref: https://pracucci.com/graceful-shutdown-of-kubernetes-pods.html
If you're going to fix/mess with it, also have a look at this line:
https://github.com/apache/incubator-airflow/blob/1d53f939669102cd0c8461ad9d756b3e0cf74dbe/airflow/jobs.py#L206
Doesn't seem like a good idea to mark a job as success upon a systemexit….
@NielsZeilemaker My suggestion would be to keep the PR's nice and small, so that might be something for you to pick up :-)
| gharchive/pull-request | 2018-12-05T20:56:23 | 2025-04-01T04:33:29.063968 | {
"authors": [
"Fokko",
"NielsZeilemaker",
"ashb",
"codecov-io"
],
"repo": "apache/incubator-airflow",
"url": "https://github.com/apache/incubator-airflow/pull/4283",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
125722608 | APEXCORE-303 #resolve evaluate the launch properties at application l…
…aunch
@davidyan74 : Can you please add test case as well so that users know how to use this enhancement?
@gauravgopi123 added unit test
| gharchive/pull-request | 2016-01-09T00:24:35 | 2025-04-01T04:33:29.065633 | {
"authors": [
"davidyan74",
"gauravgopi123"
],
"repo": "apache/incubator-apex-core",
"url": "https://github.com/apache/incubator-apex-core/pull/201",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
606663621 | fixed some doc styles by markdownlint
fixed some doc styles by markdownlint.
related issue: #1273
Scan detected 1 error(s) in 1 file(s):
[./doc/plugins/proxy-rewrite.md]:
95: line has trailing whitespace.
@dayongi please take a look.
too long...
fixed
Wen Ming notifications@github.com 于2020年4月25日周六 下午3:12写道:
Scan detected 1 error(s) in 1 file(s):
[./doc/plugins/proxy-rewrite.md]:
95: line has trailing whitespace.
@dayongi https://github.com/dayongi please take a look.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/apache/incubator-apisix/pull/1501#issuecomment-619333940,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AJ62C5DCKRZG6PLMLPX4733ROKENTANCNFSM4MQRCCMQ
.
--
tel: 18336032131
email:tianxiaoyongcs@gmail.com
address: 上海
@dayongi
This PR is huge, and I recommend splitting it into smaller PRs. Each PR can contain only a few documents.
This will make it easier to review and approve.
@dayongi How about we close this PR first? You can then split it into multiple small PRs.
ok i will do this
YuanSheng Wang notifications@github.com 于2020年5月8日周五 下午1:57写道:
@dayongi https://github.com/dayongi How about we close this PR first?
You can then split it into multiple small PRs.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/apache/incubator-apisix/pull/1501#issuecomment-625645839,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AJ62C5B2WD4RHO63UFZ3I4DRQONLVANCNFSM4MQRCCMQ
.
--
tel: 18336032131
email:tianxiaoyongcs@gmail.com
address: 上海
ok, we can close this PR first. waiting for your new small PRs.
| gharchive/pull-request | 2020-04-25T01:47:59 | 2025-04-01T04:33:29.074173 | {
"authors": [
"dayongi",
"juzhiyuan",
"membphis",
"moonming"
],
"repo": "apache/incubator-apisix",
"url": "https://github.com/apache/incubator-apisix/pull/1501",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1237324696 | Add a plugin for Ones
Description
Collect data from Ones(Cloud).
Pre-requisites
Please make sure you have access to Ones API before applying to this issue; otherwise, it's nearly impossible to finish this task.
Describe the solution you'd like
Add a plugin for ones. Please refer to other data source plugins such as jira and tapd.
This plugin performs 3 major tasks:
Collect data from Ones API. The API response will be stored in the raw data layer.
1st ETL: transform the raw data and load it into the tool layer.
2nd ETL: transform the tool layer data and load it into the domain layer. As Ones is an issue tracking tool, the final data to be loaded into the domain layer are mainly entities from issue tracking domain
How to start
Please follow the doc how to add a plugin.
Additional notes
In DevLake, there're 3 layers of storage:
raw data layer: tables starting with _raw_
tool layer: tables starting with _tool_
domain layer: tables with no prefix
You can set up DevLake and see the tables in MySQL.
I would like to tackle this one. Please approve @hezyin @Startrekzky and assign me to it :
最近几次Devlake会议都有进去参加学习
WeOpen Star
I would like to tackle this one. Please approve @hezyin @Startrekzky and assign me to it : 最近几次Devlake会议都有进去参加学习
Thanks @yujng so much, I'm looking forward to your PRs.
"Ones" only supports on-premise deployment, which means this issue is more suitable for those who already have the access to an "Ones" instance.
We still need this.
als thumb_up for this feature request.
I would like to tackle this one. Please approve @hezyin @Startrekzky and assign me to it. I went to the recent community meeting for a couple of times.
Any progress on this feature ? @hezyin @Startrekzky
| gharchive/issue | 2022-05-16T15:23:21 | 2025-04-01T04:33:29.082141 | {
"authors": [
"Startrekzky",
"morganchenjp",
"yujng",
"yumengwang03"
],
"repo": "apache/incubator-devlake",
"url": "https://github.com/apache/incubator-devlake/issues/1884",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
759951852 | page of [TaskInstance] load data slow
page of [TaskInstance] load data by interface "/dolphinscheduler/projects/stressTes/task-instance/list-paging" slow when the size of result more than 230000
Do you have any good solutions, welcome to submit PR implementation.
| gharchive/issue | 2020-12-09T03:09:35 | 2025-04-01T04:33:29.083364 | {
"authors": [
"wangxj3",
"zhuangchong"
],
"repo": "apache/incubator-dolphinscheduler",
"url": "https://github.com/apache/incubator-dolphinscheduler/issues/4184",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
957670999 | Support bitmap(binary) type in hive to import doris
Proposed changes
Describe the big picture of your changes here to communicate to the maintainers why we should accept this pull request. If it fixes a bug or resolves a feature request, be sure to link to that issue.
Types of changes
What types of changes does your code introduce to Doris?
Put an x in the boxes that apply
[ ] Bugfix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation Update (if none of the other choices apply)
[ ] Code refactor (Modify the code structure, format the code, etc...)
[ ] Optimization. Including functional usability improvements and performance improvements.
[x] Dependency. Such as changes related to third-party components.
[ ] Other.
Checklist
Put an x in the boxes that apply. You can also fill these out after creating the PR. If you're unsure about any of them, don't hesitate to ask. We're here to help! This is simply a reminder of what we are going to look for before merging your code.
[ ] I have created an issue on (Fix #ISSUE) and described the bug/feature there in detail
[ ] Compiling and unit tests pass locally with my changes
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] If these changes need document changes, I have updated the document
[ ] Any dependent changes have been merged
Further comments
If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc...
Please add User Manual for this feature
This pr would be closed temporarily and will be re-pull request after internal test
| gharchive/pull-request | 2021-08-02T02:04:58 | 2025-04-01T04:33:29.089527 | {
"authors": [
"rufeng97",
"wangbo"
],
"repo": "apache/incubator-doris",
"url": "https://github.com/apache/incubator-doris/pull/6356",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1038175849 | [Improve] Spark connector supports multiple spark versions:2.1.x/2.3.x/3.x
Proposed changes
Part of Spark Connector code is implemented using the "Spark DatasourcesV2" API . However , "Spark DatasourcesV2" API only exists in spark 2.3.x/2.4.x , not exists in spark 2.1.x/2.2.x/3.x.
That API seems to have been removed after spark 3.x .
For supporting multiple spark versions, part of the code is re-implemented using the "Spark DatasourcesV1" API.
Types of changes
What types of changes does your code introduce to Doris?
Put an x in the boxes that apply
[ ] Bugfix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation Update (if none of the other choices apply)
[ ] Code refactor (Modify the code structure, format the code, etc...)
[ ] Optimization. Including functional usability improvements and performance improvements.
[ ] Dependency. Such as changes related to third-party components.
[x] Other.
Checklist
Put an x in the boxes that apply. You can also fill these out after creating the PR. If you're unsure about any of them, don't hesitate to ask. We're here to help! This is simply a reminder of what we are going to look for before merging your code.
[ ] I have created an issue on (Fix #ISSUE) and described the bug/feature there in detail
[x] Compiling and unit tests pass locally with my changes
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] If these changes need document changes, I have updated the document
[ ] Any dependent changes have been merged
https://github.com/apache/incubator-doris/issues/7363
| gharchive/pull-request | 2021-10-28T07:29:48 | 2025-04-01T04:33:29.095900 | {
"authors": [
"chovy-3012",
"lordk911"
],
"repo": "apache/incubator-doris",
"url": "https://github.com/apache/incubator-doris/pull/6956",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
399084810 | downtime when upgrading single node overlord+coordinator to 0.13.0
I noticed this behavior while upgrading from 0.12.3 to 0.13.0.
We have a single node merged overlord+coordinator.
The assumption is that since overlord/coordinator are not in the query path, upgrading that node shouldn't result in any downtime.
However, it seems when the node restarted (after upgrade), it terminated all existing KIS tasks and restarted them. This tooks a few minutes to happen, and during these 5 minutes or so, no realtime data was available for querying.
I saw a lot of messages like this during the node startup (after upgrade):
o.a.d.i.c.IndexTaskClient [IndexTaskClient-pctile-hour-0] No TaskLocation available for task [index_kafka_pctile-hour_ec0a9bc8420bc02_ehjdgpfc], this task may not have been assigned to a worker yet or may have already completed
Is this expected behavior? If so, it might be worth documenting.
if KIS is being marked as stable (ref #6970) , this bug might be worth either fixing or atleast documenting.
| gharchive/issue | 2019-01-14T21:39:31 | 2025-04-01T04:33:29.098183 | {
"authors": [
"pdeva"
],
"repo": "apache/incubator-druid",
"url": "https://github.com/apache/incubator-druid/issues/6854",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
452491144 | First Time Launching Druid Tasks fail
Hello,
I set up an Ubuntu 19.04 Virtual Machine using VMWare Workstation.
Java JRE build 11.0.3+7-Ubuntu-1ubuntu219.04.1 and Python 2.7 are installed.
All starting Tasks fail, whenever I try to execute the single machine quicklaunch.
Here's the log:
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by com.google.inject.internal.cglib.core.$ReflectUtils$1 (file:/home/myadmin/apache-druid-0.14.2-incubating/lib/guice-4.1.0.jar) to method java.lang.ClassLoader.defineClass(java.lang.String,byte[],int,int,java.security.ProtectionDomain)
WARNING: Please consider reporting this to the maintainers of com.google.inject.internal.cglib.core.$ReflectUtils$1
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
2019-06-05T11:27:47,578 INFO [main] org.apache.druid.guice.PropertiesModule - Loading properties from common.runtime.properties
2019-06-05T11:27:47,583 INFO [main] org.apache.druid.guice.PropertiesModule - Loading properties from runtime.properties
2019-06-05T11:27:47,768 INFO [main] org.hibernate.validator.internal.util.Version - HV000001: Hibernate Validator 5.1.3.Final
Exception in thread "main" java.lang.NoClassDefFoundError: javax/xml/bind/ValidationException
at java.base/java.lang.ClassLoader.defineClass1(Native Method)
at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:1016)
at java.base/java.security.SecureClassLoader.defineClass(SecureClassLoader.java:174)
at java.base/jdk.internal.loader.BuiltinClassLoader.defineClass(BuiltinClassLoader.java:802)
at java.base/jdk.internal.loader.BuiltinClassLoader.findClassOnClassPathOrNull(BuiltinClassLoader.java:700)
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClassOrNull(BuiltinClassLoader.java:623)
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
at org.hibernate.validator.internal.engine.ConfigurationImpl.<init>(ConfigurationImpl.java:110)
at org.hibernate.validator.internal.engine.ConfigurationImpl.<init>(ConfigurationImpl.java:86)
at org.hibernate.validator.HibernateValidator.createGenericConfiguration(HibernateValidator.java:41)
at javax.validation.Validation$GenericBootstrapImpl.configure(Validation.java:276)
at javax.validation.Validation.buildDefaultValidatorFactory(Validation.java:110)
at org.apache.druid.guice.ConfigModule.configure(ConfigModule.java:39)
at com.google.inject.spi.Elements$RecordingBinder.install(Elements.java:340)
at com.google.inject.spi.Elements.getElements(Elements.java:110)
at com.google.inject.internal.InjectorShell$Builder.build(InjectorShell.java:138)
at com.google.inject.internal.InternalInjectorCreator.build(InternalInjectorCreator.java:104)
at com.google.inject.Guice.createInjector(Guice.java:99)
at com.google.inject.Guice.createInjector(Guice.java:73)
at org.apache.druid.guice.GuiceInjectors.makeStartupInjector(GuiceInjectors.java:56)
at org.apache.druid.cli.Main.main(Main.java:101)
Caused by: java.lang.ClassNotFoundException: javax.xml.bind.ValidationException
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:583)
at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
... 23 more
Is this a Java incompatibility?
I use Druid 0.14.2 incubating and I tried Zookeeper 3.5.5. , 3.4.11 and 3.4.14
Best Regards,
Calvin
Hi Calvin,
Afaik, Druid doesn't support Java 11; you'll have to use Java 8 (as mentioned in the quickstart doc). There's an initiative to support Java 11 compatibility (see #5589) but I don't think it's been completed.
Thank you! Java 8 works.
| gharchive/issue | 2019-06-05T13:07:20 | 2025-04-01T04:33:29.102192 | {
"authors": [
"CalvinSchulze",
"justinborromeo"
],
"repo": "apache/incubator-druid",
"url": "https://github.com/apache/incubator-druid/issues/7835",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
388180919 | Graceful shutdown enhancement in Spring
What is the purpose of the change
Graceful shutdown enhancement in Spring.
AbstractConfig has added a DubboShutdownHook in its constructed function,but when we used spring, SpringExtensionFactory also added a ApplicationListener to invoke DubboShutdownHook when spring fired ContextClosedEvent.
This seems to have a problem。 AbstractConfig‘s DubboShutdownHook will invoke when the application shutdown, at the same time spring boot shutDownhook will invoke to close the context。
when the application closing ,dubbo will make sure all the request has return ,if a request used bean in spring , this bean will close by context。 it will throw error,so in spring ,we should make sure dubbo shutdown before spring destory its beans。
我怕我英文没有表达清楚,用中文简单描述一下。正好也在解决这个问题,当我的应用在使用spring boot 时,spring boot 会注册一个shutdownhook 来关闭 applicationcontext ,同时 Dubbo 也会注册 shutdownhook, 但是我们没办法保证两个shutdownhook的执行顺序。因为dubbo 的优雅停机 会把已经进来的请求处理完,当这些请求依赖spring 的某些bean 时,这些bean 可能已经被spring 容器关闭了,产生了报错。我们应该确保dubbo完全停机了再关闭spring容器。
我目前的做法是取消了spring boot 注册的shutdownhook, 自己实现了一个shutdownhook ,运行时,先sleep(DEFAULT_SERVER_SHUTDOWN_TIMEOUT),再确保dubbo 停机后再关闭容器,如果 DEFAULT_SERVER_SHUTDOWN_TIMEOUT 过大的话,可能会拖延停机时间,这也是为什么我提了一个issue :https://github.com/apache/incubator-dubbo/issues/2877 。
Brief changelog
when add ShutdownHookListener ,remove AbstractConfig‘s DubboShutdownHook
Verifying this change
XXXXX
Follow this checklist to help us incorporate your contribution quickly and easily:
[x] Make sure there is a GITHUB_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GITHUB issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue.
[ ] Format the pull request title like [Dubbo-XXX] Fix UnknownException when host config not exist #XXX. Each commit in the pull request should have a meaningful subject line and body.
[ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why.
[ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add integration-test in test module.
[ ] Run mvn clean install -DskipTests & mvn clean test-compile failsafe:integration-test to make sure unit-test and integration-test pass.
[ ] If this contribution is large, please follow the Software Donation Guide.
Codecov Report
Merging #2901 into master will decrease coverage by 0.02%.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #2901 +/- ##
==========================================
- Coverage 63.74% 63.71% -0.03%
==========================================
Files 577 577
Lines 25930 25931 +1
Branches 4540 4540
==========================================
- Hits 16529 16522 -7
- Misses 7233 7235 +2
- Partials 2168 2174 +6
Impacted Files
Coverage Δ
...ava/org/apache/dubbo/config/DubboShutdownHook.java
84% <100%> (ø)
:arrow_up:
...onfig/spring/extension/SpringExtensionFactory.java
84.61% <100%> (+0.4%)
:arrow_up:
...ache/dubbo/remoting/p2p/support/AbstractGroup.java
45.45% <0%> (-11.37%)
:arrow_down:
...in/java/org/apache/dubbo/common/utils/JVMUtil.java
73.58% <0%> (-7.55%)
:arrow_down:
.../apache/dubbo/remoting/transport/AbstractPeer.java
58.69% <0%> (-4.35%)
:arrow_down:
...dubbo/remoting/exchange/support/DefaultFuture.java
66.44% <0%> (-3.36%)
:arrow_down:
...org/apache/dubbo/rpc/protocol/AbstractInvoker.java
62.9% <0%> (-3.23%)
:arrow_down:
...dubbo/rpc/protocol/dubbo/CallbackServiceCodec.java
77.2% <0%> (-2.21%)
:arrow_down:
...a/org/apache/dubbo/monitor/dubbo/DubboMonitor.java
87.85% <0%> (-1.87%)
:arrow_down:
...apache/dubbo/rpc/protocol/dubbo/DubboProtocol.java
65.83% <0%> (-0.84%)
:arrow_down:
... and 6 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 63bf28d...f6a3213. Read the comment docs.
| gharchive/pull-request | 2018-12-06T11:55:07 | 2025-04-01T04:33:29.122935 | {
"authors": [
"LiZhenNet",
"codecov-io"
],
"repo": "apache/incubator-dubbo",
"url": "https://github.com/apache/incubator-dubbo/pull/2901",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
499535808 | toolTip的formatter属性里,用function自定义时,打印出params.seriesName不能跟配置的系列名对应
Version
4.3.0
Steps to reproduce
鼠标悬浮到地图,观察分组名,全都是第一个系列的名称
What is expected?
跟系列名对应起来
What is actually happening?
全部是第一个系列名称
Could you provide a reproducible demo?
gallery.echartsjs.com 提交不了,我开了个repo作为demo:http://champyin.com/echarts-issues/
你点点图例试试,把「分组1」隐藏了试试:
是不是都是「分组2」了?
你这两个 series 是罗在一起的地图啊~
你点下图例试试:
把「分组1」隐藏,是不是都是「分组2」了?
你这两个 series 是两个罗在一起的中国地图啊= =
你点下图例试试:
把「分组1」隐藏,是不是都是「分组2」了?
你这两个 series 是两个罗在一起的两张中国地图啊= =
嗯,但是不使用formatter自定义的时候,是可以区分「分组1」和 「分组2」的,你可以拉一下我的demo源码,把formatter注释掉看看效果。
所以echarts应该是支持地图罗在一起的时候区分分组的。
我更新了问题复现demo:http://champyin.com/echarts-issues/
加入了效果对比。
额,在原楼层改了好几回……
@zhouxin860114 因为我需要展示图例,所以需要使用多个series。如果像你建议的那样,我的图例没法展示了。
说得有点乱,总结一下,我估计是这么回事:
1、你的两个 map series 是重叠的,最上层的 series 是「分组1」
2、「分组1」显示时,鼠标事件被「分组1」的图形捕获,params 里的 series 是「分组1」,所以自定义的 formatter 只能显示「分组1」的属性
3、 默认的 formatter 似乎会遍历所有 series,把有数据的 series 都显示出来,并把数值相加,如下图:
formatTooltip: function(dataIndex) {
// FIXME orignalData and data is a bit confusing
var data = this.getData();
var formattedValue = addCommas(this.getRawValue(dataIndex));
var name = data.getName(dataIndex);
var seriesGroup = this.seriesGroup;
var seriesNames = [];
for (var i = 0; i < seriesGroup.length; i++) {
var otherIndex = seriesGroup[i].originalData.indexOfName(name);
var valueDim = data.mapDimension('value');
if (!isNaN(seriesGroup[i].originalData.get(valueDim, otherIndex))) {
seriesNames.push(
encodeHTML(seriesGroup[i].name)
);
}
}
return seriesNames.join(', ') + '<br />' +
encodeHTML(name + ' : ' + formattedValue);
},
——————————————————
图例可以自定义,第一个回答里,我后来贴上的那段代码,就是包含自定义图例的
还有一个思路,就是 formatter 的回调函数里,只通过 params.name 确定省份,然后通过省份去原始数据中查询分组信息、数值等;
这样你就可以用多个 series[i]-map 了
@zhouxin860114 分析的很有道理,按照你的方式可以实现我想要的效果,3Q~
@zhouxin860114 分析的很有道理,按照你的方式可以实现我想要的效果,3Q~
不客气 :D
| gharchive/issue | 2019-09-27T16:23:24 | 2025-04-01T04:33:29.132437 | {
"authors": [
"yc111",
"yufeng04",
"zhouxin860114"
],
"repo": "apache/incubator-echarts",
"url": "https://github.com/apache/incubator-echarts/issues/11347",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
333211508 | Request: Cumulative Line Chart
One-line summary
The cumulative line chart doesn't seem to be included in ECharts.
Request
We're thinking of changing our charting library, but it will hard to do so if the new one does not support cumulative line charts.
Example (NVD3)
http://nvd3.org/examples/cumulativeLine.html
Other comments
If there is no time for this to be done by your team, could we develop it for you and have it added to the source code?
Not sure if it meets your request, cumulative line chart can be created by processing the data and use tooltip.formatter and label.formatter to display the original value. Please tell me if I missed anything.
Hi Pissang,
Thank you for your quick reply and sorry for the delay.
That's not exactly what we're looking for as we need a line (similar to the one in the example I've linked), that rebases the chart as you move it.
Kind Regards,
Meldanen
@pissang can you please explan what do you mean?
| gharchive/issue | 2018-06-18T10:20:58 | 2025-04-01T04:33:29.136266 | {
"authors": [
"Meldanen",
"klukiyan",
"pissang"
],
"repo": "apache/incubator-echarts",
"url": "https://github.com/apache/incubator-echarts/issues/8536",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
383051556 | test
[ ] I am using English in this issue. 在这个 Issue 中我使用了英文(强烈建议)。
General Questions
PLEASE MAKE SURE OF ALL THE FOLLOWING OPTIONS IN REQUIRED FIELDS ARE TICKED (with x)!
Otherwise, the issue will not be answered.
And think before you tick. :)
请确保以下每项都打上勾了!并且确保都做了这些事哦~ 这将大量简化我们的工作流程,使你的问题更快速得到解答。
[ ] Required: I have read the document and examples and tried to solve it by myself. (必填)我读过了文档和教程,并且曾试图自己解决问题。
[ ] Required: I have searched for similar issues and that didn't help. (必填)我搜索过 issue 但是没有帮助。
[ ] Required: I have tried with the latest version of ECharts and still have this problem. (必填)我试过最新版本的 ECharts,还是存在这个问题。
In this issue, I have provided information with: 在这个 issue 中我提供了以下信息:
[ ] Required: issue type;(必填)issue 类型
[ ] Required: one sentence description in issue details;(必填)一句话的问题描述
[ ] Required: demo;(必填)能反映问题的例子(如果你想提问或报 bug)
[ ] Required: ECharts version;ECharts 版本
Issue Type
[ ] I have a question to ask about how to use ECharts to ...;我想提问如何使用 ECharts 实现某功能
[ ] I have a bug to report;我想要报 bug
[ ] I have a feature to request, e.g.: I'd like a new feature that ...;我需要一个新功能
[ ] I have a feature to enhance, e.g.: The current feature should be improved in the way that ...;我希望改进某个功能
[ ] There's something wrong with the documents;文档有些问题
[ ] Others, or I'm not sure which issue types to choose here;其他,或我不知道应该选什么类型
Issue Details
{DESCRIBE_HERE}
Expected Behavior
{DESCRIBE_HERE}
Current Behavior
{DESCRIBE_HERE}
Online Example
THIS IS REQUIRED FOR ALL BUG REPORTS AND QUESTIONS!!! 如果是提问或报 bug 一定要提供这一项!
{ADD_LINK_OF_DEMO_HERE}
{PASTE_RELATED_CODE_HERE}
Topics
[ ] Legend
[ ] Tooltip
[ ] Event
[ ] Performance
[ ] SVG
[ ] Map
[ ] ECharts GL
[ ] Third-party libraries, e.g.: Vue.js, React
Anything Else We Need to Know
{DESCRIBE_HERE}
Environment
ECharts version;ECharts 版本: {DESCRIBE_HERE}
[ ] It happens only on certain browsers or operating systems. 对于特定浏览器或操作系统才会出现的问题,请提供相应环境信息:{BROWSER_VERSION_OR_OS_INFORMATION_HERE}
1
| gharchive/issue | 2018-11-21T10:27:55 | 2025-04-01T04:33:29.146495 | {
"authors": [
"Ovilia"
],
"repo": "apache/incubator-echarts",
"url": "https://github.com/apache/incubator-echarts/issues/9450",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
387240905 | tooltip不会自动消失
[x ] I am using English in this issue. 在这个 Issue 中我使用了英文(强烈建议)。
General Questions
PLEASE MAKE SURE OF ALL THE FOLLOWING OPTIONS IN REQUIRED FIELDS ARE TICKED (with x)!
Otherwise, the issue will not be answered.
And think before you tick. :)
请确保以下每项都打上勾了!并且确保都做了这些事哦~ 这将大量简化我们的工作流程,使你的问题更快速得到解答。
[ x] Required: I have read the document and examples and tried to solve it by myself. (必填)我读过了文档和教程,并且曾试图自己解决问题。
[ x] Required: I have searched for similar issues and that didn't help. (必填)我搜索过 issue 但是没有帮助。
[x ] Required: I have tried with the latest version of ECharts and still have this problem. (必填)我试过最新版本的 ECharts,还是存在这个问题。
In this issue, I have provided information with: 在这个 issue 中我提供了以下信息:
[ x] Required: issue type;(必填)issue 类型
[ x] Required: one sentence description in issue details;(必填)一句话的问题描述
[x ] Required: demo;(必填)能反映问题的例子(如果你想提问或报 bug)
[x ] Required: ECharts version;ECharts 版本
Issue Type
[ ] I have a question to ask about how to use ECharts to ...;我想提问如何使用 ECharts 实现某功能
[ x] I have a bug to report;我想要报 bug
[ ] I have a feature to request, e.g.: I'd like a new feature that ...;我需要一个新功能
[ ] I have a feature to enhance, e.g.: The current feature should be improved in the way that ...;我希望改进某个功能
[ ] There's something wrong with the documents;文档有些问题
[ ] Others, or I'm not sure which issue types to choose here;其他,或我不知道应该选什么类型
Issue Details
tooltip doesn't auto hide after pointer leave
Expected Behavior
tooltip auto hide
Current Behavior
{DESCRIBE_HERE}
Online Example
THIS IS REQUIRED FOR ALL BUG REPORTS AND QUESTIONS!!! 如果是提问或报 bug 一定要提供这一项!
{ADD_LINK_OF_DEMO_HERE}
tooltip: {
trigger: 'axis',
confine: true,
enterable: true,
backgroundColor: 'rgba(0,0,0,.8)'
}
Topics
[ ] Legend
[ x] Tooltip
[ ] Event
[ ] Performance
[ ] SVG
[ ] Map
[ ] ECharts GL
[ ] Third-party libraries, e.g.: Vue.js, React
Anything Else We Need to Know
{DESCRIBE_HERE}
Environment
ECharts version;ECharts 版本: 4.2.0-rc.2
[ ] It happens only on certain browsers or operating systems. 对于特定浏览器或操作系统才会出现的问题,请提供相应环境信息:{BROWSER_VERSION_OR_OS_INFORMATION_HERE}
请提供更完整的信息
| gharchive/issue | 2018-12-04T11:27:30 | 2025-04-01T04:33:29.157100 | {
"authors": [
"node-rookie",
"pissang"
],
"repo": "apache/incubator-echarts",
"url": "https://github.com/apache/incubator-echarts/issues/9522",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1569244913 | [ISSUE #2973]Method passes a concatenated string to SLF4J's or Log4j2's format string[SubStreamHandler]
fix #2973
@zhaoyansheng163 please solve the conflicts and the ci problems.
| gharchive/pull-request | 2023-02-03T05:27:40 | 2025-04-01T04:33:29.158166 | {
"authors": [
"xwm1992",
"zhaoyansheng163"
],
"repo": "apache/incubator-eventmesh",
"url": "https://github.com/apache/incubator-eventmesh/pull/3033",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
764754972 | Allow localhost graceful shutdown
#3581 Add gracefully shutdown
In Apache Aurora, it is necessary to allow a graceful shutdown on localhost.
Nice work.. ^^;
[I 201214 14:47:11 web:1811] 200 GET /health (127.0.0.1) 0.64ms
[I 201214 14:47:31 web:1811] 200 GET /health (127.0.0.1) 0.53ms
[I 201214 14:47:46 killexecutorhandler:51] Received 'Killing process' request
[I 201214 14:47:46 web:1811] 200 POST /quitquitquit (127.0.0.1) 1.08ms
[I 201214 14:47:46 killexecutorhandler:44] Killing parent executor
@thinker0 or @Code0x58 Is this PR ok to be further reviewed and merged?
| gharchive/pull-request | 2020-12-13T02:01:04 | 2025-04-01T04:33:29.159787 | {
"authors": [
"joshfischer1108",
"thinker0"
],
"repo": "apache/incubator-heron",
"url": "https://github.com/apache/incubator-heron/pull/3650",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
942382939 | HOP-3013 Adapt ContextDialog position to monitor
I would test tomorrow with multiple monitors, but if anyone can test on Linux / Max / WebHop
It only works when you are using your primary monitor.
If you work on your secondary monitor it trips (unless the resolutions of those monitors are the same)
This still does not really give the desired result on mac, I'll take a look tomorrow if I can gather a bit more information on what is happening. I have a feeling it is still giving the size of the primary display.
Works in Hop Web too, if you can confirm Windows it's ready for merge
Works fine on Windows, but tested with only one monitor
| gharchive/pull-request | 2021-07-12T19:39:27 | 2025-04-01T04:33:29.161797 | {
"authors": [
"hansva",
"nadment"
],
"repo": "apache/incubator-hop",
"url": "https://github.com/apache/incubator-hop/pull/932",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1824206036 | Setup lts patch
Following update done on 1.40.x-prod https://github.com/kiegroup/kogito-images/pull/1611
Many thanks for submitting your Pull Request :heart:!
Please make sure your PR meets the following requirements:
[ ] You have read the contributors guide
[ ] Pull Request title is properly formatted: [KOGITO|RHPAM-XYZ] Subject
[ ] Pull Request contains link to the JIRA issue
[ ] Pull Request contains description of the issue
[ ] Pull Request does not include fixes for issues other than the main ticket
[ ] Your feature/bug fix has a testcase that verifies it
[ ] You've tested the new feature/bug fix in an actual OpenShift cluster
[ ] You've added a RELEASE_NOTES.md entry regarding this change
How to retest this PR or trigger a specific build:
(Re)run Jenkins tests
Please add comment: Jenkins [test|retest] this
Prod tests
Please add comment: Jenkins (re)run [prod|Prod|PROD]
How to backport a pull request to a different branch?
In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch).
NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another.
Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added.
If something goes wrong, the author will be notified and at this point a manual backporting is needed.
NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created.
PR job #1019 was: FAILURE
Possible explanation: Pipeline failure or project build failure
Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images/1019/display/redirect
See console log:
Console Logs
[Pipeline] }Build KIE » kogito » main » pullrequest » kogito-images.build-image PR #1653 - kogito-management-console: https://github.com/kiegroup/kogito-images/pull/1653 completed: UNSTABLE[Pipeline] unstableWARNING: Tests on kogito-management-console seems to have failed[Pipeline] }[Pipeline] // stage[Pipeline] }Build KIE » kogito » main » pullrequest » kogito-images.build-image PR #1653 - kogito-task-console: https://github.com/kiegroup/kogito-images/pull/1653 completed: SUCCESS[Pipeline] }[Pipeline] // stage[Pipeline] }Build KIE » kogito » main » pullrequest » kogito-images.build-image PR #1653 - kogito-s2i-builder: https://github.com/kiegroup/kogito-images/pull/1653 completed: UNSTABLE[Pipeline] unstableWARNING: Tests on kogito-s2i-builder seems to have failed[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // parallel[Pipeline] }[Pipeline] // script[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] echo[INFO] Clean workspace[Pipeline] cleanWs[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] Deferred wipeout is used...[WS-CLEANUP] done[Pipeline] echo[INFO] Workspace cleaned[Pipeline] echo[INFO] Cleanup Maven artifacts[Pipeline] sh+ rm -rf /home/jenkins/.m2/repository[Pipeline] echo[INFO] .m2/repository cleaned[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images/1019/api/json[Pipeline] readJSON[Pipeline] sh
jenkins retest this
Hey @radtriste the patch itself looks good to me but I fear it could hard to maintain in this way, I mean everytime we change QUARKUS_EXTENSIONS here modules/kogito-swf/devmode/build-config/module.yaml we should always update the patch.
So why don't we fully specify all extensions? i.e., <group-id>:<artifact-id>:${KOGITO_VERSION}, is there any issue using this approach on community? if this works it will completely remove the need of having the patch
Hey @radtriste the patch itself looks good to me but I fear it could be hard to maintain in this way, I mean everytime we change QUARKUS_EXTENSIONS here modules/kogito-swf/devmode/build-config/module.yaml we should always update the patch.
So why don't we fully specify all extensions? i.e., <group-id>:<artifact-id>:${KOGITO_VERSION}, is there any issue using this approach on community? if this works it will completely remove the need of having the patch
This is exactly what I propose to Filippe here: https://github.com/kiegroup/kogito-images/pull/1653#discussion_r1276200853 :)
This is exactly what I propose to Filippe here: #1653 (comment) :)
Ouch sorry, I completely missed you comment!!
Then, I definitely agree on your proposal :)
@lampajr hi, does this PR still valid?
Hey @spolti , I think it depends on what you think about https://github.com/apache/incubator-kie-kogito-images/pull/1653#discussion_r1276200853. Because if you agree to fully specify all extensions in terms of <group-id>:<artifact-id>:${KOGITO_VERSION} I think we do NOT need this patch (therefore the PR can be closed) but we will need to create different PR to fully specify all extensions there.
@lampajr yeah, I agree with that, the full GAV is better than a patch.
feel free to merge :)
Is the CI working for the kogito-images already?
@lampajr yeah, I agree with that, the full GAV is better than a patch. feel free to merge :)
Ok so we should create a different PR specifying the full GAV for all extensions and then we can close this one as no patch will be needed at that point.
Is the CI working for the kogito-images already?
TBH I do not know IIRC the issue was just when using Quarkus LTS
| gharchive/pull-request | 2023-07-27T11:50:06 | 2025-04-01T04:33:29.178902 | {
"authors": [
"kie-ci1",
"lampajr",
"radtriste",
"spolti"
],
"repo": "apache/incubator-kie-kogito-images",
"url": "https://github.com/apache/incubator-kie-kogito-images/pull/1653",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2138409762 | kie-issues#777: Allow restricting jenkins agent labels for pipelines
part of https://github.com/apache/incubator-kie-issues/issues/777
PR job #92 was: FAILURE
Possible explanation: Pipeline failure or project build failure
Please look here: https://ci-builds.apache.org/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images.build-and-test/92/display/redirect
See console log:
Console Logs
Failed in branch kogito-base-builderBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1282 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stageBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1284 completed: FAILURE[Pipeline] }Failed in branch kogito-jobs-service-postgresql[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-jobs-service-ephemeralBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1285 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-swf-builderBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1273 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-jobs-service-allinoneBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1281 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-swf-devmodeBuild KIE » kogito » main » pullrequest » kogito-images.build-image PR #1742 - kogito-data-index-ephemeral: https://github.com/apache/incubator-kie-kogito-images/pull/1742 completed: SUCCESS[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // parallel[Pipeline] }[Pipeline] // dir[Pipeline] }[Pipeline] // script[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - 'https://ci-builds.apache.org/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images.build-and-test/92/api/json?depth=0'[Pipeline] readJSON[Pipeline] sh
| gharchive/pull-request | 2024-02-16T11:28:13 | 2025-04-01T04:33:29.183888 | {
"authors": [
"cimbalek",
"kie-ci3"
],
"repo": "apache/incubator-kie-kogito-images",
"url": "https://github.com/apache/incubator-kie-kogito-images/pull/1742",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1416402811 | Flaky Test TestSlotMigrateFromSlave
Search before asking
[X] I had searched in the issues and found no similar issues.
Motivation
--- FAIL: TestSlotMigrateFromSlave (2.36s)
server.go:106:
Error Trace: /home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/server.go:106
/home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/server.go:112
/home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/server.go:101
/home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:47
/home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:65
Error: Received unexpected error:
signal: aborted (core dumped)
Test: TestSlotMigrateFromSlave
FAIL
Solution
Perhaps we should upload server output on test failures? cc @PragmaTwice
Are you willing to submit a PR?
[ ] I'm willing to submit a PR!
https://github.com/apache/incubator-kvrocks/actions/runs/3288512905/jobs/5418990199
Looks crash after sending the terminate signal, but have no idea to reproduce it.
https://github.com/apache/incubator-kvrocks/actions/runs/3314140570/jobs/5474097740 Another flaky case
@git-hulk they're somehow different issues - not the exact failure cause, or not the exact test case:
--- FAIL: TestSlotMigrateDataType (85.48s)
--- FAIL: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL (60.85s)
slotmigrate_test.go:703:
Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:703
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:625
Error: Condition never satisfied
Test: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL
FAIL
exit status 1
--- FAIL: TestPubSub (1.20s)
server.go:106:
Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:106
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:112
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:101
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/pubsub_test.go:369
Error: Received unexpected error:
signal: abort trap
Test: TestPubSub
FAIL
exit status 1
@git-hulk they're somehow different issues - not the exact failure cause, or not the exact test case:
--- FAIL: TestSlotMigrateDataType (85.48s)
--- FAIL: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL (60.85s)
slotmigrate_test.go:703:
Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:703
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:625
Error: Condition never satisfied
Test: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL
FAIL
exit status 1
--- FAIL: TestPubSub (1.20s)
server.go:106:
Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:106
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:112
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:101
/Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/pubsub_test.go:369
Error: Received unexpected error:
signal: abort trap
Test: TestPubSub
FAIL
exit status 1
Yes, for the TestPusSub test case, it should be the same with TestSlotMigrateFromSlave, there're core when exiting the server.
@git-hulk I'm thinking of uploading the log files if there's a test failure...Don't have a concrete plan yet, though.
I'm also thinking how to download the log file if it cored
| gharchive/issue | 2022-10-20T10:51:12 | 2025-04-01T04:33:29.190564 | {
"authors": [
"git-hulk",
"tisonkun"
],
"repo": "apache/incubator-kvrocks",
"url": "https://github.com/apache/incubator-kvrocks/issues/1016",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1512864489 | error command for: georadius
Search before asking
[X] I had searched in the issues and found no similar issues.
Version
latest
Minimal reproduce step
example at : https://redis.io/commands/georadius/
redis> GEOADD Sicily 13.361389 38.115556 "Palermo" 15.087269 37.502669 "Catania"
(integer) 2
What did you expect to see?
redis> GEORADIUS Sicily 15 37 200 km WITHDIST WITHCOORD
"Catania"
"56.4413"
What did you see instead?
Expected :56.4413
Actual :3.479447370796909E15
Anything Else?
No response
Are you willing to submit a PR?
[ ] I'm willing to submit a PR!
@lt1946 Thanks for your report, but I can't reproduce the bug with your steps.
| gharchive/issue | 2022-12-28T14:43:30 | 2025-04-01T04:33:29.194792 | {
"authors": [
"git-hulk",
"lt1946"
],
"repo": "apache/incubator-kvrocks",
"url": "https://github.com/apache/incubator-kvrocks/issues/1213",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1091051758 | Turn up maven options in mvn script
Why are the changes needed?
How was this patch tested?
[ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible
[ ] Add screenshots for manual tests if appropriate
[ ] Run test locally before make a pull request
Codecov Report
Merging #1662 (4d31098) into master (8e9e8c3) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #1662 +/- ##
=========================================
Coverage 59.57% 59.57%
Complexity 193 193
=========================================
Files 265 265
Lines 13084 13084
Branches 1649 1649
=========================================
Hits 7795 7795
Misses 4631 4631
Partials 658 658
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 8e9e8c3...4d31098. Read the comment docs.
thanks, merged to master
| gharchive/pull-request | 2021-12-30T11:40:13 | 2025-04-01T04:33:29.203196 | {
"authors": [
"codecov-commenter",
"yaooqinn"
],
"repo": "apache/incubator-kyuubi",
"url": "https://github.com/apache/incubator-kyuubi/pull/1662",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1215790424 | [KYUUBI #2484] Add conf to SessionEvent and display it in EngineSessionPage
Why are the changes needed?
close #2484
How was this patch tested?
[ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible
[X] Add screenshots for manual tests if appropriate
[ ] Run test locally before make a pull request
Codecov Report
Merging #2485 (5c3bb30) into master (6187321) will decrease coverage by 0.07%.
The diff coverage is 7.69%.
@@ Coverage Diff @@
## master #2485 +/- ##
============================================
- Coverage 63.45% 63.37% -0.08%
Complexity 69 69
============================================
Files 371 371
Lines 17585 17608 +23
Branches 2348 2348
============================================
+ Hits 11158 11159 +1
- Misses 5392 5416 +24
+ Partials 1035 1033 -2
Impacted Files
Coverage Δ
.../scala/org/apache/spark/ui/EngineSessionPage.scala
2.15% <0.00%> (-0.75%)
:arrow_down:
...ache/kyuubi/engine/spark/events/SessionEvent.scala
84.00% <100.00%> (+1.39%)
:arrow_up:
...rg/apache/kyuubi/engine/spark/SparkSQLEngine.scala
77.38% <0.00%> (+0.77%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 6187321...5c3bb30. Read the comment docs.
need more unit tests
Thanks, I'll add it later
nice feature, thanks
thanks, merged to master and branch-1.5
| gharchive/pull-request | 2022-04-26T11:07:47 | 2025-04-01T04:33:29.215566 | {
"authors": [
"codecov-commenter",
"turboFei",
"wForget"
],
"repo": "apache/incubator-kyuubi",
"url": "https://github.com/apache/incubator-kyuubi/pull/2485",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1249179643 | [DOC] Replace sphinx_rtd_theme with sphinx_book_theme
Why are the changes needed?
make documentation style to fit official web site
How was this patch tested?
[ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible
[x] Add screenshots for manual tests if appropriate
[x] Run test locally before make a pull request
Codecov Report
Merging #2751 (b812352) into master (b40bcbd) will decrease coverage by 2.49%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #2751 +/- ##
============================================
- Coverage 64.24% 61.74% -2.50%
- Complexity 82 385 +303
============================================
Files 385 433 +48
Lines 18673 20345 +1672
Branches 2531 2760 +229
============================================
+ Hits 11997 12563 +566
- Misses 5532 6544 +1012
- Partials 1144 1238 +94
Impacted Files
Coverage Δ
...pache/kyuubi/sql/KyuubiSQLExtensionException.scala
0.00% <0.00%> (-100.00%)
:arrow_down:
...rg/apache/kyuubi/sql/zorder/ZorderBytesUtils.scala
0.00% <0.00%> (-94.74%)
:arrow_down:
...ubi/sql/zorder/InsertZorderBeforeWritingBase.scala
0.00% <0.00%> (-89.16%)
:arrow_down:
...cala/org/apache/kyuubi/sql/zorder/ZorderBase.scala
0.00% <0.00%> (-87.50%)
:arrow_down:
.../kyuubi/sql/zorder/OptimizeZorderCommandBase.scala
0.00% <0.00%> (-81.82%)
:arrow_down:
...yuubi/sql/zorder/OptimizeZorderStatementBase.scala
0.00% <0.00%> (-71.43%)
:arrow_down:
...g/apache/kyuubi/sql/zorder/ResolveZorderBase.scala
0.00% <0.00%> (-69.57%)
:arrow_down:
...n/spark/authz/ranger/FilteredShowObjectsExec.scala
42.10% <0.00%> (-37.90%)
:arrow_down:
...uubi/engine/flink/operation/ExecuteStatement.scala
47.61% <0.00%> (-36.80%)
:arrow_down:
...pache/kyuubi/sql/KyuubiQueryStagePreparation.scala
0.00% <0.00%> (-32.36%)
:arrow_down:
... and 112 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update b40bcbd...b812352. Read the comment docs.
+1 looks more structured!
Thanks, merging to master
| gharchive/pull-request | 2022-05-26T06:58:19 | 2025-04-01T04:33:29.235785 | {
"authors": [
"codecov-commenter",
"pan3793",
"yanghua",
"yaooqinn"
],
"repo": "apache/incubator-kyuubi",
"url": "https://github.com/apache/incubator-kyuubi/pull/2751",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1484571320 | Print the simple batch report if no log fetch to prevent misunderstand
Why are the changes needed?
With yarn-cluster mode, if spark.yarn.submit.waitAppCompletion is false and kyuubi-ctl wait completion is true, we need still print some simple batch report to prevent misunderstand.
The batch report format refer spark-submit log.
Application report for <ApplicationID> (state: RUNNING)
How was this patch tested?
[ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible
[ ] Add screenshots for manual tests if appropriate
[x] Run test locally before make a pull request
Codecov Report
Merging #3949 (9580277) into master (8305d80) will decrease coverage by 0.08%.
The diff coverage is 0.00%.
:exclamation: Current head 9580277 differs from pull request most recent head ea5597f. Consider uploading reports for the commit ea5597f to get more accurate results
@@ Coverage Diff @@
## master #3949 +/- ##
============================================
- Coverage 51.90% 51.82% -0.09%
Complexity 13 13
============================================
Files 521 521
Lines 28843 28807 -36
Branches 3864 3856 -8
============================================
- Hits 14972 14930 -42
- Misses 12494 12505 +11
+ Partials 1377 1372 -5
Impacted Files
Coverage Δ
...rg/apache/kyuubi/ctl/cmd/log/LogBatchCommand.scala
60.93% <0.00%> (-3.00%)
:arrow_down:
...g/apache/kyuubi/session/KyuubiSessionManager.scala
88.13% <0.00%> (-5.15%)
:arrow_down:
...g/apache/kyuubi/operation/BatchJobSubmission.scala
74.64% <0.00%> (-3.79%)
:arrow_down:
...apache/kyuubi/engine/JpsApplicationOperation.scala
77.41% <0.00%> (-3.23%)
:arrow_down:
.../apache/kyuubi/server/api/v1/BatchesResource.scala
69.00% <0.00%> (-3.00%)
:arrow_down:
...ache/kyuubi/server/KyuubiRestFrontendService.scala
82.95% <0.00%> (-0.75%)
:arrow_down:
...he/kyuubi/ha/client/etcd/EtcdDiscoveryClient.scala
68.50% <0.00%> (-0.56%)
:arrow_down:
...apache/kyuubi/session/KyuubiBatchSessionImpl.scala
90.66% <0.00%> (-0.13%)
:arrow_down:
...cala/org/apache/kyuubi/server/ui/JettyServer.scala
86.95% <0.00%> (ø)
...in/scala/org/apache/kyuubi/config/KyuubiConf.scala
97.50% <0.00%> (+0.06%)
:arrow_up:
... and 4 more
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
thanks, merged to master and 1.6.2
| gharchive/pull-request | 2022-12-08T12:49:32 | 2025-04-01T04:33:29.253797 | {
"authors": [
"codecov-commenter",
"turboFei"
],
"repo": "apache/incubator-kyuubi",
"url": "https://github.com/apache/incubator-kyuubi/pull/3949",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1464545624 | update how-to-email.md
email-reply.png The path is wrongly written and submitted
LGTM
| gharchive/pull-request | 2022-11-25T13:22:43 | 2025-04-01T04:33:29.255669 | {
"authors": [
"GuoPhilipse",
"binbinCheng"
],
"repo": "apache/incubator-linkis-website",
"url": "https://github.com/apache/incubator-linkis-website/pull/610",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1345746316 | [Question]
Before asking
[X] I have read the How to ask for help (我已阅读如何寻求帮助指引文档).
Your environment
Linkis version used: 1.1.2
Environment name and version:
hdp-3.1.0
hive-3.1.0
spark-2.4.3
scala-2.11.12
jdk 1.8.0_161-b14
....
Describe your questions
Q1. ...
在dss中使用hive建表语法
create table test_78 as
select
*
from
*.;
时候报错,在后台用hive 登录执行以上语句正常执行。
Eureka service list
eg:
Some logs info or acctch file
linkis-xxx.log:
ERROR [Linkis-Default-Scheduler-Thread-19] org.apache.hadoop.hive.ql.Driver 1247 printError - FAILED: Error in acquiring locks: calling recordValidTxn() more than once in the same txnid:705363368
java.lang.IllegalStateException: calling recordValidTxn() more than once in the same txnid:705363368
at org.apache.hadoop.hive.ql.Driver.acquireLocks(Driver.java:1614)
at org.apache.hadoop.hive.ql.Driver.lockAndRespond(Driver.java:1795)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1965)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.linkis.engineplugin.hive.executor.HiveDriverProxy.run(HiveEngineConnExecutor.scala:523)
at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor.org$apache$linkis$engineplugin$hive$executor$HiveEngineConnExecutor$$executeHQL(HiveEngineConnExecutor.scala:195)
at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor$$anon$1.run(HiveEngineConnExecutor.scala:146)
at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor$$anon$1.run(HiveEngineConnExecutor.scala:139)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor.executeLine(HiveEngineConnExecutor.scala:139)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10$$anonfun$apply$11.apply(ComputationExecutor.scala:183)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10$$anonfun$apply$11.apply(ComputationExecutor.scala:182)
at org.apache.linkis.common.utils.Utils$.tryCatch(Utils.scala:40)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10.apply(ComputationExecutor.scala:184)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10.apply(ComputationExecutor.scala:178)
at scala.collection.immutable.Range.foreach(Range.scala:160)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2.apply(ComputationExecutor.scala:177)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2.apply(ComputationExecutor.scala:153)
at org.apache.linkis.common.utils.Utils$.tryFinally(Utils.scala:61)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor.toExecuteTask(ComputationExecutor.scala:230)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$3.apply(ComputationExecutor.scala:245)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$3.apply(ComputationExecutor.scala:245)
at org.apache.linkis.common.utils.Utils$.tryFinally(Utils.scala:61)
at org.apache.linkis.engineconn.acessible.executor.entity.AccessibleExecutor.ensureIdle(AccessibleExecutor.scala:55)
at org.apache.linkis.engineconn.acessible.executor.entity.AccessibleExecutor.ensureIdle(AccessibleExecutor.scala:49)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor.ensureOp(ComputationExecutor.scala:137)
at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor.execute(ComputationExecutor.scala:244)
at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl.org$apache$linkis$engineconn$computation$executor$service$TaskExecutionServiceImpl$$executeTask(TaskExecutionServiceImpl.scala:296)
at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2$$anonfun$run$2.apply$mcV$sp(TaskExecutionServiceImpl.scala:229)
at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2$$anonfun$run$2.apply(TaskExecutionServiceImpl.scala:227)
at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2$$anonfun$run$2.apply(TaskExecutionServiceImpl.scala:227)
at org.apache.linkis.common.utils.Utils$.tryCatch(Utils.scala:40)
at org.apache.linkis.common.utils.Utils$.tryAndWarn(Utils.scala:69)
at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2.run(TaskExecutionServiceImpl.scala:227)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
log file:
linkis-xxx.log
err.log
Does it happen every time?
Is there any useful information I can get from here? https://www.mail-archive.com/issues@hive.apache.org/msg116947.html
| gharchive/issue | 2022-08-22T03:13:53 | 2025-04-01T04:33:29.264932 | {
"authors": [
"casionone",
"mowenfuhua123"
],
"repo": "apache/incubator-linkis",
"url": "https://github.com/apache/incubator-linkis/issues/2918",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
474848922 | DMLC link removal
Replaced v1.5.x Julia page with the content on master. Also, replaced dmlc.ml links with available content.
@mxnet-label-bot Add [Doc]
Preview: http://100.24.115.165/api/julia/index.html
@aaronmarkham
@marcoabreu Sorry about rushing that through without addressing your questions. We needed to be rid of the malware links ASAP.
@IvyBazan Can you answer? We spoke about this offline and you had some reasoning why it might be ok to just kill the links with a #.
@marcoabreu I removed those links because I couldn't find any replacement for the content that was hosted on dmlc.ml. It seems better to temporarily have replaced the links with a #, while waiting for that content to be recovered and rehosted, rather than having malware links still active on the site. The current experience is that the links redirect to the current page.
Issues related to missing content:
Missing nin content-https://github.com/apache/incubator-mxnet/issues/15729#issuecomment-518131446
Missing nnvm blog: https://github.com/apache/incubator-mxnet/issues/15730#issuecomment-518131577
| gharchive/pull-request | 2019-07-30T22:23:47 | 2025-04-01T04:33:29.269738 | {
"authors": [
"IvyBazan",
"aaronmarkham",
"piyushghai"
],
"repo": "apache/incubator-mxnet",
"url": "https://github.com/apache/incubator-mxnet/pull/15708",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
607241504 | Netdb multiple dns servers
Summary
PR adds capability to add multiple nameservers on run-time even when not using resolv.conf and reset the list of nameservers back to default setting (no nameserver at all or single predefined nameserver in Kconfig). This is useful for applications that change their cellular network frequently.
Also make getaddrinfo re-entrant when doing service name query
Impact
Testing
@juniskane there are two minor nxstyle, please fix them, thanks.
@juniskane , since the 2nd patch generate the nxstyle issues, it's better to merge the 3rd patch into the 2nd one:
git rebase --interactive
change pick to squach for the 3rd patch
git push -f origin netdb_multiple_dns_servers_pr
| gharchive/pull-request | 2020-04-27T05:17:36 | 2025-04-01T04:33:29.272283 | {
"authors": [
"juniskane",
"xiaoxiang781216"
],
"repo": "apache/incubator-nuttx",
"url": "https://github.com/apache/incubator-nuttx/pull/892",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
611887034 | sched/semaphore/sem_holder.c: Remove DEBUGPANIC
Summary
There is a DEBUGPANIC in some logic. This happens if a a task exists at certain points with priority inheritance enabled. This event was not expected in the original design (although logic was provided to support it). Since, apparently, it does happen, the DEBUGPANIC must be removed.
Impact
Should eliminate a crash found by Brennan Ashton using the X86_64 Qemu port
Testing
This makes sense to me.
Should we also remove the first line of the assert in nxsem_restorebaseprio that was disabled and re-enable it. I think this explains the behavior. I think the second half of the assert is correct.
This makes sense to me.
Should we also remove the first line of the assert in nxsem_restorebaseprio that was disabled and re-enable it. I think this explains the behavior. I think the second half of the assert is correct.
I removed the code because I don't like commented out code. It just leaves questions for the next person who deals with the code (and no one to answer the questions). It is always better to be courageous and remove wrong code.
I did re-enable the debug output because I think it still has valid information. I dropped from the error to a warning. I think the output has value if you are debugging with scheduler debug turned off.. although perhaps. It is not a error any more. A warning is for unusual events that you might want to know about while debugging. info() is just chit-char, flow of control stuff.
So generally, my preference would be to leave the warning in place because it provides information for people who are silly enough to use scheduler debug output. I really doesn't work. If you enable the scheduler debug output, the system becomes very unstable. You can't really use the OS to monitor itself in the that way. syslog output also requires the scheduler.
This makes sense to me.
Should we also remove the first line of the assert in nxsem_restorebaseprio that was disabled and re-enable it. I think this explains the behavior. I think the second half of the assert is correct.
I removed the code because I don't like commented out code. It just leaves questions for the next person who deals with the code (and no one to answer the questions). It is always better to be courageous and remove wrong code.
I did re-enable the debug output because I think it still has valid information. I dropped from the error to a warning. I think the output has value if you are debugging with scheduler debug turned off.. although perhaps. It is not a error any more. A warning is for unusual events that you might want to know about while debugging. info() is just chit-char, flow of control stuff.
So generally, my preference would be to leave the warning in place because it provides information for people who are silly enough to use scheduler debug output. I really doesn't work. If you enable the scheduler debug output, the system becomes very unstable. You can't really use the OS to monitor itself in the that way. syslog output also requires the scheduler.
These are bogus build failures. The macOS builds fail a large percentage of the time when there is no problem with the changes. I have gotten so that I ignore these tests because the are not reliable indicators of anything. I will restart the checks. maybe we will be lucky next time.
| gharchive/pull-request | 2020-05-04T13:45:24 | 2025-04-01T04:33:29.278392 | {
"authors": [
"btashton",
"patacongo"
],
"repo": "apache/incubator-nuttx",
"url": "https://github.com/apache/incubator-nuttx/pull/967",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1989952643 | RFC: Concurrent Stat in List
Description
relate: #3097
Inviting more committers to review the RFC.
| gharchive/pull-request | 2023-11-13T06:08:19 | 2025-04-01T04:33:29.279559 | {
"authors": [
"Xuanwo",
"morristai"
],
"repo": "apache/incubator-opendal",
"url": "https://github.com/apache/incubator-opendal/pull/3574",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
334605673 | Allow for activation store to accept user and request information
Other ActivationStore implementations, Elasticsearch for instance, will need user and request information.
Description
Related issue and scope
[ ] I opened an issue to propose and discuss this change (#????)
My changes affect the following components
[ ] API
[ ] Controller
[ ] Message Bus (e.g., Kafka)
[ ] Loadbalancer
[ ] Invoker
[ ] Intrinsic actions (e.g., sequences, conductors)
[ ] Data stores (e.g., CouchDB)
[ ] Tests
[ ] Deployment
[ ] CLI
[ ] General tooling
[ ] Documentation
Types of changes
[ ] Bug fix (generally a non-breaking change which closes an issue).
[ ] Enhancement or new feature (adds new functionality).
[ ] Breaking change (a bug fix or enhancement which changes existing behavior).
Checklist:
[x] I signed an Apache CLA.
[x] I reviewed the style guides and followed the recommendations (Travis CI will check :).
[ ] I added tests to cover my changes.
[ ] My changes require further changes to the documentation.
[ ] I updated the documentation where necessary.
PG4 1863 ⏳
Codecov Report
Merging #3798 into master will increase coverage by 0.01%.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #3798 +/- ##
==========================================
+ Coverage 75.15% 75.16% +0.01%
==========================================
Files 132 132
Lines 6142 6145 +3
Branches 373 380 +7
==========================================
+ Hits 4616 4619 +3
Misses 1526 1526
Impacted Files
Coverage Δ
...la/whisk/core/entity/ArtifactActivationStore.scala
92.85% <ø> (ø)
:arrow_up:
...main/scala/whisk/core/controller/Activations.scala
96.77% <100%> (+0.16%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 92a64c2...0140629. Read the comment docs.
@markusthoemmes, very concerned about this polling behavior. https://github.com/apache/incubator-openwhisk/blob/1d79fa395de32c6a774efeb4f53563689e9cbb7e/core/controller/src/main/scala/whisk/core/controller/actions/PrimitiveActions.scala#L585
@dubee concerned in which way? I believe this is the database fallback polling in case an active ack goes missing.
@dubee Can you share some more details on how Elasticsearch implementation would make use of user and request data. Would it like to use any aspect of request or we can extract explicit parts from request and pass it on.
For now it appears that passing request param is making SPI usage more coupled to web requests
@chetanmeh, certain Elasticsearch deployments require additional authentication which can be passed through the controller via HTTP headers. This allows the controller to communicate to Elasitcsearch for users making requests. Currently our Elasticsearch log store works in the same way. When required headers are specified via the OpenWhisk Ansible configuration for the log store, the log store will pass those headers from the client to Elasticsearch to perform a query.
See the log store for example:
https://github.com/apache/incubator-openwhisk/blob/92a64c291156a2cd3d6b304babc2a193a46d0699/common/scala/src/main/scala/whisk/core/containerpool/logging/ElasticSearchLogStore.scala#L103
@dubee Now its more clear. So we pass on some sort of "user context" to ES. May be we make that notion explicit via having a UserContext as parameter type
case class UserContext(user:Option[Identity], request: Option[HttpRequest])
- def get(activationId: ActivationId, user: Option[Identity] = None, request: Option[HttpRequest] = None)(
- def get(activationId: ActivationId, userContext: Option[UserContext] = None)(
PG2 3475 🔵
@markusthoemmes, anything else on this one?
@chetanmeh, where do you think the UserContext case class should live?
Added UserContext in the third commit. The case class definition should probably be moved elsewhere though.
PG4 2057 🔵
where do you think the UserContext case class should live
Adding it to ActivationStore looks fine
@markusthoemmes, any other comments on this one?
@mdeuser, can you review this one?
| gharchive/pull-request | 2018-06-21T18:23:38 | 2025-04-01T04:33:29.298428 | {
"authors": [
"chetanmeh",
"codecov-io",
"dubee",
"markusthoemmes"
],
"repo": "apache/incubator-openwhisk",
"url": "https://github.com/apache/incubator-openwhisk/pull/3798",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2145734474 | [Spark]Spark sql support get/list function
Spark sql support get/list function
CC @YannByron
LGTM. @waywtdcc please rebase master and fix the spotless style.
LGTM. @waywtdcc please rebase master and fix the spotless style.
ok
+1
| gharchive/pull-request | 2024-02-21T03:29:24 | 2025-04-01T04:33:29.300844 | {
"authors": [
"JingsongLi",
"YannByron",
"waywtdcc"
],
"repo": "apache/incubator-paimon",
"url": "https://github.com/apache/incubator-paimon/pull/2880",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
811591448 | Controller Crashing - repetition of same config causing multi value setting
I have following settings in controller config where validation.frequencyInSeconds is accidentally repeated. This is causing controller to crash as the value is derived as multi value. it will be better to take either last defined value or avoid this crash.
controller.realtime.segment.validation.frequencyInSeconds=300
controller.realtime.segment.validation.frequencyInSeconds=900
controller.broker.resource.validation.frequencyInSeconds=900
Error:
2021/02/18 14:46:44.389 ERROR [StartServiceManagerCommand] [main] Failed to start a Pinot [CONTROLLER] at 39.246 since launch
java.lang.NumberFormatException: For input string: "[300, 900]"
at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) ~[?:1.8.0_282]
at java.lang.Integer.parseInt(Integer.java:580) ~[?:1.8.0_282]
I feel we should try to validate those configures and provide more meaningful error msgs, e.g. the config controller.realtime.segment.validation.frequencyInSeconds is invalid.
| gharchive/issue | 2021-02-19T00:52:39 | 2025-04-01T04:33:29.303798 | {
"authors": [
"fx19880617",
"pabrahamusa"
],
"repo": "apache/incubator-pinot",
"url": "https://github.com/apache/incubator-pinot/issues/6595",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
624126880 | Add UDFs for String Transformation
Functions Added in Function Registry
Generic Transform Function modified to handle more types
Missing:
The return type of Generic Transform function is still restricted to String.
This is great!
Could you add tests in: CalciteSqlCompilerTest to check sql parsing logic.
@fx19880617 @siddharthteotia Should I add tests in CalciteSQL for all the functions?
@fx19880617 @siddharthteotia Should I add tests in CalciteSQL for all the functions?
@KKcorps , sorry missed seeing this. Yes, the query compilation tests should be in CalciteSqlCompilerTest. Here we can verify that PinotQuery is being built correctly and that gets converted to BrokerRequest correctly. Most other tests in this file do this.
The other suggestion was to also add unit tests for exercising end-to-end execution path. Please consider adding these tests to an appropriate file in /incubator-pinot/pinot-core/src/test/java/org/apache/pinot/queries/. May be TransformQueriesTest
Is the plan to use this wrapper solely for invoking scalar functions (like already done in this PR for StringFunctions) or are we expecting follow-ups to integrate it with rest of the transform functions. I think only the former?
@sidd only for scalarfunctions.
| gharchive/pull-request | 2020-05-25T08:12:52 | 2025-04-01T04:33:29.307502 | {
"authors": [
"KKcorps",
"fx19880617",
"kishoreg",
"siddharthteotia"
],
"repo": "apache/incubator-pinot",
"url": "https://github.com/apache/incubator-pinot/pull/5440",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
311675274 | Rest endpoint to query compaction status
Currently returns whether compaction has not run, is running, is
complete or has failed. If it has failed, the last error is returned
also.
retest this please // failing on the mock issue
retest this please // flake in c++ AuthPluginTest.testAthenz
retest this please // PersistentQueueE2ETest.testCancelReadRequestOnLastDisconnect
retest this please // many failures this time
org.apache.pulsar.broker.service.PersistentFailoverE2ETest.testSimpleConsumerEventsWithoutPartition
org.apache.pulsar.broker.service.ReplicatorTest.testCloseReplicatorStartProducer
org.apache.pulsar.client.impl.BrokerClientIntegrationTest.testUnsupportedBatchMessageConsumer
| gharchive/pull-request | 2018-04-05T16:10:04 | 2025-04-01T04:33:29.310350 | {
"authors": [
"ivankelly"
],
"repo": "apache/incubator-pulsar",
"url": "https://github.com/apache/incubator-pulsar/pull/1501",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
721558295 | RATIS-1097. DataStreamOutputImpl should use streamOffset instead of messageId
See https://issues.apache.org/jira/browse/RATIS-1097
@runzhiwang , thanks for reviewing the change. I just have addressed your comments. Could you review it again?
@szetszwo LGTM, +1. Just wait the fix of checkstyle.
Reopen PR to trigger CI.
@szetszwo The failed testAsyncConfiguration seems related to this PR.
Oops, I have accidentally changed OrderedAsync.java. It should be OrderedStreamAsync.java
@szetszwo Thanks the patch, @amaliujia Thanks for review. I have merged it.
| gharchive/pull-request | 2020-10-14T15:25:18 | 2025-04-01T04:33:29.313246 | {
"authors": [
"runzhiwang",
"szetszwo"
],
"repo": "apache/incubator-ratis",
"url": "https://github.com/apache/incubator-ratis/pull/224",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
577524266 | Decouple sharding-proxy-transport's dependencies
For #4594.
Pull Request Test Coverage Report for Build 10119
2 of 2 (100.0%) changed or added relevant lines in 1 file are covered.
3 unchanged lines in 1 file lost coverage.
Overall coverage decreased (-0.005%) to 59.047%
Files with Coverage Reduction
New Missed Lines
%
sharding-orchestration/sharding-orchestration-core/src/main/java/org/apache/shardingsphere/orchestration/internal/util/IpUtils.java
3
76.0%
Totals
Change from base Build 1021:
-0.005%
Covered Lines:
12113
Relevant Lines:
20514
💛 - Coveralls
| gharchive/pull-request | 2020-03-08T16:09:22 | 2025-04-01T04:33:29.319508 | {
"authors": [
"coveralls",
"terrymanu"
],
"repo": "apache/incubator-shardingsphere",
"url": "https://github.com/apache/incubator-shardingsphere/pull/4664",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1083051248 | [BUG] abnormal return of file request. serious
Is there an existing issue for this?
[X] I have searched the existing issues
Current Behavior
Service A provides a file / picture download.
shenyu-gateway return is an abnormal file character stream.
Expected Behavior
shenyu-gateway will return file stream.
Steps To Reproduce
No response
Environment
ShenYu version(s):2.4.2
Debug logs
I don't think the file return should be converted to string.
Anything else?
No response
I can fix this problem.
Can you provide responseHeader?
| gharchive/issue | 2021-12-17T09:25:01 | 2025-04-01T04:33:29.323150 | {
"authors": [
"MarsOu1995",
"impactCn"
],
"repo": "apache/incubator-shenyu",
"url": "https://github.com/apache/incubator-shenyu/issues/2611",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1305539139 | [type:refactor] refactor-spring-cloud-loadbalancer
Make sure that:
[ ] You have read the contribution guidelines.
[ ] You submit test cases (unit or integration tests) that back your changes.
[ ] Your local test passed ./mvnw clean install -Dmaven.javadoc.skip=true.
Codecov Report
Merging #3714 (9627a7a) into master (ec17f7e) will increase coverage by 0.07%.
The diff coverage is 88.73%.
@@ Coverage Diff @@
## master #3714 +/- ##
============================================
+ Coverage 63.42% 63.49% +0.07%
+ Complexity 5947 5938 -9
============================================
Files 897 897
Lines 24681 24630 -51
Branches 2249 2238 -11
============================================
- Hits 15653 15638 -15
+ Misses 7607 7568 -39
- Partials 1421 1424 +3
Impacted Files
Coverage Δ
...d/loadbalance/ShenyuSpringCloudServiceChooser.java
87.69% <87.69%> (ø)
...e/shenyu/plugin/springcloud/SpringCloudPlugin.java
94.11% <100.00%> (+23.52%)
:arrow_up:
...in/springcloud/SpringCloudPluginConfiguration.java
100.00% <100.00%> (ø)
...yu/sync/data/http/refresh/AbstractDataRefresh.java
68.96% <0.00%> (-13.80%)
:arrow_down:
...a/org/apache/shenyu/common/utils/VersionUtils.java
67.85% <0.00%> (-10.72%)
:arrow_down:
...henyu/admin/service/impl/UpstreamCheckService.java
54.66% <0.00%> (-9.34%)
:arrow_down:
...rg/apache/shenyu/plugin/rewrite/RewritePlugin.java
66.66% <0.00%> (-5.56%)
:arrow_down:
...in/response/strategy/NettyClientMessageWriter.java
66.66% <0.00%> (-4.17%)
:arrow_down:
...che/shenyu/sync/data/http/HttpSyncDataService.java
81.63% <0.00%> (-4.09%)
:arrow_down:
...icsearch/client/ElasticSearchLogCollectClient.java
70.00% <0.00%> (-2.00%)
:arrow_down:
... and 28 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update ec17f7e...9627a7a. Read the comment docs.
| gharchive/pull-request | 2022-07-15T03:50:19 | 2025-04-01T04:33:29.341763 | {
"authors": [
"codecov-commenter",
"moremind"
],
"repo": "apache/incubator-shenyu",
"url": "https://github.com/apache/incubator-shenyu/pull/3714",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
389696941 | 请问什么时候可以兼容spring cloud gateway呢?
Please answer these questions before submitting your issue.
Why do you submit this issue?
[ ] Question or discussion
[ ] Bug
[ ] Requirement
[ ] Feature or performance improvement
Question
What do you want to know?
Bug
Which version of SkyWalking, OS and JRE?
Which company or project?
What happen?
If possible, provide a way for reproducing the error. e.g. demo application, component version.
Requirement or improvement
Please describe about your requirements or improvement suggestions.
English only. Requirement only accept when some contributor wants it.
spring cloud gateway 现在不兼容
| gharchive/issue | 2018-12-11T10:35:55 | 2025-04-01T04:33:29.345857 | {
"authors": [
"github24518",
"monkeyAlwaysWin",
"wu-sheng"
],
"repo": "apache/incubator-skywalking",
"url": "https://github.com/apache/incubator-skywalking/issues/2034",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
428581362 | Increase the test coverage rate
Today, SkyWalking's tests are mostly around integration tests, which are benifit for end users. Now we hope through this, people from community could help on Unit Tests.
JUnit and Mockito are provided. This is a good start for everyone wants to contribute codes. When you submit test cases PR, please ref to this issue.
Also, in this kind of pull request, we don't limit how many test cases you provide. Even one new test case, which could effectively test one more line, it will be acceptable. Just try your best to make tests more sense.
Where can I find the code coverage report?
The coverage is here, https://coveralls.io/github/apache/skywalking
Update here, we are only just increasing the test case coverage, we are doing more.
[x] @kezhenxu94 set up the Jenkins of ASF for power test env.
[ ] @ascrutae lead the agent test in Jenkins jobs of every PR when necessary
[ ] @kezhenxu94 lead the e2e test, make agent/backend integration test done in every PR.
We have e2e tests fro single node w/ H2 and cluster mode w/ ES. We are trying to build more. Including MySQL storage tests, alarm tests, and aggregation accuracy tests.
After #3561, we have 30%+ coverage now.
@ascrutae lead the agent test in Jenkins jobs of every PR when necessary
This task is lead by @dmsolr on #3528. Ownership replaced.
Agent autotest framework has been provided, and 300+ test cases have been added. Seems working well. Just working for @dmsolr to provide the official documents, rather than guide users to do that one by one.
| gharchive/issue | 2019-04-03T06:04:35 | 2025-04-01T04:33:29.350512 | {
"authors": [
"flycash",
"wu-sheng"
],
"repo": "apache/incubator-skywalking",
"url": "https://github.com/apache/incubator-skywalking/issues/2438",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
742904233 | docs: fix docker config paths
SUMMARY
Just pointin' people in the right direction
BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF
TEST PLAN
ADDITIONAL INFORMATION
[ ] Has associated issue:
[ ] Changes UI
[ ] Requires DB Migration.
[ ] Confirm DB Migration upgrade and downgrade tested.
[ ] Introduces new feature or API
[ ] Removes existing feature or API
this is a docs change and e2e has failed 3 times :(
xit!
Codecov Report
Merging #11703 (942b6f6) into master (68693c7) will decrease coverage by 0.04%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #11703 +/- ##
==========================================
- Coverage 62.86% 62.81% -0.05%
==========================================
Files 889 454 -435
Lines 43055 16185 -26870
Branches 4017 4017
==========================================
- Hits 27065 10167 -16898
+ Misses 15811 5839 -9972
Partials 179 179
Flag
Coverage Δ
javascript
62.81% <ø> (ø)
python
?
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
...t/annotation_layers/annotations/commands/create.py
superset/examples/multi_line.py
superset/queries/filters.py
superset/datasets/commands/refresh.py
...uperset/migrations/versions/1d2ddd543133_log_dt.py
...s/versions/3c3ffe173e4f_add_sql_string_to_table.py
superset/stats_logger.py
superset/db_engine_specs/redshift.py
superset/examples/misc_dashboard.py
superset/db_engine_specs/bigquery.py
... and 425 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 68693c7...942b6f6. Read the comment docs.
| gharchive/pull-request | 2020-11-14T02:19:37 | 2025-04-01T04:33:29.367002 | {
"authors": [
"codecov-io",
"mistercrunch",
"suddjian"
],
"repo": "apache/incubator-superset",
"url": "https://github.com/apache/incubator-superset/pull/11703",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
342021375 | Migrated to click
Fixed merge conflicts and linting errors from https://github.com/apache/incubator-superset/pull/4078.
Please fix the merge conflicts and build errors
Codecov Report
Merging #5410 into master will increase coverage by <.01%.
The diff coverage is 92.85%.
@@ Coverage Diff @@
## master #5410 +/- ##
=========================================
+ Coverage 59.09% 59.1% +<.01%
=========================================
Files 372 372
Lines 23747 23759 +12
Branches 2758 2758
=========================================
+ Hits 14033 14042 +9
- Misses 9699 9702 +3
Partials 15 15
Impacted Files
Coverage Δ
superset/cli.py
46.59% <92.85%> (+1.9%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 99ce7b7...aa6d503. Read the comment docs.
@mistercrunch Fixed.
Can we now remove flask-script from our setup.py & requeirments.txt ?
@mistercrunch Oh, yeah good call. Sorry about that.
| gharchive/pull-request | 2018-07-17T17:51:35 | 2025-04-01T04:33:29.375524 | {
"authors": [
"codecov-io",
"gbates101",
"mistercrunch"
],
"repo": "apache/incubator-superset",
"url": "https://github.com/apache/incubator-superset/pull/5410",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
440256903 | feat: Scheduling queries from SQL Lab (#7416)
merged from commit 90eef51 on lyft-release-sp8
@betodealmeida @xtinec
Codecov Report
Merging #7446 into master will decrease coverage by 0.03%.
The diff coverage is 52%.
@@ Coverage Diff @@
## master #7446 +/- ##
==========================================
- Coverage 65.26% 65.22% -0.04%
==========================================
Files 430 431 +1
Lines 21078 21147 +69
Branches 2338 2343 +5
==========================================
+ Hits 13757 13794 +37
- Misses 7205 7237 +32
Partials 116 116
Impacted Files
Coverage Δ
...src/components/FilterableTable/FilterableTable.jsx
91.01% <ø> (ø)
:arrow_up:
.../assets/src/SqlLab/components/QueryAutoRefresh.jsx
11.9% <0%> (-0.92%)
:arrow_down:
superset/views/sql_lab.py
94.64% <100%> (ø)
:arrow_up:
superset/assets/src/featureFlags.ts
88.88% <100%> (+1.38%)
:arrow_up:
...sets/src/SqlLab/components/ScheduleQueryButton.jsx
13.33% <13.33%> (ø)
...uperset/assets/src/SqlLab/components/SqlEditor.jsx
59.8% <50%> (-0.2%)
:arrow_down:
superset/db_engine_specs.py
61.45% <86.11%> (+1.22%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update c8bb7e0...4b5e54f. Read the comment docs.
| gharchive/pull-request | 2019-05-03T23:25:13 | 2025-04-01T04:33:29.387073 | {
"authors": [
"DiggidyDave",
"codecov-io"
],
"repo": "apache/incubator-superset",
"url": "https://github.com/apache/incubator-superset/pull/7446",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
613817538 | Fusion base directory is not mounted
I see u guys made a lot of changes. And I cloned ur the newest code to give it a shot. I have to say this is much better than the old version. But when I was try to launch Teaclave Services with docker-compose command, I got a situation here. Like below.
teaclave-execution-service | [2020-05-07T04:25:14Z ERROR teaclave_execution_service_enclave] Failed to start the service: Fusion base directory is not mounted: /tmp/fusion_data
teaclave-execution-service | Service exit with error: ServiceError
teaclave-execution-service exited with code 0
It seems the execution-service is down because of a mount thing. Waiting for any help. BTW, if I run it successfully, what would I see when I hit the link(https://api.trustedservices.intel.com:443). I don not see any doc about it. BTW AGAIN, the new docs ur uploaded explain my lots of questions.
"Fusion base directory" is a directory for store temporary (encrypted) fusion data and can be set in the runtime config (https://github.com/apache/incubator-teaclave/blob/master/config/runtime.config.toml#L34).
The issue is caused by a check in the execution service (https://github.com/apache/incubator-teaclave/blob/master/services/execution/enclave/src/lib.rs#L66). You can create a directory by your self in /tmp/fusion_data to start the service.
Thanks for your report. We will propose a better implementation for this case.
At last, the link you mentioned (https://api.trustedservices.intel.com:443) is the URL of attestation service. API specifications can be found here: https://software.intel.com/content/dam/develop/public/us/en/documents/sgx-attestation-api-spec.pdf.
| gharchive/issue | 2020-05-07T06:49:56 | 2025-04-01T04:33:29.392006 | {
"authors": [
"czzmmc",
"mssun"
],
"repo": "apache/incubator-teaclave",
"url": "https://github.com/apache/incubator-teaclave/issues/287",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
189362585 | Fix TC-43: TR prints 'No certificate data for https' for a ds with cert
@dneuman64 @trevorackerman Could you help to review if this is needed? Thanks
https://issues.apache.org/jira/browse/TC-43
Looks good to me.
I don't think we need a debug log entry but it's not a huge deal.
| gharchive/pull-request | 2016-11-15T11:06:35 | 2025-04-01T04:33:29.393796 | {
"authors": [
"trevorackerman",
"weifensh"
],
"repo": "apache/incubator-trafficcontrol",
"url": "https://github.com/apache/incubator-trafficcontrol/pull/81",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
811860997 | alter doc for check style
alter doc for check style
@qiaojialin @jixuan1989 @HTHou hi, could you please review this?
@qiaojialin @jixuan1989 @HTHou hi, could you please review this?
Nice catch!
| gharchive/pull-request | 2021-02-19T09:33:20 | 2025-04-01T04:33:29.395086 | {
"authors": [
"HTHou",
"wangchao316"
],
"repo": "apache/iotdb",
"url": "https://github.com/apache/iotdb/pull/2697",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1525336773 | JAMES-3756 JMAP endpoint (upload/download) should support being called with accountIds of delegated accounts
resolve https://github.com/linagora/james-project/issues/4679
Please rebase @vttranlina
rebased
green
| gharchive/pull-request | 2023-01-09T10:49:28 | 2025-04-01T04:33:29.397468 | {
"authors": [
"chibenwa",
"vttranlina"
],
"repo": "apache/james-project",
"url": "https://github.com/apache/james-project/pull/1382",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
841164374 | KAFKA-12556: Add --under-preferred-replica-partitions option to describe topics command
Whether the preferred replica is the partition leader directly affects the external output traffic of the broker. When the preferred replica of all partitions becomes the leader, the external output traffic of the broker will be in a balanced state. When there are a large number of partition leaders that are not preferred replicas, it will be destroyed this state of balance.
Currently, the controller will periodically check the unbalanced ratio of the partition preferred replicas (if enabled) to trigger the preferred replica election, or manually trigger the election through the kafka-leader-election tool. However, if we want to know which partition leader is in the non-preferred replica, we need to look it up in the controller log or judge ourselves from the topic details list.
We can add the --under-preferred-replica-partitions configuration option in TopicCommand describe topics to query the list of partitions in the current cluster that are in non-preferred replicas.
Committer Checklist (excluded from commit message)
[ ] Verify design and implementation
[ ] Verify test coverage and CI build status
[ ] Verify documentation (including upgrade notes)
@chia7712 Can you take a look at this pr?Thanks. :)
@wenbingshen Thanks for the PR. We need a KIP for adding a new flag to a command line tool as it is considered as part of the public API.
@wenbingshen Thanks for the PR. We need a KIP for adding a new flag to a command line tool as it is considered as part of the public API.
Thank you for your comment, I will create a KIP for this. :)
| gharchive/pull-request | 2021-03-25T17:43:07 | 2025-04-01T04:33:29.421520 | {
"authors": [
"dajac",
"wenbingshen"
],
"repo": "apache/kafka",
"url": "https://github.com/apache/kafka/pull/10403",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1497675098 | KAFKA-14492: Extract a method to create LogManager, in order to be overrided by subclass of KafkaServer
KAFKA-14492
In our scene, we want to implement a subclass of LogManager, add do some interception like encrypting/auditing.
Even more, we want enhance the write log performance in the future.
Split alone a create method can provide an extension point to allow user define the action from outside kafka core.
Base kafka, entend it, rather than modification.
hi, those failed checks are not caused by my modification in this time.
what can i do next ?
Thanks for the PR. It is important to note that KafkaServer is not considered as a public interface in Kafka. Hence, there is not guarantee that the interface will remain like this in the future. It can basically change anytime. Therefore, I am personally not comfortable with this change. Have you considered alternative approaches? @ijuma What do you think?
Yes, a KIP would be required to turn this into a public API. For non public APIs, the code can change at any time.
Thanks for your reply.
I agree that the KafkaServer is not designed for extend, attempt to replace the LogManager by override a createLogManager, maybe not a permanently method.
As your said, a KIP is needed to propose a complete approache that specially focus on which components can be extended in KafkaServer. Design an interface describe which components support extend is obviously a better way.
Before an api or interface be provided, allow to extend an object's method properly can be a temporary way with less drawback. At least is better than that fork a standalone repository, then modify with the code. The extend way is just add a kafka-core.jar, then override inside classes.
Except the extend side, provide a createLogManager method is coincide with createReplicaManager below those code. It's also a good style only by the readability.
Further on, only this modification is negligible, we could meet more problems in the way to extend LogManager. In that time, i am willing to provide a comprehensive plan as a KIP to resolve the extension problem.
Thank you both again.
| gharchive/pull-request | 2022-12-15T02:10:23 | 2025-04-01T04:33:29.426670 | {
"authors": [
"PhantomMaa",
"dajac",
"ijuma"
],
"repo": "apache/kafka",
"url": "https://github.com/apache/kafka/pull/12997",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2062313351 | KAFKA-14412: Decouple RocksDB access from CF
To support future use-cases that use different strategies for accessing RocksDB, we need to de-couple the RocksDB access strategy from the Column Family access strategy.
To do this, we now have two separate accessors:
DBAccessor: dictates how we access RocksDB. Currently only one strategy is supported: DirectDBAccessor, which access RocksDB directly, via the RocksDB class for all operations. In the future, a BatchedDBAccessor will be added, which enables transactions via WriteBatch.
ColumnFamilyAccessor: maps StateStore operations to operations on one or more column families. This is a rename of the old RocksDBDBAccessor.
@cadonna @mjsax @ableegoldman @lucasbru @wcarlson5 @bbejeck @vvcephei @guozhangwang
This is part of KIP-892, and has been broken out into a separate PR to reduce the review burden on the main KIP-892 implementation, since it can be merged independently.
There are no tests, because there are no behavioural changes, just a refactoring. The existing test suite should ensure no regressions.
@nicktelford Seems like all build jobs timed out. Could you take a look?
@lucasbru I think this is just the CI causing trouble again. The build passes locally. Is there a way to rerun the build without pushing more commits?
@nicktelford I'll rerun it. That being said, the last jobs on trunk have all finished within 3-5 hours, so this must be caused by either infrastructure or the code in this PR.
@lucasbru OK, my bad. It turns out I did a minor refactoring after I ran the test suite yesterday that was so insignificant I didn't think I needed to run the tests again... Turns out I was wrong :see_no_evil:
I've fixed the bug now and the tests pass locally for real now.
| gharchive/pull-request | 2024-01-02T11:52:50 | 2025-04-01T04:33:29.431227 | {
"authors": [
"lucasbru",
"nicktelford"
],
"repo": "apache/kafka",
"url": "https://github.com/apache/kafka/pull/15105",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
110295179 | KAFKA-2620: Introduce Scalariform
This is a WIP. There will likely need to be discussion around a rule set we want to use and if we actually want to include this in the build. This could also be used as a one time rebase.
This commit does not include the actual formatting changes. To see what effect the patch has run
gradle formatScala
@granthenke Would you like to continue working on this JIRA and PR?
I think we may want to use scalafmt instead as it supports a maxColumn setting:
http://scalameta.org/scalafmt/#FAQ/Troubleshooting
Closing as this has been inactive for a while.
This is a WIP. There will likely need to be discussion around a rule set we want to use and if we actually want to include this in the build. This could also be used as a one time rebase.
This commit does not include the actual formatting changes. To see what effect the patch has run
gradle formatScala
| gharchive/pull-request | 2015-10-07T19:06:30 | 2025-04-01T04:33:29.434572 | {
"authors": [
"b-goyal",
"granthenke",
"guozhangwang",
"ijuma"
],
"repo": "apache/kafka",
"url": "https://github.com/apache/kafka/pull/284",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
588005308 | Suppress detailed responses for security-sensitive (PCI-DSS) environments
Before this PR, Kafka Connect's exception mapper would print an unchecked messages from unhandled stack traces. I am contributing this code change so that the community has the option to change this behavior based on a configuration option, "error.rest.response.message.detail.enabled"
As this functionality is an API-level change, as it changes the response of handled errors, it should not be enabled by default. I have provided a unit test for handling the expected response if the option is enabled.
Issue: https://issues.apache.org/jira/browse/KAFKA-9766
| gharchive/pull-request | 2020-03-25T21:34:51 | 2025-04-01T04:33:29.436247 | {
"authors": [
"cwpenhale"
],
"repo": "apache/kafka",
"url": "https://github.com/apache/kafka/pull/8355",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
672717996 | KAFKA-10314: KafkaStorageException on reassignment when offline log d…
…irectories exist
Make sure that we set the isNew field in LeaderAndIsrRequest correctly for brokers
that gets added to the replica set on reassignment.
This is tested by creating a variant of ControllerIntergationTest.testPartitionReassignment()
that makes one of the log directories on the target broker offline before initiating the
reassignment. Without the change to the way isNew is set, this fails after a timeout. With
the change, it succeeds.
To facilitate calling causeLogDirFailure() both from ControllerIntegrationTest and
LogDirFailureTest, the method was moved to TestUtils along with the other helper
methods that deals with interacting with KafkaServer instances for test cases.
More detailed description of your change,
if necessary. The PR title and PR message become
the squashed commit message, so use a separate
comment to ping reviewers.
Summary of testing strategy (including rationale)
for the feature or bug fix. Unit and/or integration
tests are expected for any behaviour change and
system tests should be considered for larger changes.
Committer Checklist (excluded from commit message)
[ ] Verify design and implementation
[ ] Verify test coverage and CI build status
[ ] Verify documentation (including upgrade notes)
Ok to test
@nresare Good catch! The change looks good.
Before merging I'd like to get this change tested in the CI pipeline. The Apache Jenkins setup just changed a few days ago and it needs 241e1447fae8b85b3bb491f371357b8f92d2da72 to run. So can you rebase on trunk?
Trying to navigate the tests, it seems they are related to flakiness in other parts of the overall test suite. Is it possible to re-trigger tests?
| gharchive/pull-request | 2020-08-04T11:26:09 | 2025-04-01T04:33:29.440980 | {
"authors": [
"mimaison",
"nresare"
],
"repo": "apache/kafka",
"url": "https://github.com/apache/kafka/pull/9122",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1733886208 | Extract table from ResolvedIdentifier for DropTable in Spark 3.4
Why are the changes needed?
adapting changes in logical plan of DropTable in Spark 3.4 by extracting table object from ResolvedIdntifier, to fix test w/ Spark 3.4 ut "DropTable"
How was this patch tested?
[ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible
[ ] Add screenshots for manual tests if appropriate
[x] Run test locally before make a pull request
cc @zhouyifan279
Codecov Report
Merging #4910 (53c76f6) into master (8f61835) will not change coverage.
The diff coverage is 0.00%.
@@ Coverage Diff @@
## master #4910 +/- ##
======================================
Coverage 0.00% 0.00%
======================================
Files 557 557
Lines 30700 30703 +3
Branches 3995 3996 +1
======================================
- Misses 30700 30703 +3
Impacted Files
Coverage Δ
...ubi/plugin/spark/authz/serde/tableExtractors.scala
0.00% <0.00%> (ø)
:mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
Thanks, merged to master.
| gharchive/pull-request | 2023-05-31T10:26:31 | 2025-04-01T04:33:29.448507 | {
"authors": [
"bowenliang123",
"codecov-commenter"
],
"repo": "apache/kyuubi",
"url": "https://github.com/apache/kyuubi/pull/4910",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2083636373 | [KYUUBI #5964][BUG] Avoid check not fully optimized query for InsertIntoDataSourceDirCommand and InsertIntoDataSourceCommand
:mag: Description
Issue References 🔗
This pull request fixes #5964
Describe Your Solution 🔧
InsertIntoDataSourceDirCommand and InsertIntoDataSourceCommand‘s query is not fully optimized, we direct check it's query will cause request privilege that we haven't used.
We can directly ignore the query's check. Since we will check it's generated plan. Still will request the correct privilege of the SQL
Types of changes :bookmark:
[x] Bugfix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Test Plan 🧪
Behavior Without This Pull Request :coffin:
Behavior With This Pull Request :tada:
Related Unit Tests
Checklist 📝
[ ] This patch was not authored or co-authored using Generative Tooling
Be nice. Be informative.
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Comparison is base (3915fe8) 61.21% compared to head (1adcf8d) 61.06%.
Report is 1 commits behind head on master.
Additional details and impacted files
@@ Coverage Diff @@
## master #5983 +/- ##
============================================
- Coverage 61.21% 61.06% -0.15%
Complexity 23 23
============================================
Files 622 622
Lines 36897 37036 +139
Branches 5016 5023 +7
============================================
+ Hits 22585 22617 +32
- Misses 11877 11973 +96
- Partials 2435 2446 +11
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
ping @yaooqinn
| gharchive/pull-request | 2024-01-16T10:54:02 | 2025-04-01T04:33:29.456010 | {
"authors": [
"AngersZhuuuu",
"codecov-commenter"
],
"repo": "apache/kyuubi",
"url": "https://github.com/apache/kyuubi/pull/5983",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2286801315 | Improve closeEngine method names and logs due to closeSession
:mag: Description
Issue References 🔗
This pull request aims to improve closeEngine method names and logs due to closeSession.
when the share level is CONNECTION, closeSession triggers the closeEngine, there are some minor issues with logging and method name.
Describe Your Solution 🔧
Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.
Types of changes :bookmark:
[ ] Bugfix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Test Plan 🧪
Behavior Without This Pull Request :coffin:
Behavior With This Pull Request :tada:
Related Unit Tests
Checklist 📝
[x] This patch was not authored or co-authored using Generative Tooling
Be nice. Be informative.
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 58.44%. Comparing base (12c5568) to head (1b5a061).
Additional details and impacted files
@@ Coverage Diff @@
## master #6376 +/- ##
============================================
- Coverage 58.51% 58.44% -0.08%
Complexity 24 24
============================================
Files 653 653
Lines 39895 39895
Branches 5482 5482
============================================
- Hits 23345 23315 -30
- Misses 14055 14081 +26
- Partials 2495 2499 +4
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
@pan3793 Please take a look if you find a time.
thanks, merged to master.
| gharchive/pull-request | 2024-05-09T03:01:58 | 2025-04-01T04:33:29.464245 | {
"authors": [
"codecov-commenter",
"yikf"
],
"repo": "apache/kyuubi",
"url": "https://github.com/apache/kyuubi/pull/6376",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
473539609 | Fix hash validation with stream uploads
Fix hash validation with stream uploads
Description
Bug fix for cloudfiles upload_object_via_stream() erroneously throwing a hash mismatch when using a file handle as the iterator. Example:
from libcloud.storage.providers import get_driver
from libcloud.storage.types import ObjectDoesNotExistError
driver = get_driver('cloudfiles')
driver = driver('user', 'key')
container = driver.get_container(container_name='test')
with open('/tmp/test.file', 'w') as fh:
fh.write('test')
with open('/tmp/test.file') as fh:
driver.upload_object_via_stream(iterator=fh, container=container, object_name='test.file')
This would result in the exception:
libcloud.storage.types.ObjectHashMismatchError: <ObjectHashMismatchError in <libcloud.storage.drivers.cloudfiles.CloudFilesStorageDriver object at 0x103f43d68>, value=MD5 hash checksum does not match (expected=d41d8cd98f00b204e9800998ecf8427e, actual=098f6bcd4621d373cade4e832627b4f6), object = test.file>
d41d8cd98f00b204e9800998ecf8427e is the hash for an empty file. The problem is the upload happens before the hash is calculated and the file handle is read to the EOF. The fix is to seek the file handle to the beginning before feeding it into the hash function.
Status
done, ready for review
Checklist (tick everything that applies)
[x] Code linting (required, can be done after the PR checks)
[ ] Documentation
[ ] Tests
[ ] ICLA (required for bigger changes)
Thanks for the contribution.
Can you please also add a corresponding test case?
I've added a test case (858d9e6e16d0f1a2160537d9acac1ca0f897ef70) and will go ahead and merge this into trunk.
It's also worth pointing out that you made change in the base driver class so this change potentially affects all the drivers. I think the change is the correct one, but we should still keep an eye on potential issues once a new version is released.
Sorry, I've been busy lately, but did intend to add tests. Thanks for doing it for me, @Kami!
| gharchive/pull-request | 2019-07-26T21:17:49 | 2025-04-01T04:33:29.469222 | {
"authors": [
"Kami",
"gvengel"
],
"repo": "apache/libcloud",
"url": "https://github.com/apache/libcloud/pull/1326",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.