id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
293928891
DRILL-6130: Fix NPE during physical plan submission for various storage plugins Fixed ser / de issues for Hive, Kafka, Hbase plugins. Added physical plan submission unit test for all storage plugins in contrib module. Refactoring. @vdiravka please review. @vdiravka thanks for the code review. Addressed two comments.
gharchive/pull-request
2018-02-02T15:39:09
2025-04-01T04:33:28.244567
{ "authors": [ "arina-ielchiieva" ], "repo": "apache/drill", "url": "https://github.com/apache/drill/pull/1108", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
744391154
Doc Fix : modify druid.historical.cache.maxEntrySize property in Unified format Description The property druid.historical.cache.maxEntrySize mentioned in Configuration reference -> Historical -> Historical Caching is inconsistent with the existing format, lack of Possible Values. This PR fills in this column as positive integer This PR has: [x] been self-reviewed. Key changed/added classes in this PR docs/configuration/index.md @a2l007 @nishantmonu51 Thanks for your review and merge!
gharchive/pull-request
2020-11-17T04:04:09
2025-04-01T04:33:28.248375
{ "authors": [ "zhangyue19921010" ], "repo": "apache/druid", "url": "https://github.com/apache/druid/pull/10590", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2555617371
[Backport] Fix maxRowsInMemory default for streaming (#17028) Backport apache/druid#17028 Thank you for the backport, @kfaraz
gharchive/pull-request
2024-09-30T05:42:45
2025-04-01T04:33:28.249268
{ "authors": [ "AmatyaAvadhanula", "kfaraz" ], "repo": "apache/druid", "url": "https://github.com/apache/druid/pull/17186", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
733657298
DubboHandlerRegistry#removeService NPE问题 [x] I have searched the issues of this repository and believe that this is not a duplicate. [x] I have checked the FAQ of this repository and believe that this is not a duplicate. Environment Dubbo version: 2.7.7 Operating System version: macOs Mojave 10.14.5 Java version: 1.8 Steps to reproduce this issue 使用dubbo启动一个任意grpc Provider进程 关闭进程 Expected Result 进程正常关闭,资源回收,无任何异常 Actual Result 有NullPointerException报出 java.lang.NullPointerException: null at org.apache.dubbo.rpc.protocol.grpc.DubboHandlerRegistry.removeService(DubboHandlerRegistry.java:66) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.rpc.protocol.grpc.GrpcProtocol.lambda$doExport$1(GrpcProtocol.java:97) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.rpc.protocol.AbstractProxyProtocol$1.unexport(AbstractProxyProtocol.java:95) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.rpc.listener.ListenerExporterWrapper.unexport(ListenerExporterWrapper.java:71) [dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.config.ServiceConfig.unexport(ServiceConfig.java:170) [dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.config.bootstrap.DubboBootstrap.lambda$unexportServices$17(DubboBootstrap.java:952) [dubbo-2.7.7.jar:2.7.7] at java.util.ArrayList.forEach(ArrayList.java:1257) ~[na:1.8.0_231] at org.apache.dubbo.config.bootstrap.DubboBootstrap.unexportServices(DubboBootstrap.java:950) [dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.config.bootstrap.DubboBootstrap.destroy(DubboBootstrap.java:1072) [dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.config.bootstrap.DubboBootstrap$1.callback(DubboBootstrap.java:193) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.common.function.ThrowableAction.execute(ThrowableAction.java:46) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.common.lang.ShutdownHookCallbacks.lambda$callback$0(ShutdownHookCallbacks.java:70) ~[dubbo-2.7.7.jar:2.7.7] at java.lang.Iterable.forEach(Iterable.java:75) ~[na:1.8.0_231] at org.apache.dubbo.common.lang.ShutdownHookCallbacks.callback(ShutdownHookCallbacks.java:70) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.config.DubboShutdownHook.callback(DubboShutdownHook.java:85) ~[dubbo-2.7.7.jar:2.7.7] at org.apache.dubbo.config.DubboShutdownHook.run(DubboShutdownHook.java:73) ~[dubbo-2.7.7.jar:2.7.7] 如下两个地方都会去调用 DubboHandlerRegistry#removeService,但这个方法貌似没有重复调用的情况,重复调用会NPE DubboShutdownHook#destroyProtocols DubboBootstrap#unexportServices void removeService(String serviceKey) { ServerServiceDefinition service = services.remove(serviceKey); // 此处service可能为空 for (ServerMethodDefinition<?, ?> method : service.getMethods()) { methods.remove(method.getMethodDescriptor().getFullMethodName(), method); } } Can you provide detailed steps to reproduce? I tested it here and it didn’t reproduce the NPE.
gharchive/issue
2020-10-31T07:21:04
2025-04-01T04:33:28.260090
{ "authors": [ "xianliguang", "xiaoheng1" ], "repo": "apache/dubbo", "url": "https://github.com/apache/dubbo/issues/6856", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
565841475
some code optimization What is the purpose of the change some code optimization Brief changelog XXXXX Verifying this change XXXXX Follow this checklist to help us incorporate your contribution quickly and easily: [x] Make sure there is a GITHUB_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GITHUB issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue. [ ] Format the pull request title like [Dubbo-XXX] Fix UnknownException when host config not exist #XXX. Each commit in the pull request should have a meaningful subject line and body. [ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why. [ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add sample in dubbo samples project. [ ] Run mvn clean install -DskipTests=false & mvn clean test-compile failsafe:integration-test to make sure unit-test and integration-test pass. [ ] If this contribution is large, please follow the Software Donation Guide. Codecov Report Merging #5746 into master will decrease coverage by 0.02%. The diff coverage is 33.33%. @@ Coverage Diff @@ ## master #5746 +/- ## ============================================ - Coverage 61.24% 61.22% -0.03% - Complexity 422 495 +73 ============================================ Files 922 928 +6 Lines 37498 37899 +401 Branches 5431 5447 +16 ============================================ + Hits 22967 23204 +237 - Misses 12055 12176 +121 - Partials 2476 2519 +43 Impacted Files Coverage Δ Complexity Δ ...g/apache/dubbo/common/bytecode/ClassGenerator.java 61.53% <0%> (ø) 0 <0> (ø) :arrow_down: ...a/org/apache/dubbo/monitor/dubbo/DubboMonitor.java 88.57% <100%> (ø) 16 <0> (ø) :arrow_down: ...g/context/properties/DefaultDubboConfigBinder.java 0% <0%> (-100%) 0% <0%> (ø) ...t/annotation/DubboLifecycleComponentRegistrar.java 0% <0%> (-100%) 0% <0%> (ø) ...ache/dubbo/rpc/cluster/support/ForkingCluster.java 0% <0%> (-100%) 0% <0%> (ø) .../context/properties/AbstractDubboConfigBinder.java 0% <0%> (-92.31%) 0% <0%> (ø) ...ring/context/config/DubboConfigBeanCustomizer.java 0% <0%> (-66.67%) 0% <0%> (ø) ...che/dubbo/rpc/cluster/support/FailfastCluster.java 50% <0%> (-50%) 0% <0%> (ø) ...va/org/apache/dubbo/remoting/TimeoutException.java 0% <0%> (-22.23%) 0% <0%> (ø) ...dubbo/remoting/exchange/support/DefaultFuture.java 71.96% <0%> (-14.53%) 0% <0%> (ø) ... and 120 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update b288ad7...b110fa4. Read the comment docs. thanks for your contribution :)
gharchive/pull-request
2020-02-16T02:49:02
2025-04-01T04:33:28.279528
{ "authors": [ "ZZMarquis", "codecov-io", "htynkn" ], "repo": "apache/dubbo", "url": "https://github.com/apache/dubbo/pull/5746", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
925251311
Contribute to options documentation Is there a way for me to contribute to the options documentation? For example, in the screenshot below, you can see that the "label.position" for bar charts has the proper "uiControl" setting from the option.series-bar.js file. However, the "uiControl" is missing for pie charts and I'd like to add it. There are a lot of errors and missing items that I come across and would love to contribute. Sure, this is the repo for Apache ECharts option doc. You can fork it and then make PRs.
gharchive/issue
2021-06-18T22:57:27
2025-04-01T04:33:28.281727
{ "authors": [ "ataft", "susiwen8" ], "repo": "apache/echarts-doc", "url": "https://github.com/apache/echarts-doc/issues/188", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
796063781
5.0版本resize是否取消掉了缓动动画,为什么执行resize()后缓动效果没有了? `let myChart = this.$Echarts.init(document.getElementById('myDiv')); let option = { xAxis: { type: 'category', data: ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun'] }, yAxis: { type: 'value' }, series: [{ data: [120, 200, 150, 80, 70, 110, 130], type: 'bar' }], animation: true }; myChart.setOption(option); myChart.resize();` 这段代码在4.x下,即使执行了resize,缓动效果还是存在的,可是升级了5.0后,执行了resize之后缓动效果就没有了 resize动画效果确实没了,请问答主加上了吗
gharchive/issue
2021-01-28T14:36:24
2025-04-01T04:33:28.284152
{ "authors": [ "JManToGithub", "zhoufanglu" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/14146", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
833525888
about bmap, Incomplete rendering load(关于echart中地图,最近开始加载不全的问题) Version 5.0.2 Reproduction link https://echarts.apache.org/examples/zh/editor.html?c=lines-bmap-effect Steps to reproduce No code modifiednothing What is expected? complete rendering load What is actually happening? No code modified Is it a network problem? 这是百度地图 2.0 自定义地图用到的瓦片地图服务器的问题,可能会遇到白块的地方,ECharts 无法解决此问题。建议尝试使用百度地图 3.0。 这是百度地图 2.0 自定义地图用到的瓦片地图服务器的问题,可能会遇到白块的地方,ECharts 无法解决此问题。建议尝试使用百度地图 3.0。 我用的是 "echarts": "^5.0.0",按理来说应该是3.0了 ECharts 5.0 与百度地图 3.0 没有关系,百度地图 API 脚本是需要开发者自行引入的,所以使用的版本由开发者来决定。 ECharts 5.0 与百度地图 3.0 没有关系,百度地图 API 脚本是需要开发者自行引入的,所以使用的版本由开发者来决定。 因为我写的是vue3.0的框架,所以引用了echarts官方最新的依赖。看了下node_moudles源码包里面bmap的拓展,确实引用3.0版本了 不太清楚你是否有理解,ECharts 并不会帮我们自动引入百度地图 API script,而是需要开发者在代码里手动引入。 <script type="text/javascript" src="https://api.map.baidu.com/api?v=3.0&ak=您的密钥"></script> 不太清楚你是否有理解,ECharts 并不会帮我们自动引入百度地图 API script,而是需要开发者在代码里手动引入。 <script type="text/javascript" src="https://api.map.baidu.com/api?v=3.0&ak=您的密钥"></script> 理解的,看上图 所以 3.0 仍然出现瓦片地图加载不全的问题吗? 所以 3.0 仍然出现瓦片地图加载不全的问题吗? 是的,2.0,3.0,都会出来。 这种情况只能建议向百度地图提工单咨询了。 这种情况只能建议向百度地图提工单咨询了。 感谢回答
gharchive/issue
2021-03-17T08:28:06
2025-04-01T04:33:28.291014
{ "authors": [ "joel785454405", "plainheart" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/14496", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
916823335
Export image of the charts using CORS resources Background In most cases, one chart may be referring to some external resources, some of them allow requests from any origin with the response header like Access-Control-Allow-Origin: '*' but some of them don't. For those resources with CORS header, we hope they can be exported. See this simple example, option = { xAxis: { type: 'category', data: ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun'] }, yAxis: { type: 'value' }, toolbox: { feature: { saveAsImage: {} } }, series: [{ data: [150, 230, 224, 218, 135, 147, 260], type: 'line', symbolSize: [95, 22], symbol: 'image://https://cdn.jsdelivr.net/gh/apache/echarts-website@asf-site/zh/images/logo.png' }] }; If we hope to get a screenshot through the current API getDataURL, we would get an unexpected error like this, Uncaught DOMException: Failed to execute 'toDataURL' on 'HTMLCanvasElement': Tainted canvases may not be exported. To avoid this, we will have to create some Image objects and specify their own crossOrigin property. Like this, const logoImg = new Image(); logoImg.src = 'image://https://cdn.jsdelivr.net/gh/apache/echarts-website@asf-site/zh/images/logo.png'; logoImg.crossOrigin = ''; // or 'anonymous' logoImg.onload = () => { // xxx }; But here is another limit, our series.symbol doesn't support the Image object yet! So we can't directly use the created image objects with the crossOrigin property. But we can try another approach: convert it to a Base64 URL. The code will be, const symbolSize = [95, 22]; const logoImg = new Image(); logoImg.src = 'https://cdn.jsdelivr.net/gh/apache/echarts-website@asf-site/zh/images/logo.png'; logoImg.crossOrigin = ''; // or 'anonymous' logoImg.onload = () => { const tmpCanvas = echarts.zrUtil.createCanvas() const tmpCtx = tmpCanvas.getContext('2d') const dpr = myChart.getDevicePixelRatio() tmpCanvas.width = symbolSize[0] * dpr tmpCanvas.height = symbolSize[1] * dpr tmpCanvas.style.width = symbolSize[0] + 'px' tmpCanvas.style.height = symbolSize[1] + 'px' tmpCtx.drawImage(logoImg, 0, 0, tmpCanvas.width, tmpCanvas.height) myChart.setOption({ series: [{ symbol: 'image://' + tmpCanvas.toDataURL() }] }) } option = { xAxis: { type: 'category', data: ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun'] }, yAxis: { type: 'value' }, toolbox: { feature: { saveAsImage: {} } }, series: [{ data: [150, 230, 224, 218, 135, 147, 260], type: 'line', symbolSize: symbolSize, symbol: 'none' }] }; Now let's export again, we would found it can be exported and downloaded successfully. But everything was going to be complex when we hope to set a different image for each symbol item. It would be simpler and easier if there could be an option to specify whether the image URL is cross-origin. This feature will involve some required changes in zrender, but that's underlying. Let's leave it alone now and just discuss the things related to ECharts. Proposed Usage For symbol with image://, I prefer specifying the crossOrigin in the URL, either image://crossOrigin/URL or image://URL?crossOrigin. For the common pattern object, { image: imageDom, // Image, HTMLImageElement, and HTMLCanvasElement are supported, while string path is not supported repeat: 'repeat' // whether to repeat texture, whose value can be repeat-x, repeat-y, or no-repeat } The crossOrigin can be set in imageDom in advance. So it should be OK. We can also provide a new option crossOrigin if needed. Demo https://www.makeapie.com/editor.html?c=xntd1QC3kI&v=1 Yeah, it's a very useful feature. ECharts GL also has a similar issue. I hardcoded cross-origin to anonymous(or it will not even display because of the strict origin policy in WebGL). But how to make it configurable still bothers me. Specifying it in the URL may be a good idea. But image://URL?crossOrigin may be conflicts with the exists ? in the URL itself. Another way is to specify this image policy at the beginning of initializing the echarts instead of adding it on every image. const chart = echarts.init(dom, null, { crossOrigin: 'anonymous' }) I'm not sure if using different CORS policies in the same echarts instance is a common sceneraio. Images are not very often used in echarts. Perhaps we can support the global CORS policy currently and see if any developers need more precisely control on each image. In either way, I think it will be convenient for developers if they only need to specify crossOrigin policy once. Any status on this ? i tried the method suggested by @pissang but i am getting type error for the same here is my code snippet : const chart = echarts.init(chartRef.current, "100ms-echarts", { crossOrigin: "anonymous", }) Argument of type '{ crossOrigin: string; }' is not assignable to parameter of type 'EChartsInitOpts'. Object literal may only specify known properties, and 'crossOrigin' does not exist in type 'EChartsInitOpts'. @hdz-666 This feature is under discussion, it's not implemented yet. Any info about, when this feature will be available?
gharchive/issue
2021-06-10T02:27:01
2025-04-01T04:33:28.300674
{ "authors": [ "AndreyPatseiko", "hdz-666", "pissang", "plainheart" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/15117", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1163874216
[Bug] 数据量少的时候起点和终点会有很多留白 Version 5.2.1 Link to Minimal Reproduction No response Steps to Reproduce 数据量很少的时候就会这样 Current Behavior 数据量少的时候起点和终点会有很多留白 Expected Behavior 希望能像highchart一样处理,在起点向前移,终点往后移 Environment - OS: - Browser: - Framework: Any additional comments? No response 设置 xAxis.boundaryGap 为 false. 设置 xAxis.boundaryGap 为 false. 感谢
gharchive/issue
2022-03-09T12:25:15
2025-04-01T04:33:28.305939
{ "authors": [ "aninchiron", "plainheart" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/16658", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2457285553
[Bug] textStyle is invalid when formatter and richText renderMode are being used in the tooltip Version 5.5.0 Link to Minimal Reproduction PYBwLglsB2AEC8sDeAoWsA2BTA5l6AJgFzJrqxigkBEARsGJQLbUA0Z6BAhmFyQNrVA8IaBEFWoBdMgF926SsAyQQJVOQoAnCDjzqA8tBrQYWNh1hcMAdy4BPAM4BlABbBLAYRhh8YEmHUBXLFlyEGA7CEgYEgAzf2gAY0joAApwAEpSNXR1LDB_dTh Steps to Reproduce use richText renderMode use formatter add textStyle tooltip: { renderMode: "richText", formatter: Data: {c}, textStyle: { fontSize: 60, fontWeight: "500", fontFamily: "My-Custom-Font", }, }, Current Behavior Expected Behavior Environment - OS: macOS Ventura - Browser: - Framework: React / React native Any additional comments? I have used the same chart option for tooltip in react native framework also, but the result is same as web (as mentioned above). Do you really need to use renderMode:'richText' ? formatter and textStyle work fine together without it. @helgasoft 1 - Yes i need to to use 'richText' for react native frame work, because html option is not working in mobile. If you any suggestion on this please do share, i will definitely give a try. 2 - 'formatter and textStyle work together fine without it' - Yes it is but as mentioned above fontSize, fontWeight and fontFamily are not implementing on the tool tip text while using formatter and textStyle together. One more request if you have any suggestions regarding the above points, please do share it would be highly appreciated. https://github.com/apache/echarts/issues/19966#issuecomment-2538259499 @souravpany
gharchive/issue
2024-08-09T07:23:02
2025-04-01T04:33:28.312342
{ "authors": [ "XHL-study", "helgasoft", "souravpany" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/20240", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2577718533
[Feature] 自行车 What problem does this feature solve? 暗室逢灯 What does the proposed API look like? 阿斯顿发到付十分大方撒旦法所发生的发撒说法大萨达发达 S这是打发斯蒂芬
gharchive/issue
2024-10-10T06:04:50
2025-04-01T04:33:28.313598
{ "authors": [ "purpletastes" ], "repo": "apache/echarts", "url": "https://github.com/apache/echarts/issues/20401", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2496839338
[FLINK-36188] Fix disable buffer flush lose efficacy The user lookup joins the hbase table, adds 1 to the col value, and writes it back to hbase @Test void testTableSinkDisabledBufferFlush() throws Exception { StreamExecutionEnvironment execEnv = StreamExecutionEnvironment.getExecutionEnvironment(); StreamTableEnvironment tEnv = StreamTableEnvironment.create(execEnv, streamSettings); tEnv.executeSql( "CREATE TABLE hTableForSink (" + " rowkey INT PRIMARY KEY NOT ENFORCED," + " family1 ROW<col1 INT>" + ") WITH (" + " 'connector' = 'hbase-2.2'," + " 'sink.buffer-flush.max-size' = '0'," + " 'sink.buffer-flush.max-rows' = '0'," + " 'table-name' = '" + TEST_TABLE_6 + "'," + " 'zookeeper.quorum' = '" + getZookeeperQuorum() + "'" + ")"); String insert = "INSERT INTO hTableForSink VALUES(1, ROW(1))"; tEnv.executeSql(insert).await(); tEnv.executeSql( "CREATE VIEW user_click AS " + " SELECT user_id, proctime() AS proc_time" + " FROM ( " + " VALUES(1), (1), (1), (1), (1)" + " ) AS t (user_id);"); tEnv.executeSql( "INSERT INTO hTableForSink SELECT " + " user_id as rowkey," + " ROW(CAST(family1.col1 + 1 AS INT))" + " FROM user_click INNER JOIN hTableForSink" + " FOR SYSTEM_TIME AS OF user_click.proc_time" + " ON hTableForSink.rowkey = user_click.user_id;"); tEnv.executeSql( "CREATE TABLE hTableForQuery (" + " rowkey INT PRIMARY KEY NOT ENFORCED," + " family1 ROW<col1 INT>" + ") WITH (" + " 'connector' = 'hbase-2.2'," + " 'table-name' = '" + TEST_TABLE_6 + "'," + " 'zookeeper.quorum' = '" + getZookeeperQuorum() + "'" + ")"); String query = "SELECT rowkey, family1.col1 FROM hTableForQuery"; TableResult firstResult = tEnv.executeSql(query); List<Row> firstResults = CollectionUtil.iteratorToList(firstResult.collect()); String firstExpected = "+I[1, 6]"; TestBaseUtils.compareResultAsText(firstResults, firstExpected); } test failed org.junit.ComparisonFailure: Different elements in arrays: expected 1 elements and received 1 expected: [+I[1, 6]] received: [+I[1, 2]] expected:<+I[1, [6]]> but was:<+I[1, [2]]> Expected :+I[1, 6] Actual :+I[1, 2] Thanks for submitting this fix! Pls. check the test the change broke. And since new changes will be necessary anyways, pls. reword the commit msg as well to contain the jira key to make "Title Validator" happy. Thanks for submitting this fix! Pls. check the test the change broke. And since new changes will be necessary anyways, pls. reword the commit msg as well to contain the jira key to make "Title Validator" happy. done,Could you please review it again? thank you @ferenc-csaky @MartijnVisser can you help to trigger a CI run, thanks~
gharchive/pull-request
2024-08-30T10:25:03
2025-04-01T04:33:28.317583
{ "authors": [ "MOBIN-F", "ferenc-csaky" ], "repo": "apache/flink-connector-hbase", "url": "https://github.com/apache/flink-connector-hbase/pull/49", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1962022279
[FLINK-31859] Update CycloneDX to 2.7.9 What is the purpose of the change The issue with current cylcone version is that it generates lots of exceptions in logs like org.codehaus.plexus.util.xml.pull.XmlPullParserException: UTF-8 BOM plus xml decl of ISO-8859-1 is incompatible (position: START_DOCUMENT seen <?xml version="1.0" encoding="ISO-8859-1"... @1:42) at org.codehaus.plexus.util.xml.pull.MXParser.parseXmlDeclWithVersion (MXParser.java:3423) at org.codehaus.plexus.util.xml.pull.MXParser.parseXmlDecl (MXParser.java:3345) at org.codehaus.plexus.util.xml.pull.MXParser.parsePI (MXParser.java:3197) at org.codehaus.plexus.util.xml.pull.MXParser.parseProlog (MXParser.java:1828) ... at org.apache.maven.model.io.xpp3.MavenXpp3Reader.read (MavenXpp3Reader.java:627) at org.cyclonedx.maven.BaseCycloneDxMojo.readPom (BaseCycloneDxMojo.java:759) at org.cyclonedx.maven.BaseCycloneDxMojo.readPom (BaseCycloneDxMojo.java:746) at org.cyclonedx.maven.BaseCycloneDxMojo.retrieveParentProject (BaseCycloneDxMojo.java:694) ... The PR is aiming to fix this Brief change log pom.xml Verifying this change This change is a trivial rework / code cleanup without any test coverage. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): (yes ) The public API, i.e., is any changes to the CustomResourceDescriptors: ( no) Core observer or reconciler logic that is regularly executed: ( no) Documentation Does this pull request introduce a new feature? ( no) If yes, how is the feature documented? (not applicable ) Please add the contents of this PR to https://github.com/apache/flink-kubernetes-operator/pull/691 to avoid fragmentation @gyfora this is not related to jdk17 this is a separate issue which i just noticed while comiling locally ok, I am just not a big fan of these tiny PRs/commits. It could have been simply part of the other PR as you noticed and fixed it while you worked on it. Opening one-liner PRs puts a strain on CI and also takes up everyone's time. Simply not worth it. ok, i see, i will put them together closing since now it is a part of https://github.com/apache/flink-kubernetes-operator/pull/691
gharchive/pull-request
2023-10-25T18:47:18
2025-04-01T04:33:28.323233
{ "authors": [ "gyfora", "snuyanzin" ], "repo": "apache/flink-kubernetes-operator", "url": "https://github.com/apache/flink-kubernetes-operator/pull/693", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
725124878
[FLINK-19624][table-planner-blink] Update deadlock break-up algorithm to cover more cases What is the purpose of the change Current deadlock breakup algorithm in Blink planner fails to cover some cases. See FLINK-19624 and the appendix of the design doc for more information. We're going to introduce a new deadlock breakup algorithm to solve this problem. This algorithm is also used in the upcoming multiple input node creation algorithm. Brief change log Introduce a new deadlock breakup algorithm. Verifying this change This change added tests and can be verified as follows: Run the added test cases. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): no The public API, i.e., is any changed class annotated with @Public(Evolving): no The serializers: no The runtime per-record code paths (performance sensitive): no Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no The S3 file system connector: no Documentation Does this pull request introduce a new feature? no If yes, how is the feature documented? not applicable @flinkbot run azure
gharchive/pull-request
2020-10-20T02:13:22
2025-04-01T04:33:28.329407
{ "authors": [ "TsReaper", "godfreyhe" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/13692", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
850587152
[hotfix][docs] Fix max parallelism behavior for Reactive Mode With FLINK-21844, Reactive Mode now has the same behavior when the same max parallelism as the "default" mode. This also includes minor typos/ fixes. Signed-off-by: austin ce austin.cawley@gmail.com What is the purpose of the change With FLINK-21844, Reactive Mode now has the same behavior when the same max parallelism as the "default" mode. This also includes minor typos/ fixes. Brief change log Update Elastic Scaling page with description of new behavior Verifying this change This change is a trivial rework / code cleanup without any test coverage. Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): no The public API, i.e., is any changed class annotated with @Public(Evolving): no The serializers: no The runtime per-record code paths (performance sensitive): no Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no The S3 file system connector: no Documentation Does this pull request introduce a new feature? no If yes, how is the feature documented? docs @rmetzger can you have a look at this when you have a chance? @flinkbot attention rmetzger (?) Thanks for merging Chesnay & Thanks for fixing Austin. I think I ignored these notifications because I thought it is the PR fixing the test instability in ReactiveModeITCase. I'll pay more attention in the future! No worries! Should've updated these with the original ticket
gharchive/pull-request
2021-04-05T19:03:18
2025-04-01T04:33:28.334610
{ "authors": [ "austince", "rmetzger" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/15488", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
338831941
[FLINK-9755][network] forward exceptions in RemoteInputChannel#notifyBufferAvailable() to the responsible thread What is the purpose of the change Exceptions in RemoteInputChannel#notifyBufferAvailable(), e.g. state checks, were swallowed inside LocalBufferPool#recycle() and neither logged nor otherwise processed and may have lead to stalling processes waiting for a notification that never comes. Please note that this PR builds upon #6271 which also touched the unit tests which we change here. @tillrohrmann, @zhijiangW can you have a look at this cleanup of LocalBufferPool#recycle()? Brief change log cleanup of LocalBufferPool#recycle() also clarifying the contract of BufferListener#notifyBufferAvailable() which should recycle the given buffer (one implementation already did that; RemoteInputChannel did not) forward exceptions in RemoteInputChannel#notifyBufferAvailable() to the responsible channel and recycle the given buffer in that case Verifying this change This change added tests and can be verified as follows: added RemoteInputChannelTest#testFailureInNotifyBufferAvailable() Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): no The public API, i.e., is any changed class annotated with @Public(Evolving): no The serializers: no The runtime per-record code paths (performance sensitive): no Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: no The S3 file system connector: no Documentation Does this pull request introduce a new feature? no If yes, how is the feature documented? JavaDocs Thanks for fixing this potential bug. It makes sense to handle exception during notifyBufferAvailable on listener side. Just some thoughts on my side above. :) thanks for the review - I changed the code as requested actually, I forgot to adapt the unit test which relied on the previous order of the calls and assumed that the buffer got recycled. With the changes from the fixup, the buffer is not recycled by the failure directly but rather during error handling which will eventually release the channel's resources. -> this is now fixed as well rebased to solve the merge conflict (auto-solved by git though) Thanks for the review, merging...
gharchive/pull-request
2018-07-06T07:19:44
2025-04-01T04:33:28.341987
{ "authors": [ "NicoK", "zhijiangW" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/6272", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
464849390
[FLINK-13107][table-planner-blink] Copy TableApi IT and UT to Blink planner What is the purpose of the change The issue aims to copy the testcases in the following packages from flink-planner and original blink to Blink-planner: org.apache.flink.table.api.batch.table org.apache.flink.table.api.stream.table org.apache.flink.table.runtime.batch.table org.apache.flink.table.runtime.stream.table Brief change log Copy UT and ITCase of TableApi Fix some Bugs, commit message contains detail information about those Bugs Verifying this change UT, ITCase Does this pull request potentially affect one of the following parts: Dependencies (does it add or upgrade a dependency): (no) The public API, i.e., is any changed class annotated with @Public(Evolving): (no) The serializers: (no) The runtime per-record code paths (performance sensitive): (no) Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) The S3 file system connector: (no) Documentation Does this pull request introduce a new feature? (no) If yes, how is the feature documented? (not applicable) We need copy all org.apache.flink.table.api.*? I think we have these duplicate plan cases in org.apache.flink.table.table.*? Why not put cases to org.apache.flink.table.table.* to be consistent with previous SQL plan tests? What are the implications on the test times when we copy so many tests? According to this: https://travis-ci.org/beyond1920/flink/builds/557233740, the test time increases less than 2 mins. travis passed here: https://travis-ci.org/beyond1920/flink/builds/557233740 I'm merging this
gharchive/pull-request
2019-07-06T12:23:58
2025-04-01T04:33:28.348759
{ "authors": [ "JingsongLi", "KurtYoung", "beyond1920" ], "repo": "apache/flink", "url": "https://github.com/apache/flink/pull/9006", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
383288365
GEODE-6079: Attempt to sync filesystem before searching for suspect s… …trings. Thank you for submitting a contribution to Apache Geode. In order to streamline the review of the contribution we ask you to ensure the following steps have been taken: For all changes: [ ] Is there a JIRA ticket associated with this PR? Is it referenced in the commit message? [ ] Has your PR been rebased against the latest commit within the target branch (typically develop)? [ ] Is your initial contribution a single, squashed commit? [ ] Does gradlew build run cleanly? [ ] Have you written or updated unit tests to verify your changes? [ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0? Note: Please ensure that once the PR is submitted, you check travis-ci for build issues and submit an update to your PR as soon as possible. If you need help, please send an email to dev@geode.apache.org. Because we run tests in parallel, and because they all seem to be writing to the same suspect file, it may well be that either the DUnit run will timeout with the added time of the sync(), and/or the call will throw an IOException after timing out with frequency that makes this change pointless. The precheckin will be interesting to review.
gharchive/pull-request
2018-11-21T20:52:46
2025-04-01T04:33:28.353427
{ "authors": [ "PurelyApplied" ], "repo": "apache/geode", "url": "https://github.com/apache/geode/pull/2892", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
388536752
GEODE-6164: CacheClientProxy's closeSocket should be called atomically Thank you for submitting a contribution to Apache Geode. @jhuynh1 @Bill In order to streamline the review of the contribution we ask you to ensure the following steps have been taken: For all changes: [ ] Is there a JIRA ticket associated with this PR? Is it referenced in the commit message? [ ] Has your PR been rebased against the latest commit within the target branch (typically develop)? [ ] Is your initial contribution a single, squashed commit? [ ] Does gradlew build run cleanly? [ ] Have you written or updated unit tests to verify your changes? [ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0? Note: Please ensure that once the PR is submitted, you check travis-ci for build issues and submit an update to your PR as soon as possible. If you need help, please send an email to dev@geode.apache.org. Looks good to me now.
gharchive/pull-request
2018-12-07T07:28:47
2025-04-01T04:33:28.357998
{ "authors": [ "Bill", "gesterzhou" ], "repo": "apache/geode", "url": "https://github.com/apache/geode/pull/2972", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
916700809
[GOBBLIN-1466] Make meters shared between DagManagerThreads Dear Gobblin maintainers, Please accept this PR. I understand that it will not be reviewed until I have checked off all the steps below! JIRA [x] My PR addresses the following Gobblin JIRA issues and references them in the PR title. For example, "[GOBBLIN-XXX] My Gobblin PR" https://issues.apache.org/jira/browse/GOBBLIN-1466 Description [x] Here are some details about my PR, including screenshots (if applicable): Each of the DagManagerThreads had its own meter (but with the same names) previously, which causing confusing results when reading them, instead there should only be meters shared across the threads. For service level meter, created one meter in the DagManager and passed it to each thread instead For group level meters, created a static map keyed by group name to be shared by all threads Tests [x] My PR adds the following unit tests OR does not need testing for this extremely good reason: Tested locally by printing meter values in each thread Commits [x] My commits all reference JIRA issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message": Subject is separated from body by a blank line Subject is limited to 50 characters Subject does not end with a period Subject uses the imperative mood ("add", not "adding") Body wraps at 72 characters Body explains "what" and "why", not "how" Codecov Report Merging #3306 (048841f) into master (bf37c76) will decrease coverage by 3.51%. The diff coverage is 75.00%. @@ Coverage Diff @@ ## master #3306 +/- ## ============================================ - Coverage 46.50% 42.98% -3.52% + Complexity 10041 9311 -730 ============================================ Files 2041 2041 Lines 79349 79356 +7 Branches 8845 8846 +1 ============================================ - Hits 36898 34112 -2786 - Misses 39020 42007 +2987 + Partials 3431 3237 -194 Impacted Files Coverage Δ ...blin/service/modules/orchestration/DagManager.java 72.64% <75.00%> (-0.34%) :arrow_down: .../org/apache/gobblin/util/filters/HiddenFilter.java 0.00% <0.00%> (-100.00%) :arrow_down: ...g/apache/gobblin/cluster/HelixMessageSubTypes.java 0.00% <0.00%> (-100.00%) :arrow_down: ...gobblin/runtime/mapreduce/GobblinOutputFormat.java 0.00% <0.00%> (-100.00%) :arrow_down: ...obblin/compaction/source/CompactionFailedTask.java 0.00% <0.00%> (-100.00%) :arrow_down: ...n/cluster/event/ClusterManagerShutdownRequest.java 0.00% <0.00%> (-100.00%) :arrow_down: ...in/compaction/action/CompactionCompleteAction.java 0.00% <0.00%> (-100.00%) :arrow_down: ...n/compaction/mapreduce/orc/OrcKeyDedupReducer.java 0.00% <0.00%> (-100.00%) :arrow_down: ...action/audit/KafkaAuditCountHttpClientFactory.java 0.00% <0.00%> (-100.00%) :arrow_down: ...askStateCollectorServiceHiveRegHandlerFactory.java 0.00% <0.00%> (-100.00%) :arrow_down: ... and 138 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update bf37c76...048841f. Read the comment docs. rebasing to master to fix travis +1.
gharchive/pull-request
2021-06-09T21:52:00
2025-04-01T04:33:28.380975
{ "authors": [ "codecov-commenter", "jack-moseley", "sv2000" ], "repo": "apache/gobblin", "url": "https://github.com/apache/gobblin/pull/3306", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
964405608
[GOBBLIN-1509] Announce flow failure on DagManager::addDag error Dear Gobblin maintainers, Please accept this PR. I understand that it will not be reviewed until I have checked off all the steps below! JIRA [ ] My PR addresses the following Gobblin JIRA issues and references them in the PR title. For example, "[GOBBLIN-XXX] My Gobblin PR" [ ] https://issues.apache.org/jira/browse/GOBBLIN-1509 Description [ ] Here are some details about my PR, including screenshots (if applicable): Announce flow failure on DagManager::addDag error Additionally, migrate Orchestrator overall away from deprecated EventSubmitter::getTimingEvent factory method. Presently, addDag failure leaves the flow marooned in the COMPILED state, as the warranted FLOW_FAILED event is never sent. Particularly insidious is that scheduled flows with their execution stuck in COMPILED miss their next execution, unless flow.allowConcurrentExecutions is set. Thus the scheduled flow is stuck in its entirety, not merely a single execution. One observed cause of addDag failure is when the DagStateStore is backed by a replicated DB (e.g. MySqlDagStateStore) that just switched leaders. Cached connections in the pool may suddenly point to a read-only follower unable to DagStateStore::writeCheckpoint. Tests [ ] My PR adds the following unit tests OR does not need testing for this extremely good reason: Manual testing while running GaaS locally: I added a flow to runImmediatly, which I soon after observed COMPLETE then I locally patched DagManager::addDag to mimic the motivating failure scenario by invariably throwing an IOException. I again added the same flow as 1.) (adjusted only to bear a unique name and target location) to runImmediately I observed that second flow as FAILED with the shimmed exception conveyed in the message. a. patched DagManager: --- a/gobblin-service/src/main/java/org/apache/gobblin/service/modules/orchestration/DagManager.java +++ b/gobblin-service/src/main/java/org/apache/gobblin/service/modules/orchestration/DagManager.java @@ -263,7 +263,8 @@ public class DagManager extends AbstractIdleService { synchronized void addDag(Dag<JobExecutionPlan> dag, boolean persist, boolean setStatus) throws IOException { if (persist) { //Persist the dag - this.dagStateStore.writeCheckpoint(dag); + throw new IOException("No, I won't add the DAG, dawg!"); + // this.dagStateStore.writeCheckpoint(dag); } int queueId = DagManagerUtils.getDagQueueId(dag, this.numThreads); // Add the dag to the specific queue determined by flowExecutionId b. submitted flow: { "id": { "flowName": "test005", "flowGroup": "testKip" }, "templateUris": "FS:///", "properties": { "gobblin.flow.sourceIdentifier": "<<redacted-source>>", "gobblin.flow.destinationIdentifier": "<<(same) redacted-source>>", "user.to.proxy": "gobblintest", "gobblin.flow.input.dataset.descriptor.path": "<<redacted-path>>", "gobblin.flow.output.dataset.descriptor.path": "/tmp/gaas-testing/kip/test005", "gobblin.flow.input.dataset.descriptor.partition.type": "none", "gobblin.flow.output.dataset.descriptor.partition.type": "none", "gobblin.copy.simulate": "false", "flow.applyRetention": "false", "dataset.datetimePattern": "yyyy/MM/dd", "copy.date.pattern": "yyyy/MM/dd" }, "schedule": { "cronSchedule": "0 0 8 * * ? *", "runImmediately": true } } c. observed (FAILED) status: curli -k --dv-auth SELF "https://localhost:6956/sharedgobblinservice/flowexecutions?q=latestFlowExecution&flowId=(flowGroup:testKip,flowName:test005)" -X GET -H 'X-RestLi-Protocol-Version: 2.0.0' % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 361 0 361 0 0 44 0 --:--:-- 0:00:08 --:--:-- 74 { "elements": [ { "id": { "flowGroup": "testKip", "flowExecutionId": 1628209427967, "flowName": "test005" }, "message": "Failed to add Job Execution Plan due to: No, I won't add the DAG, dawg!", "executionStatistics": { "executionEndTime": 1628209428045, "executionStartTime": 1628209427967 }, "jobStatuses": [], "executionStatus": "FAILED" } ], "paging": { "count": 10, "start": 0, "links": [] } } Commits [ ] My commits all reference JIRA issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message": Subject is separated from body by a blank line Subject is limited to 50 characters Subject does not end with a period Subject uses the imperative mood ("add", not "adding") Body wraps at 72 characters Body explains "what" and "why", not "how" Codecov Report Merging #3357 (41ec0a0) into master (de94515) will decrease coverage by 3.44%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #3357 +/- ## ============================================ - Coverage 46.43% 42.99% -3.45% + Complexity 10120 1940 -8180 ============================================ Files 2052 394 -1658 Lines 79656 16886 -62770 Branches 8896 2080 -6816 ============================================ - Hits 36991 7260 -29731 + Misses 39226 8826 -30400 + Partials 3439 800 -2639 Impacted Files Coverage Δ ...in/java/org/apache/gobblin/cluster/HelixUtils.java 32.23% <0.00%> (-2.48%) :arrow_down: ...a/org/apache/gobblin/cluster/GobblinHelixTask.java 58.06% <0.00%> (-2.16%) :arrow_down: .../apache/gobblin/util/test/RetentionTestHelper.java ...gobblin/runtime/mapreduce/GobblinOutputFormat.java ...e/gobblin/runtime/api/GobblinInstanceLauncher.java ...g/apache/gobblin/service/monitoring/JobStatus.java ...va/org/apache/gobblin/publisher/DataPublisher.java ...org/apache/gobblin/source/extractor/Extractor.java .../apache/gobblin/http/ApacheHttpResponseStatus.java ...blin/data/management/copy/RecursivePathFinder.java ... and 1649 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update de94515...41ec0a0. Read the comment docs.
gharchive/pull-request
2021-08-09T21:30:26
2025-04-01T04:33:28.406251
{ "authors": [ "codecov-commenter", "phet" ], "repo": "apache/gobblin", "url": "https://github.com/apache/gobblin/pull/3357", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2482454348
[#4662] improve(IT): Add ranger authorization Hive E2E test What changes were proposed in this pull request? Use an isolated class loader to create an authorization Ranger plugin. Add authorization for Hive E2E tests using Ranger via Gravitino RESTful access control. Copy the authorization ranger jar to the distribution/package/authorizations/ranger/libs. Add Apache Ranger to license.bin Why are the changes needed? #4662 Does this PR introduce any user-facing change? N/A How was this patch tested? CI @jerryshao Please help me review this PR, thanks.
gharchive/pull-request
2024-08-23T06:58:11
2025-04-01T04:33:28.409811
{ "authors": [ "xunliu" ], "repo": "apache/gravitino", "url": "https://github.com/apache/gravitino/pull/4651", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2485068850
GUACAMOLE-1866: Clean up style of recent connection removal button. This change restyles the recent connection removal button such that: It's a gray-colored "X" that becomes black when hovered. The button itself is not visible unless the associated recent connection is being hovered. Closing in favor of broader UI cleanup around the newly configurable "Recent Connections" section.
gharchive/pull-request
2024-08-25T05:33:50
2025-04-01T04:33:28.411282
{ "authors": [ "mike-jumper" ], "repo": "apache/guacamole-client", "url": "https://github.com/apache/guacamole-client/pull/1008", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
506966656
HDDS-2194. Replication of Container fails with "Only closed containers could be exported" https://issues.apache.org/jira/browse/HDDS-2194 The issue is because when the Replication Manager is considering to replicate containers which are under replicated, it considers replicas in QuasiClosed/Closed. Whereas in Datanode we have a check of Closed. This has caused the issue. So, that is why we see IllegealStateException in the logs. Test failures are not related. Thanks @bharatviswa504 for the contribution and @adoroszlai for review.
gharchive/pull-request
2019-10-15T02:29:57
2025-04-01T04:33:28.413002
{ "authors": [ "bharatviswa504", "nandakumar131" ], "repo": "apache/hadoop-ozone", "url": "https://github.com/apache/hadoop-ozone/pull/25", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
504997511
YARN-7243. Moving logging APIs over to slf4j in hadoop-yarn-server-resourcemanager. Backport https://issues.apache.org/jira/browse/YARN-7243 to branch-3.2 to fix compile failure caused by https://issues.apache.org/jira/browse/YARN-9873 All the conflicts are import sections. Ran {{mvn install -DskipTests}} successfully. :broken_heart: -1 overall Vote Subsystem Runtime Comment 0 reexec 531 Docker mode activated. _ Prechecks _ +1 dupname 6 No case conflicting files found. +1 @author 0 The patch does not contain any @author tags. +1 test4tests 0 The patch appears to include 74 new or modified test files. _ branch-3.2 Compile Tests _ 0 mvndep 17 Maven dependency ordering for branch -1 mvninstall 399 root in branch-3.2 failed. -1 compile 367 root in branch-3.2 failed. -0 checkstyle 207 The patch fails to run checkstyle in root -1 mvnsite 29 hadoop-yarn-server-resourcemanager in branch-3.2 failed. -1 shadedclient 550 branch has errors when building and testing our client artifacts. +1 javadoc 83 branch-3.2 passed 0 spotbugs 103 Used deprecated FindBugs config; considering switching to SpotBugs. -1 findbugs 27 hadoop-yarn-server-resourcemanager in branch-3.2 failed. _ Patch Compile Tests _ 0 mvndep 18 Maven dependency ordering for patch +1 mvninstall 91 the patch passed +1 compile 906 the patch passed -1 javac 906 root generated 517 new + 804 unchanged - 8 fixed = 1321 total (was 812) -0 checkstyle 219 The patch fails to run checkstyle in root +1 mvnsite 134 the patch passed +1 whitespace 0 The patch has no whitespace issues. +1 shadedclient 707 patch has no errors when building and testing our client artifacts. +1 javadoc 114 the patch passed +1 findbugs 227 the patch passed _ Other Tests _ +1 unit 536 hadoop-common in the patch passed. -1 unit 4292 hadoop-yarn-server-resourcemanager in the patch failed. +1 asflicense 49 The patch does not generate ASF License warnings. 9655 Reason Tests Failed junit tests hadoop.yarn.server.resourcemanager.metrics.TestSystemMetricsPublisherForV2 hadoop.yarn.server.resourcemanager.metrics.TestCombinedSystemMetricsPublisher Subsystem Report/Notes Docker Client=19.03.3 Server=19.03.3 base: https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/1634 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle uname Linux c9a61fad2458 4.15.0-58-generic #64-Ubuntu SMP Tue Aug 6 11:12:41 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality personality/hadoop.sh git revision branch-3.2 / ae8ca70 Default Java 1.8.0_222 mvninstall https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-mvninstall-root.txt compile https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-compile-root.txt checkstyle https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out//home/jenkins/jenkins-slave/workspace/hadoop-multibranch_PR-1634/out/maven-branch-checkstyle-root.txt mvnsite https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-mvnsite-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt findbugs https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt javac https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/diff-compile-javac-root.txt checkstyle https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out//home/jenkins/jenkins-slave/workspace/hadoop-multibranch_PR-1634/out/maven-patch-checkstyle-root.txt unit https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt Test Results https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/testReport/ Max. process+thread count 1532 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: . Console output https://builds.apache.org/job/hadoop-multibranch/job/PR-1634/1/console versions git=2.7.4 maven=3.3.9 findbugs=3.1.0-RC1 Powered by Apache Yetus 0.10.0 http://yetus.apache.org This message was automatically generated. The test failures are not related to the patch. Merging this into branch-3.2.
gharchive/pull-request
2019-10-10T02:39:59
2025-04-01T04:33:28.443369
{ "authors": [ "aajisaka", "hadoop-yetus" ], "repo": "apache/hadoop", "url": "https://github.com/apache/hadoop/pull/1634", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
655719130
MAPREDUCE-7284. TestCombineFileInputFormat#testMissingBlocks fails JIRA: https://issues.apache.org/jira/browse/MAPREDUCE-7284 :broken_heart: -1 overall Vote Subsystem Runtime Comment +0 :ok: reexec 0m 31s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 1 new or modified test files. _ trunk Compile Tests _ +1 :green_heart: mvninstall 22m 2s trunk passed +1 :green_heart: compile 0m 36s trunk passed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04 +1 :green_heart: compile 0m 33s trunk passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09 +1 :green_heart: checkstyle 0m 30s trunk passed +1 :green_heart: mvnsite 0m 32s trunk passed +1 :green_heart: shadedclient 15m 11s branch has no errors when building and testing our client artifacts. -1 :x: javadoc 0m 26s hadoop-mapreduce-client-jobclient in trunk failed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04. +1 :green_heart: javadoc 0m 24s trunk passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09 +0 :ok: spotbugs 0m 46s Used deprecated FindBugs config; considering switching to SpotBugs. +1 :green_heart: findbugs 0m 44s trunk passed _ Patch Compile Tests _ +1 :green_heart: mvninstall 0m 31s the patch passed +1 :green_heart: compile 0m 30s the patch passed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04 +1 :green_heart: javac 0m 30s the patch passed +1 :green_heart: compile 0m 27s the patch passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09 +1 :green_heart: javac 0m 27s the patch passed +1 :green_heart: checkstyle 0m 20s the patch passed +1 :green_heart: mvnsite 0m 29s the patch passed +1 :green_heart: whitespace 0m 0s The patch has no whitespace issues. +1 :green_heart: shadedclient 14m 9s patch has no errors when building and testing our client artifacts. -1 :x: javadoc 0m 21s hadoop-mapreduce-client-jobclient in the patch failed with JDK Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04. +1 :green_heart: javadoc 0m 17s the patch passed with JDK Private Build-1.8.0_252-8u252-b09-1~18.04-b09 +1 :green_heart: findbugs 0m 45s the patch passed _ Other Tests _ +1 :green_heart: unit 123m 21s hadoop-mapreduce-client-jobclient in the patch passed. -1 :x: asflicense 0m 42s The patch generated 1 ASF License warnings. 185m 20s Subsystem Report/Notes Docker ClientAPI=1.40 ServerAPI=1.40 base: https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/2136 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle uname Linux 65816539f21c 4.15.0-60-generic #67-Ubuntu SMP Thu Aug 22 16:55:30 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality personality/hadoop.sh git revision trunk / b9fa5e0182c Default Java Private Build-1.8.0_252-8u252-b09-1~18.04-b09 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_252-8u252-b09-1~18.04-b09 javadoc https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/branch-javadoc-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient-jdkUbuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04.txt javadoc https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/patch-javadoc-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient-jdkUbuntu-11.0.7+10-post-Ubuntu-2ubuntu218.04.txt Test Results https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/testReport/ asflicense https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/artifact/out/patch-asflicense-problems.txt Max. process+thread count 1330 (vs. ulimit of 5500) modules C: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient U: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient Console output https://builds.apache.org/job/hadoop-multibranch/job/PR-2136/1/console versions git=2.17.1 maven=3.6.0 findbugs=3.1.0-RC1 Powered by Apache Yetus 0.12.0 https://yetus.apache.org This message was automatically generated. Thank you @iwasakims
gharchive/pull-request
2020-07-13T09:58:44
2025-04-01T04:33:28.474157
{ "authors": [ "aajisaka", "hadoop-yetus" ], "repo": "apache/hadoop", "url": "https://github.com/apache/hadoop/pull/2136", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1329141741
HADOOP-18392. Propagate vectored s3a input stream stats to file system stats. Description of PR Propagate vectored s3a input stream stats to file system stats. How was this patch tested? Re-ran the vectored io tests using us-east-1 endpoint. For code changes: [x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')? [x] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation? [ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0? [ ] If applicable, have you updated the LICENSE, LICENSE-binary, NOTICE-binary files? :confetti_ball: +1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 59s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 2 new or modified test files. _ trunk Compile Tests _ +0 :ok: mvndep 15m 14s Maven dependency ordering for branch +1 :green_heart: mvninstall 25m 24s trunk passed +1 :green_heart: compile 23m 14s trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: compile 21m 2s trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: checkstyle 4m 24s trunk passed +1 :green_heart: mvnsite 3m 47s trunk passed +1 :green_heart: javadoc 2m 54s trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javadoc 2m 34s trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: spotbugs 5m 3s trunk passed +1 :green_heart: shadedclient 22m 12s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +0 :ok: mvndep 0m 33s Maven dependency ordering for patch +1 :green_heart: mvninstall 1m 45s the patch passed +1 :green_heart: compile 22m 32s the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javac 22m 32s the patch passed +1 :green_heart: compile 21m 1s the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: javac 21m 1s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 4m 22s the patch passed +1 :green_heart: mvnsite 3m 37s the patch passed +1 :green_heart: javadoc 2m 50s the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javadoc 2m 39s the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: spotbugs 5m 12s the patch passed +1 :green_heart: shadedclient 22m 33s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 18m 50s hadoop-common in the patch passed. +1 :green_heart: unit 3m 34s hadoop-aws in the patch passed. +1 :green_heart: asflicense 1m 35s The patch does not generate ASF License warnings. 243m 3s Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/4704 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux 7114ac705ada 4.15.0-65-generic #74-Ubuntu SMP Tue Sep 17 17:06:04 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / fef7f5f6e5c79af4f16cd5a52af56317cf260a30 Default Java Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/1/testReport/ Max. process+thread count 1330 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws U: . Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/1/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 57s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 0s codespell was not available. +0 :ok: detsecrets 0m 0s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 2 new or modified test files. _ trunk Compile Tests _ +0 :ok: mvndep 15m 30s Maven dependency ordering for branch +1 :green_heart: mvninstall 25m 36s trunk passed +1 :green_heart: compile 23m 20s trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: compile 20m 45s trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: checkstyle 4m 22s trunk passed +1 :green_heart: mvnsite 3m 35s trunk passed +1 :green_heart: javadoc 2m 57s trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javadoc 2m 41s trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: spotbugs 5m 5s trunk passed +1 :green_heart: shadedclient 22m 17s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +0 :ok: mvndep 0m 30s Maven dependency ordering for patch +1 :green_heart: mvninstall 1m 47s the patch passed +1 :green_heart: compile 22m 28s the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javac 22m 28s the patch passed +1 :green_heart: compile 20m 43s the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: javac 20m 43s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. -0 :warning: checkstyle 4m 9s /results-checkstyle-root.txt root: The patch generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) +1 :green_heart: mvnsite 3m 44s the patch passed +1 :green_heart: javadoc 2m 48s the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javadoc 2m 41s the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: spotbugs 5m 11s the patch passed +1 :green_heart: shadedclient 22m 28s patch has no errors when building and testing our client artifacts. _ Other Tests _ -1 :x: unit 18m 55s /patch-unit-hadoop-common-project_hadoop-common.txt hadoop-common in the patch passed. +1 :green_heart: unit 3m 25s hadoop-aws in the patch passed. +1 :green_heart: asflicense 1m 37s The patch does not generate ASF License warnings. 242m 36s Reason Tests Failed junit tests hadoop.fs.contract.localfs.TestLocalFSContractVectoredRead hadoop.fs.contract.rawlocal.TestRawLocalContractVectoredRead Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/2/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/4704 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux e9a142c077c3 4.15.0-65-generic #74-Ubuntu SMP Tue Sep 17 17:06:04 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / a65a68dda3b020b35b116cb97a59e5e1b09d98d3 Default Java Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/2/testReport/ Max. process+thread count 1859 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws U: . Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/2/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. :confetti_ball: +1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 56s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 2 new or modified test files. _ trunk Compile Tests _ +0 :ok: mvndep 15m 15s Maven dependency ordering for branch +1 :green_heart: mvninstall 25m 25s trunk passed +1 :green_heart: compile 23m 13s trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: compile 20m 50s trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: checkstyle 4m 23s trunk passed +1 :green_heart: mvnsite 3m 47s trunk passed +1 :green_heart: javadoc 2m 50s trunk passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javadoc 2m 36s trunk passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: spotbugs 5m 1s trunk passed +1 :green_heart: shadedclient 22m 25s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +0 :ok: mvndep 0m 35s Maven dependency ordering for patch +1 :green_heart: mvninstall 1m 42s the patch passed +1 :green_heart: compile 22m 29s the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javac 22m 29s the patch passed +1 :green_heart: compile 20m 54s the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: javac 20m 54s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 4m 11s the patch passed +1 :green_heart: mvnsite 3m 36s the patch passed +1 :green_heart: javadoc 2m 51s the patch passed with JDK Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 +1 :green_heart: javadoc 2m 40s the patch passed with JDK Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 +1 :green_heart: spotbugs 5m 16s the patch passed +1 :green_heart: shadedclient 22m 32s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 18m 48s hadoop-common in the patch passed. +1 :green_heart: unit 3m 22s hadoop-aws in the patch passed. +1 :green_heart: asflicense 1m 36s The patch does not generate ASF License warnings. 242m 33s Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/3/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/4704 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux c9c07ee729e1 4.15.0-65-generic #74-Ubuntu SMP Tue Sep 17 17:06:04 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / cad85597b88ae52871472fed31e0f30c3cc49c68 Default Java Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Private Build-11.0.15+10-Ubuntu-0ubuntu0.20.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_312-8u312-b07-0ubuntu1~20.04-b07 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/3/testReport/ Max. process+thread count 3159 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws U: . Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-4704/3/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated.
gharchive/pull-request
2022-08-04T20:27:17
2025-04-01T04:33:28.576851
{ "authors": [ "hadoop-yetus", "mukund-thakur" ], "repo": "apache/hadoop", "url": "https://github.com/apache/hadoop/pull/4704", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1537931087
YARN-11419. Simplify code and improve readability Description of PR Remove YarnException and IOException, because they are all subclasses of Throwable. Use try-resource. A non-empty string should be on the left side of the equal. How was this patch tested? Origin uts. For code changes: [x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')? [ ] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation? [ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0? [ ] If applicable, have you updated the LICENSE, LICENSE-binary, NOTICE-binary files? :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 49s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 43m 29s trunk passed +1 :green_heart: compile 0m 37s trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: compile 0m 35s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: checkstyle 0m 35s trunk passed +1 :green_heart: mvnsite 0m 39s trunk passed -1 :x: javadoc 0m 40s /branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 29s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 1m 7s trunk passed +1 :green_heart: shadedclient 23m 19s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 0m 34s the patch passed +1 :green_heart: compile 0m 28s the patch passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: javac 0m 28s the patch passed +1 :green_heart: compile 0m 24s the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: javac 0m 24s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. -0 :warning: checkstyle 0m 17s /results-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client: The patch generated 2 new + 14 unchanged - 0 fixed = 16 total (was 14) +1 :green_heart: mvnsite 0m 28s the patch passed -1 :x: javadoc 0m 22s /patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 21s the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 0m 56s the patch passed +1 :green_heart: shadedclient 23m 8s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 28m 10s hadoop-yarn-client in the patch passed. +1 :green_heart: asflicense 0m 41s The patch does not generate ASF License warnings. 129m 37s Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5309 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux 29d2cd1531f3 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 15d3246fa4e299adf24dd9335998d886b117a0e9 Default Java Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/1/testReport/ Max. process+thread count 577 (vs. ulimit of 5500) modules C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/1/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. Hi @smallzhongfeng , Could you be more specific with the title of PR / jira about what are you improving / cleaning up? Class name should be in it at least. Hi @smallzhongfeng , Could you be more specific with the title of PR / jira about what are you improving / cleaning up? Class name should be in it at least. Thank you for your reminder, could you help me review this? @szilard-nemeth @slfan1989 Thx ! :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 58s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 48m 51s trunk passed +1 :green_heart: compile 0m 44s trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: compile 0m 40s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: checkstyle 0m 34s trunk passed +1 :green_heart: mvnsite 0m 46s trunk passed -1 :x: javadoc 0m 46s /branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 32s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 1m 26s trunk passed -1 :x: shadedclient 28m 36s branch has errors when building and testing our client artifacts. _ Patch Compile Tests _ -1 :x: mvninstall 0m 23s /patch-mvninstall-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt hadoop-yarn-client in the patch failed. -1 :x: compile 0m 22s /patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. -1 :x: javac 0m 22s /patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. -1 :x: compile 0m 22s /patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkPrivateBuild-1.8.0_352-8u352-ga-1~20.04-b08.txt hadoop-yarn-client in the patch failed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08. -1 :x: javac 0m 22s /patch-compile-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkPrivateBuild-1.8.0_352-8u352-ga-1~20.04-b08.txt hadoop-yarn-client in the patch failed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08. +1 :green_heart: blanks 0m 0s The patch has no blanks issues. -0 :warning: checkstyle 0m 20s /buildtool-patch-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt The patch fails to run checkstyle in hadoop-yarn-client -1 :x: mvnsite 0m 23s /patch-mvnsite-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt hadoop-yarn-client in the patch failed. -1 :x: javadoc 0m 22s /patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. -1 :x: javadoc 0m 22s /patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkPrivateBuild-1.8.0_352-8u352-ga-1~20.04-b08.txt hadoop-yarn-client in the patch failed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08. -1 :x: spotbugs 0m 22s /patch-spotbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt hadoop-yarn-client in the patch failed. +1 :green_heart: shadedclient 31m 9s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 28m 16s hadoop-yarn-client in the patch passed. +1 :green_heart: asflicense 0m 41s The patch does not generate ASF License warnings. 145m 33s Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/2/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5309 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux faa74cdeb796 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / e6cf7b764f1feb0dee2a9ddbe5d1de516f65a254 Default Java Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/2/testReport/ Max. process+thread count 576 (vs. ulimit of 5500) modules C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/2/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. @smallzhongfeng Thank you very much for your contribution, We need to fix checkstyle :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 50s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 44m 4s trunk passed +1 :green_heart: compile 0m 38s trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: compile 0m 34s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: checkstyle 0m 34s trunk passed +1 :green_heart: mvnsite 0m 39s trunk passed -1 :x: javadoc 0m 39s /branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 27s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 1m 7s trunk passed +1 :green_heart: shadedclient 23m 4s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 0m 33s the patch passed +1 :green_heart: compile 0m 29s the patch passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: javac 0m 29s the patch passed +1 :green_heart: compile 0m 25s the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: javac 0m 25s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 0m 17s the patch passed +1 :green_heart: mvnsite 0m 28s the patch passed -1 :x: javadoc 0m 22s /patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 21s the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 0m 58s the patch passed +1 :green_heart: shadedclient 23m 10s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 28m 14s hadoop-yarn-client in the patch passed. +1 :green_heart: asflicense 0m 38s The patch does not generate ASF License warnings. 130m 3s Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/3/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5309 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux 08b67f6886cb 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / f59718b431a01b9652003d4cadf54192e4fedf33 Default Java Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/3/testReport/ Max. process+thread count 576 (vs. ulimit of 5500) modules C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/3/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 52s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 0s codespell was not available. +0 :ok: detsecrets 0m 0s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 43m 20s trunk passed +1 :green_heart: compile 0m 38s trunk passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: compile 0m 35s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: checkstyle 0m 35s trunk passed +1 :green_heart: mvnsite 0m 39s trunk passed -1 :x: javadoc 0m 41s /branch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in trunk failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 29s trunk passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 1m 7s trunk passed +1 :green_heart: shadedclient 23m 28s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 0m 34s the patch passed +1 :green_heart: compile 0m 29s the patch passed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 +1 :green_heart: javac 0m 29s the patch passed +1 :green_heart: compile 0m 26s the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: javac 0m 25s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 0m 18s the patch passed +1 :green_heart: mvnsite 0m 28s the patch passed -1 :x: javadoc 0m 22s /patch-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client-jdkUbuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04.txt hadoop-yarn-client in the patch failed with JDK Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04. +1 :green_heart: javadoc 0m 21s the patch passed with JDK Private Build-1.8.0_352-8u352-ga-1~20.04-b08 +1 :green_heart: spotbugs 0m 56s the patch passed +1 :green_heart: shadedclient 23m 6s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 28m 14s hadoop-yarn-client in the patch passed. +1 :green_heart: asflicense 0m 40s The patch does not generate ASF License warnings. 129m 47s Subsystem Report/Notes Docker ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/4/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5309 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux 06bd08ac6e2e 4.15.0-200-generic #211-Ubuntu SMP Thu Nov 24 18:16:04 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 05b14c1f60b61487ed06fabbba84b6c1ac373f35 Default Java Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.17+8-post-Ubuntu-1ubuntu220.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_352-8u352-ga-1~20.04-b08 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/4/testReport/ Max. process+thread count 576 (vs. ulimit of 5500) modules C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5309/4/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. merged trunk, @smallzhongfeng Thanks for your contribution! Thanks for your review. @slfan1989 👍
gharchive/pull-request
2023-01-18T12:37:33
2025-04-01T04:33:28.719020
{ "authors": [ "hadoop-yetus", "slfan1989", "smallzhongfeng", "szilard-nemeth" ], "repo": "apache/hadoop", "url": "https://github.com/apache/hadoop/pull/5309", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1660810905
HADOOP-18691. Client.Connection#updateAddress needs to ensure that address is resolved before updating Description of PR When Client.Connection#setupConnection encounters an IOException, it will try to update the server address. (HADOOP-18365) When the address is re-parsed, it may be an unresolved address (UnknownHostException), which causes Client.Connection#setupConnection to fail to reconnect. while (true) { try { if (server.isUnresolved()) { Especially when DN is connected to NN, BPServiceActor#bpNamenode is only initialized once, which causes DN to never connect to NN before restarting. How was this patch tested? For code changes: [ ] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')? [ ] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation? [ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0? [ ] If applicable, have you updated the LICENSE, LICENSE-binary, NOTICE-binary files? :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 52s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 2s codespell was not available. +0 :ok: detsecrets 0m 2s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 58m 43s trunk passed -1 :x: compile 31m 46s /branch-compile-root-jdkUbuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.txt root in trunk failed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1. -1 :x: compile 0m 54s /branch-compile-root-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt root in trunk failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09. +1 :green_heart: checkstyle 2m 8s trunk passed -1 :x: mvnsite 1m 30s /branch-mvnsite-hadoop-common-project_hadoop-common.txt hadoop-common in trunk failed. +1 :green_heart: javadoc 1m 38s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 57s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 3m 51s trunk passed -1 :x: shadedclient 25m 44s branch has errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 1m 29s the patch passed -1 :x: compile 10m 14s /patch-compile-root-jdkUbuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.txt root in the patch failed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1. -1 :x: javac 10m 14s /patch-compile-root-jdkUbuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1.txt root in the patch failed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1. -1 :x: compile 0m 30s /patch-compile-root-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt root in the patch failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09. -1 :x: javac 0m 30s /patch-compile-root-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt root in the patch failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09. +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 1m 15s the patch passed +1 :green_heart: mvnsite 1m 55s the patch passed +1 :green_heart: javadoc 1m 4s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 -1 :x: javadoc 0m 25s /patch-javadoc-hadoop-common-project_hadoop-common-jdkPrivateBuild-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09.txt hadoop-common in the patch failed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09. +1 :green_heart: spotbugs 3m 51s the patch passed -1 :x: shadedclient 8m 39s patch has errors when building and testing our client artifacts. _ Other Tests _ -1 :x: unit 0m 25s /patch-unit-hadoop-common-project_hadoop-common.txt hadoop-common in the patch failed. +0 :ok: asflicense 0m 25s ASF License check generated no output? 155m 59s Subsystem Report/Notes Docker ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5542 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux 7f78a852880b 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 29ac1d27728be1d3e68a4ca158ed21fd673d5e02 Default Java Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/1/testReport/ Max. process+thread count 186 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/1/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 48s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 43m 50s trunk passed +1 :green_heart: compile 25m 5s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: compile 21m 42s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: checkstyle 1m 5s trunk passed +1 :green_heart: mvnsite 1m 39s trunk passed +1 :green_heart: javadoc 1m 9s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 42s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 2m 43s trunk passed +1 :green_heart: shadedclient 25m 37s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 1m 0s the patch passed +1 :green_heart: compile 24m 37s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javac 24m 37s the patch passed +1 :green_heart: compile 21m 43s the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: javac 21m 43s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 0m 59s the patch passed +1 :green_heart: mvnsite 1m 35s the patch passed +1 :green_heart: javadoc 1m 1s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 41s the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 2m 40s the patch passed +1 :green_heart: shadedclient 25m 14s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 18m 9s hadoop-common in the patch passed. +1 :green_heart: asflicense 0m 53s The patch does not generate ASF License warnings. 223m 6s Subsystem Report/Notes Docker ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/2/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5542 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux b1511504b561 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 596bb839cdbdf52a34f324c8313e5777aa604894 Default Java Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/2/testReport/ Max. process+thread count 1375 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/2/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. :confetti_ball: +1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 17m 31s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 0s codespell was not available. +0 :ok: detsecrets 0m 0s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 1 new or modified test files. _ trunk Compile Tests _ +1 :green_heart: mvninstall 42m 46s trunk passed +1 :green_heart: compile 25m 14s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: compile 21m 43s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: checkstyle 1m 6s trunk passed +1 :green_heart: mvnsite 1m 38s trunk passed +1 :green_heart: javadoc 1m 9s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 43s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 2m 47s trunk passed +1 :green_heart: shadedclient 25m 24s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 1m 1s the patch passed +1 :green_heart: compile 27m 2s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javac 27m 2s the patch passed +1 :green_heart: compile 22m 55s the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: javac 22m 55s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 1m 0s the patch passed +1 :green_heart: mvnsite 1m 42s the patch passed +1 :green_heart: javadoc 1m 0s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 40s the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 2m 46s the patch passed +1 :green_heart: shadedclient 25m 12s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 18m 25s hadoop-common in the patch passed. +1 :green_heart: asflicense 0m 54s The patch does not generate ASF License warnings. 242m 38s Subsystem Report/Notes Docker ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/3/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5542 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux c172ac28a844 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 9620ed8db83b5f4c5bc73ecf86b6efc3b5f05275 Default Java Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/3/testReport/ Max. process+thread count 1375 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/3/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. :confetti_ball: +1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 0m 47s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 1s codespell was not available. +0 :ok: detsecrets 0m 1s detect-secrets was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 1 new or modified test files. _ trunk Compile Tests _ +1 :green_heart: mvninstall 42m 41s trunk passed +1 :green_heart: compile 25m 20s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: compile 21m 48s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: checkstyle 1m 7s trunk passed +1 :green_heart: mvnsite 1m 37s trunk passed +1 :green_heart: javadoc 1m 11s trunk passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 41s trunk passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 2m 43s trunk passed +1 :green_heart: shadedclient 25m 34s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 1m 0s the patch passed +1 :green_heart: compile 24m 33s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javac 24m 33s the patch passed +1 :green_heart: compile 21m 36s the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: javac 21m 36s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: checkstyle 1m 0s the patch passed +1 :green_heart: mvnsite 1m 36s the patch passed +1 :green_heart: javadoc 0m 59s the patch passed with JDK Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 +1 :green_heart: javadoc 0m 42s the patch passed with JDK Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 +1 :green_heart: spotbugs 2m 43s the patch passed +1 :green_heart: shadedclient 25m 11s patch has no errors when building and testing our client artifacts. _ Other Tests _ +1 :green_heart: unit 18m 31s hadoop-common in the patch passed. +1 :green_heart: asflicense 0m 52s The patch does not generate ASF License warnings. 222m 12s Subsystem Report/Notes Docker ClientAPI=1.42 ServerAPI=1.42 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/4/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/5542 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets uname Linux 8d8fd1b002d9 4.15.0-206-generic #217-Ubuntu SMP Fri Feb 3 19:10:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 4c1c871292b40c5d38ce0e232bc19ef8949d81ef Default Java Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.18+10-post-Ubuntu-0ubuntu120.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_362-8u362-ga-0ubuntu1~20.04.1-b09 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/4/testReport/ Max. process+thread count 2253 (vs. ulimit of 5500) modules C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-5542/4/console versions git=2.25.1 maven=3.6.3 spotbugs=4.2.2 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated.
gharchive/pull-request
2023-04-10T13:50:01
2025-04-01T04:33:28.852365
{ "authors": [ "cxzl25", "hadoop-yetus" ], "repo": "apache/hadoop", "url": "https://github.com/apache/hadoop/pull/5542", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2205076172
HDFS-17441. Fix junit dependency by adding missing library in hadoop-hdfs-rbf. Description of PR We need to add some missing junit libraries in hadoop-hdfs-rbf. Based on Ayush's advice, we need to add junit-platform-launcher. We might also need junit-jupiter-api, similar to other modules. How was this patch tested? For code changes: [x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')? @ayushtkn Could you please review it if you have bandwidth? :broken_heart: -1 overall Vote Subsystem Runtime Logfile Comment +0 :ok: reexec 6m 41s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: codespell 0m 0s codespell was not available. +0 :ok: detsecrets 0m 0s detect-secrets was not available. +0 :ok: xmllint 0m 0s xmllint was not available. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. -1 :x: test4tests 0m 0s The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. _ trunk Compile Tests _ +1 :green_heart: mvninstall 31m 42s trunk passed +1 :green_heart: compile 0m 25s trunk passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1 +1 :green_heart: compile 0m 23s trunk passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06 +1 :green_heart: mvnsite 0m 26s trunk passed +1 :green_heart: javadoc 0m 30s trunk passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1 +1 :green_heart: javadoc 0m 22s trunk passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06 +1 :green_heart: shadedclient 53m 29s branch has no errors when building and testing our client artifacts. _ Patch Compile Tests _ +1 :green_heart: mvninstall 0m 20s the patch passed +1 :green_heart: compile 0m 19s the patch passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1 +1 :green_heart: javac 0m 19s the patch passed +1 :green_heart: compile 0m 18s the patch passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06 +1 :green_heart: javac 0m 18s the patch passed +1 :green_heart: blanks 0m 0s The patch has no blanks issues. +1 :green_heart: mvnsite 0m 22s the patch passed +1 :green_heart: javadoc 0m 19s the patch passed with JDK Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1 +1 :green_heart: javadoc 0m 18s the patch passed with JDK Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06 +1 :green_heart: shadedclient 20m 44s patch has no errors when building and testing our client artifacts. _ Other Tests _ -1 :x: unit 26m 7s /patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt hadoop-hdfs-rbf in the patch passed. +1 :green_heart: asflicense 0m 26s The patch does not generate ASF License warnings. 112m 4s Reason Tests Failed junit tests hadoop.hdfs.server.federation.router.TestRouterRpc hadoop.hdfs.server.federation.router.TestRouterRpcMultiDestination Subsystem Report/Notes Docker ClientAPI=1.45 ServerAPI=1.45 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6669/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hadoop/pull/6669 Optional Tests dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient codespell detsecrets xmllint uname Linux c1ea8b931ca6 5.15.0-94-generic #104-Ubuntu SMP Tue Jan 9 15:25:40 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/bin/hadoop.sh git revision trunk / 41759a731f419a773c73fe0f62ab5586be4ffcfa Default Java Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06 Multi-JDK versions /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.22+7-post-Ubuntu-0ubuntu220.04.1 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_402-8u402-ga-2ubuntu1~20.04-b06 Test Results https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6669/1/testReport/ Max. process+thread count 3981 (vs. ulimit of 5500) modules C: hadoop-hdfs-project/hadoop-hdfs-rbf U: hadoop-hdfs-project/hadoop-hdfs-rbf Console output https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6669/1/console versions git=2.25.1 maven=3.6.3 Powered by Apache Yetus 0.14.0 https://yetus.apache.org This message was automatically generated. The failed tests are filed by HDFS-17435. @ayushtkn Thanks for your review!
gharchive/pull-request
2024-03-25T07:24:50
2025-04-01T04:33:28.886378
{ "authors": [ "hadoop-yetus", "tasanuma" ], "repo": "apache/hadoop", "url": "https://github.com/apache/hadoop/pull/6669", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1106441239
Backport HBASE-26579 Set storage policy of recovered edits when wal storage type is configured https://issues.apache.org/jira/browse/HBASE-26678 :confetti_ball: +1 overall Vote Subsystem Runtime Comment +0 :ok: reexec 4m 42s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +1 :green_heart: hbaseanti 0m 0s Patch does not have any anti-patterns. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. +1 :green_heart: test4tests 0m 0s The patch appears to include 1 new or modified test files. _ branch-1 Compile Tests _ +1 :green_heart: mvninstall 10m 36s branch-1 passed +1 :green_heart: compile 0m 44s branch-1 passed with JDK Azul Systems, Inc.-1.8.0_262-b19 +1 :green_heart: compile 0m 49s branch-1 passed with JDK Azul Systems, Inc.-1.7.0_272-b10 +1 :green_heart: checkstyle 1m 43s branch-1 passed +1 :green_heart: shadedjars 3m 15s branch has no errors when building our shaded downstream artifacts. +1 :green_heart: javadoc 0m 48s branch-1 passed with JDK Azul Systems, Inc.-1.8.0_262-b19 +1 :green_heart: javadoc 0m 42s branch-1 passed with JDK Azul Systems, Inc.-1.7.0_272-b10 +0 :ok: spotbugs 3m 9s Used deprecated FindBugs config; considering switching to SpotBugs. +1 :green_heart: findbugs 3m 7s branch-1 passed _ Patch Compile Tests _ +1 :green_heart: mvninstall 2m 7s the patch passed +1 :green_heart: compile 0m 45s the patch passed with JDK Azul Systems, Inc.-1.8.0_262-b19 +1 :green_heart: javac 0m 45s the patch passed +1 :green_heart: compile 0m 48s the patch passed with JDK Azul Systems, Inc.-1.7.0_272-b10 +1 :green_heart: javac 0m 48s the patch passed +1 :green_heart: checkstyle 1m 31s the patch passed +1 :green_heart: whitespace 0m 0s The patch has no whitespace issues. +1 :green_heart: shadedjars 3m 4s patch has no errors when building our shaded downstream artifacts. +1 :green_heart: hadoopcheck 5m 10s Patch does not cause any errors with Hadoop 2.8.5 2.9.2. +1 :green_heart: javadoc 0m 32s the patch passed with JDK Azul Systems, Inc.-1.8.0_262-b19 +1 :green_heart: javadoc 0m 41s the patch passed with JDK Azul Systems, Inc.-1.7.0_272-b10 +1 :green_heart: findbugs 2m 59s the patch passed _ Other Tests _ +1 :green_heart: unit 132m 14s hbase-server in the patch passed. +1 :green_heart: asflicense 0m 41s The patch does not generate ASF License warnings. 180m 55s Subsystem Report/Notes Docker ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/HBase/job/HBase-PreCommit-GitHub-PR/job/PR-4037/1/artifact/out/Dockerfile GITHUB PR https://github.com/apache/hbase/pull/4037 Optional Tests dupname asflicense javac javadoc unit spotbugs findbugs shadedjars hadoopcheck hbaseanti checkstyle compile uname Linux ef48a0add413 4.15.0-112-generic #113-Ubuntu SMP Thu Jul 9 23:41:39 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality /home/jenkins/jenkins-home/workspace/Base-PreCommit-GitHub-PR_PR-4037/out/precommit/personality/provided.sh git revision branch-1 / 598b453a41 Default Java Azul Systems, Inc.-1.7.0_272-b10 Multi-JDK versions /usr/lib/jvm/zulu-8-amd64:Azul Systems, Inc.-1.8.0_262-b19 /usr/lib/jvm/zulu-7-amd64:Azul Systems, Inc.-1.7.0_272-b10 Test Results https://ci-hadoop.apache.org/job/HBase/job/HBase-PreCommit-GitHub-PR/job/PR-4037/1/testReport/ Max. process+thread count 4746 (vs. ulimit of 10000) modules C: hbase-server U: hbase-server Console output https://ci-hadoop.apache.org/job/HBase/job/HBase-PreCommit-GitHub-PR/job/PR-4037/1/console versions git=2.17.1 maven=3.6.0 findbugs=3.0.1 Powered by Apache Yetus 0.12.0 https://yetus.apache.org This message was automatically generated.
gharchive/pull-request
2022-01-18T03:36:40
2025-04-01T04:33:28.915661
{ "authors": [ "Apache-HBase", "YutSean" ], "repo": "apache/hbase", "url": "https://github.com/apache/hbase/pull/4037", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2015726686
HBASE-28212 Do not need to maintain rollback step when root procedure… … does not support rollback (#5538) Signed-off-by: GeorryHuang huangzhuoyue@apache.org (cherry picked from commit 4b015e6a5486394d70bbf5fc0197e469c0987913) (cherry picked from commit 7d6de8b948a088d316124d6587e6a540bc5b38f1) :broken_heart: -1 overall Vote Subsystem Runtime Comment +0 :ok: reexec 0m 44s Docker mode activated. _ Prechecks _ +1 :green_heart: dupname 0m 0s No case conflicting files found. +0 :ok: prototool 0m 0s prototool was not available. +1 :green_heart: hbaseanti 0m 0s Patch does not have any anti-patterns. +1 :green_heart: @author 0m 0s The patch does not contain any @author tags. _ branch-2 Compile Tests _ +0 :ok: mvndep 0m 13s Maven dependency ordering for branch +1 :green_heart: mvninstall 3m 14s branch-2 passed +1 :green_heart: compile 3m 24s branch-2 passed +1 :green_heart: checkstyle 0m 50s branch-2 passed -1 :x: spotless 0m 44s branch has 1 errors when running spotless:check, run spotless:apply to fix. +1 :green_heart: spotbugs 3m 50s branch-2 passed _ Patch Compile Tests _ +0 :ok: mvndep 0m 15s Maven dependency ordering for patch +1 :green_heart: mvninstall 2m 45s the patch passed +1 :green_heart: compile 3m 23s the patch passed +1 :green_heart: cc 3m 23s the patch passed +1 :green_heart: javac 3m 23s the patch passed +1 :green_heart: checkstyle 0m 48s the patch passed +1 :green_heart: whitespace 0m 0s The patch has no whitespace issues. +1 :green_heart: hadoopcheck 15m 10s Patch does not cause any errors with Hadoop 2.10.2 or 3.2.4 3.3.6. +1 :green_heart: hbaseprotoc 1m 13s the patch passed +1 :green_heart: spotless 0m 46s patch has no errors when running spotless:check. +1 :green_heart: spotbugs 4m 14s the patch passed _ Other Tests _ +1 :green_heart: asflicense 0m 23s The patch does not generate ASF License warnings. 43m 51s Subsystem Report/Notes Docker ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-general-check/output/Dockerfile GITHUB PR https://github.com/apache/hbase/pull/5547 Optional Tests dupname asflicense javac spotbugs hadoopcheck hbaseanti spotless checkstyle compile cc hbaseprotoc prototool uname Linux 7a6175b64081 5.4.0-1103-aws #111~18.04.1-Ubuntu SMP Tue May 23 20:04:10 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/hbase-personality.sh git revision branch-2 / ea11923239 Default Java Eclipse Adoptium-11.0.17+8 spotless https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-general-check/output/branch-spotless.txt Max. process+thread count 81 (vs. ulimit of 30000) modules C: hbase-protocol-shaded hbase-procedure hbase-server U: . Console output https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/console versions git=2.34.1 maven=3.8.6 spotbugs=4.7.3 Powered by Apache Yetus 0.12.0 https://yetus.apache.org This message was automatically generated. :confetti_ball: +1 overall Vote Subsystem Runtime Comment +0 :ok: reexec 0m 44s Docker mode activated. -0 :warning: yetus 0m 4s Unprocessed flag(s): --brief-report-file --spotbugs-strict-precheck --whitespace-eol-ignore-list --whitespace-tabs-ignore-list --quick-hadoopcheck _ Prechecks _ _ branch-2 Compile Tests _ +0 :ok: mvndep 0m 12s Maven dependency ordering for branch +1 :green_heart: mvninstall 3m 14s branch-2 passed +1 :green_heart: compile 1m 26s branch-2 passed +1 :green_heart: shadedjars 5m 33s branch has no errors when building our shaded downstream artifacts. +1 :green_heart: javadoc 0m 43s branch-2 passed _ Patch Compile Tests _ +0 :ok: mvndep 0m 18s Maven dependency ordering for patch +1 :green_heart: mvninstall 2m 45s the patch passed +1 :green_heart: compile 1m 28s the patch passed +1 :green_heart: javac 1m 28s the patch passed +1 :green_heart: shadedjars 5m 32s patch has no errors when building our shaded downstream artifacts. +1 :green_heart: javadoc 0m 41s the patch passed _ Other Tests _ +1 :green_heart: unit 0m 30s hbase-protocol-shaded in the patch passed. +1 :green_heart: unit 1m 27s hbase-procedure in the patch passed. +1 :green_heart: unit 207m 47s hbase-server in the patch passed. 236m 34s Subsystem Report/Notes Docker ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-jdk11-hadoop3-check/output/Dockerfile GITHUB PR https://github.com/apache/hbase/pull/5547 Optional Tests javac javadoc unit shadedjars compile uname Linux 12789fc35368 5.4.0-1103-aws #111~18.04.1-Ubuntu SMP Tue May 23 20:04:10 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/hbase-personality.sh git revision branch-2 / ea11923239 Default Java Eclipse Adoptium-11.0.17+8 Test Results https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/testReport/ Max. process+thread count 4759 (vs. ulimit of 30000) modules C: hbase-protocol-shaded hbase-procedure hbase-server U: . Console output https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/console versions git=2.34.1 maven=3.8.6 Powered by Apache Yetus 0.12.0 https://yetus.apache.org This message was automatically generated. :confetti_ball: +1 overall Vote Subsystem Runtime Comment +0 :ok: reexec 0m 52s Docker mode activated. -0 :warning: yetus 0m 7s Unprocessed flag(s): --brief-report-file --spotbugs-strict-precheck --whitespace-eol-ignore-list --whitespace-tabs-ignore-list --quick-hadoopcheck _ Prechecks _ _ branch-2 Compile Tests _ +0 :ok: mvndep 0m 15s Maven dependency ordering for branch +1 :green_heart: mvninstall 3m 49s branch-2 passed +1 :green_heart: compile 1m 46s branch-2 passed +1 :green_heart: shadedjars 6m 34s branch has no errors when building our shaded downstream artifacts. +1 :green_heart: javadoc 0m 53s branch-2 passed _ Patch Compile Tests _ +0 :ok: mvndep 0m 10s Maven dependency ordering for patch +1 :green_heart: mvninstall 3m 22s the patch passed +1 :green_heart: compile 1m 48s the patch passed +1 :green_heart: javac 1m 48s the patch passed +1 :green_heart: shadedjars 6m 27s patch has no errors when building our shaded downstream artifacts. +1 :green_heart: javadoc 0m 53s the patch passed _ Other Tests _ +1 :green_heart: unit 0m 38s hbase-protocol-shaded in the patch passed. +1 :green_heart: unit 1m 57s hbase-procedure in the patch passed. +1 :green_heart: unit 264m 25s hbase-server in the patch passed. 299m 1s Subsystem Report/Notes Docker ClientAPI=1.43 ServerAPI=1.43 base: https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/artifact/yetus-jdk8-hadoop2-check/output/Dockerfile GITHUB PR https://github.com/apache/hbase/pull/5547 Optional Tests javac javadoc unit shadedjars compile uname Linux f8d79484a930 5.4.0-163-generic #180-Ubuntu SMP Tue Sep 5 13:21:23 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux Build tool maven Personality dev-support/hbase-personality.sh git revision branch-2 / ea11923239 Default Java Temurin-1.8.0_352-b08 Test Results https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/testReport/ Max. process+thread count 4169 (vs. ulimit of 30000) modules C: hbase-protocol-shaded hbase-procedure hbase-server U: . Console output https://ci-hbase.apache.org/job/HBase-PreCommit-GitHub-PR/job/PR-5547/1/console versions git=2.34.1 maven=3.8.6 Powered by Apache Yetus 0.12.0 https://yetus.apache.org This message was automatically generated. The spotless error is not a real style problem... [ERROR] An internal error occurred during: "Periodic workspace save.". java.lang.IllegalStateException: Job manager has been shut down. at org.eclipse.core.internal.jobs.JobManager.schedule (JobManager.java:1295) at org.eclipse.core.internal.jobs.InternalJob.schedule (InternalJob.java:385) at org.eclipse.core.runtime.jobs.Job.schedule (Job.java:684) at org.eclipse.core.internal.events.AutoBuildJob.build (AutoBuildJob.java:110) at org.eclipse.core.internal.events.BuildManager.endTopLevel (BuildManager.java:604) at org.eclipse.core.internal.resources.Workspace.endOperation (Workspace.java:1518) at org.eclipse.core.internal.resources.SaveManager.save (SaveManager.java:1246) at org.eclipse.core.internal.resources.SaveManager.save (SaveManager.java:1143) at org.eclipse.core.internal.resources.DelayedSnapshotJob.run (DelayedSnapshotJob.java:55) at org.eclipse.core.internal.jobs.Worker.run (Worker.java:63) Let me merge.
gharchive/pull-request
2023-11-29T03:26:18
2025-04-01T04:33:28.978706
{ "authors": [ "Apache-HBase", "Apache9" ], "repo": "apache/hbase", "url": "https://github.com/apache/hbase/pull/5547", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2502892236
[BUG] [customize dashboard]Multiple rows of data are not displayed correctly Is there an existing issue for this? [X] I have searched the existing issues Current Behavior Multiple rows of data are not displayed correctly Expected Behavior No response Steps To Reproduce No response Environment HertzBeat version(s): Debug logs No response Anything else? No response How to reproduce this bug I have repeatedly tested the code from the latest branch several times but failed to reproduce the issue.
gharchive/issue
2024-09-03T13:29:40
2025-04-01T04:33:28.990631
{ "authors": [ "kerwin612", "zhangshenghang", "zqr10159" ], "repo": "apache/hertzbeat", "url": "https://github.com/apache/hertzbeat/issues/2663", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1458755560
HIVE-23559: Optimise Hive::moveAcidFiles for cloud storage. What changes were proposed in this pull request? Transforming Hive::moveAcidFiles for cloud storage to work in parallel instead of sequentially. Why are the changes needed? Currently this function ends up transferring DELTA, DELETE_DELTA, BASE prefixes sequentially from staging to final location. This causes delays even with simple updates statements, which updates smaller number of records in cloud storage. The proposed change fixes that. Does this PR introduce any user-facing change? No. How was this patch tested? Hive pre-commit testing. Hi @ramesh0201, can you please review? LGTM. +1
gharchive/pull-request
2022-11-21T23:38:46
2025-04-01T04:33:28.994000
{ "authors": [ "difin", "ramesh0201" ], "repo": "apache/hive", "url": "https://github.com/apache/hive/pull/3795", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2526485386
feat: support storage_options param when reading from table To integrate hudi-rs with AWS SDK for Pandas (aws wrangler), we must be able to pass boto_session related aws authentication params (mostly AWS_* params) directly and not only rely on env variable inference. I want to propose adding an option to handle this: storage_options = {"AWS_ACCESS_KEY_ID": "xxxx", "AWS_SECRET_ACCESS_KEY":"xxxx", "AWS_SECRET_ACCESS_TOKEN":"xxxx"} hudi_table = HudiTable("/tmp/trips_table", storage_options=storage_options) records = hudi_table.read_snapshot() Although I want to add this for S3, it should work for other storage backends. I'm happy to contribute and add this. @kazdy sounds good. feel free to take this up and send a pr I'll wait until #72 gets merged. I did the first strawman impl and it requires some refactoring in the Table itself. @xushiyan I also have some questions about this, maybe you can give me your opinion on these: Should we rename Table to HudiTable? I don't know why Timeline and FileSystemView both use separate storage instances, can't they share it, maybe there's a reason why it's done this way I can't see atm? Does it make sense to introduce something that will hold both Timeline and FileSystemView (basically table state) and expose coherent API? thanks hey @kazdy we keep name Table within hudi-core to avoid redundant prefix; everything in hudi-core is about Hudi. When import to other crates, we can give it an alias like HudiTable. We can also add an alias in hudi crate for external facing API when needed. As of now, no strong need for this. Timeline is responsible for data stored in timeline files under .hoodie/, and FileSystemView is responsible for the data stored under the table excluding .hoodie/. It's good to keep things less coupled, unless there is a need for sharing - it's a stateless client performing IO anyway. Maybe you can make a case about why sharing it? Currently Table holds Timeline and FileSystemView. You want to elaborate on what you meant by coherent API?
gharchive/issue
2024-09-14T15:50:55
2025-04-01T04:33:28.999087
{ "authors": [ "kazdy", "xushiyan" ], "repo": "apache/hudi-rs", "url": "https://github.com/apache/hudi-rs/issues/139", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1097088676
[HUDI-3157] Remove aws jars from hudi bundles Tips Thank you very much for contributing to Apache Hudi. Please review https://hudi.apache.org/contribute/how-to-contribute before opening a pull request. What is the purpose of the pull request Remove aws jars from hudi bundles to align with flink-bundle ref: https://github.com/apache/hudi/issues/4474 Brief change log (for example:) Modify AnnotationLocation checkstyle rule in checkstyle.xml Verify this pull request (Please pick either of the following options) This pull request is a trivial rework / code cleanup without any test coverage. (or) This pull request is already covered by existing tests, such as (please describe tests). (or) This change added tests and can be verified as follows: (example:) Added integration tests for end-to-end. Added HoodieClientWriteTest to verify the change. Manually verified the change by running a job locally. Committer checklist [ ] Has a corresponding JIRA in PR title & commit [ ] Commit message is descriptive of the change [ ] CI is green [ ] Necessary doc changes done or have another open PR [ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. @xushiyan @nsivabalan pls take a took. @boneanxs what about also relocating the aws dependency in hudi-aws ? because this alone don't help for aws sdk version mismatch in case they use the dynamodb-lock . they still need to have there own hudi build with shadow relocation ? @xushiyan I think a better approach would be to provide a hudi-aws-bundle with all the different dependencies for aws separately? As @parisni points out as well, there are more issues here? Can we summarize how we are on master and if anything needs to be done before the release. @xushiyan I think a better approach would be to provide a hudi-aws-bundle with all the different dependencies for aws separately? As @parisni points out as well, there are more issues here? Can we summarize how we are on master and if anything needs to be done before the release. as discussed, we should have the aws bundle with some dep issue resolved https://issues.apache.org/jira/browse/HUDI-3878
gharchive/pull-request
2022-01-09T02:40:02
2025-04-01T04:33:29.008006
{ "authors": [ "boneanxs", "parisni", "vinothchandar", "xushiyan" ], "repo": "apache/hudi", "url": "https://github.com/apache/hudi/pull/4542", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1587402330
[HUDI-5812] Optimize the data size check in HoodieBaseParquetWriter Change Logs Describe context and summary for this change. Highlight if any code was copied. Finish TODO, we can actually do evaluation more accurately: if we cache last data size check, since we account for how many records were written we can accurately project avg record size, and therefore estimate how many more records we can write before cut off Impact Describe any public API or user-facing feature change or any performance impact. None Risk level (write none, low medium or high below) If medium or high, explain what verification was done to mitigate the risks. none Documentation Update Describe any necessary documentation update if there is any new feature, config, or user-facing change The config description must be updated if new configs are added or the default value of the configs are changed Any new feature or user-facing change requires updating the Hudi website. Please create a Jira ticket, attach the ticket number here and follow the instruction to make changes to the website. Contributor's checklist [ ] Read through contributor's guide [ ] Change Logs and Impact were stated clearly [ ] Adequate tests were added if applicable [ ] CI passed @alexeykudinkin @danny0405 could you please help to review this? Gentle ping @danny0405, all comments are addressed, is it good to merge now?
gharchive/pull-request
2023-02-16T10:13:31
2025-04-01T04:33:29.013566
{ "authors": [ "boneanxs" ], "repo": "apache/hudi", "url": "https://github.com/apache/hudi/pull/7978", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
791908386
IGNITE-14035: Table access API. Draft. Simple examples added. Tip: use draft PRs https://github.blog/2019-02-14-introducing-draft-pull-requests/
gharchive/pull-request
2021-01-22T10:57:39
2025-04-01T04:33:29.040934
{ "authors": [ "AMashenkov", "ptupitsyn" ], "repo": "apache/ignite-3", "url": "https://github.com/apache/ignite-3/pull/33", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
278589215
IGNITE-6565: Use long type for size and keySize in cache metrics https://issues.apache.org/jira/browse/IGNITE-6565 @akuznetsov-gridgain Sorry, but I don't understand how it connected with the task? @akuznetsov-gridgain done. I also fixed some conflicts which appeared since PR was created. Hi! Thanks for changes, they needed for compatibility with previous versions. Code now almost looks good for me, except minor note. Could you please change VisorCacheMetrics#writeExternalData in this way: @Override protected void writeExternalData(ObjectOutput out) throws IOException { ... write. ALL old metrics... out.writeLong(sizeLong); out.writeLong(keySizeLong); } and ` @Override protected void readExternalData(byte protoVer, ObjectInput in) throws IOException, ClassNotFoundException { .... read ALL old metrics .... if (protoVer > V1) { sizeLong = in.readLong(); keySizeLong = in.readLong(); } ` @akuznetsov-gridgain done. merged
gharchive/pull-request
2017-12-01T20:37:43
2025-04-01T04:33:29.044586
{ "authors": [ "SharplEr", "akuznetsov-gridgain" ], "repo": "apache/ignite", "url": "https://github.com/apache/ignite/pull/3130", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
317306409
IGNITE-7896 FilePageStore truncate now actually remove redundant partition page file. The ticket is resolved. Closing PR
gharchive/pull-request
2018-04-24T16:37:37
2025-04-01T04:33:29.045453
{ "authors": [ "dspavlov", "ivandasch" ], "repo": "apache/ignite", "url": "https://github.com/apache/ignite/pull/3910", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
387942049
[AIRFLOW-3450] Remove unnecessary sigint handler This obfuscate the actual exit signal. Make sure you have checked all steps below. Jira [x] My PR addresses the following Airflow Jira issues and references them in the PR title. For example, "[AIRFLOW-3450] My Airflow PR" https://issues.apache.org/jira/browse/AIRFLOW-3450 In case you are fixing a typo in the documentation you can prepend your commit with [AIRFLOW-3450], code changes always need a Jira issue. Description [x] Here are some details about my PR, including screenshots of any UI changes: The signal handler obfuscates the actual exit signal. Tests [x] My PR adds the following unit tests OR does not need testing for this extremely good reason: Commits [x] My commits all reference Jira issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message": Subject is separated from body by a blank line Subject is limited to 50 characters (not including Jira issue reference) Subject does not end with a period Subject uses the imperative mood ("add", not "adding") Body wraps at 72 characters Body explains "what" and "why", not "how" Documentation [x] In case of new functionality, my PR adds documentation that describes how to use it. When adding new operators/hooks/sensors, the autoclass documentation generation needs to be added. Code Quality [x] Passes flake8 @bolkedebruin Any idea why this empty sighandle is there? Codecov Report Merging #4283 into master will increase coverage by 0.02%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #4283 +/- ## ========================================= + Coverage 78.08% 78.1% +0.02% ========================================= Files 201 201 Lines 16458 16450 -8 ========================================= - Hits 12851 12849 -2 + Misses 3607 3601 -6 Impacted Files Coverage Δ airflow/bin/cli.py 65.09% <ø> (+0.5%) :arrow_up: airflow/models.py 92.29% <0%> (-0.05%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 9dce1f0...66f2e46. Read the comment docs. I'm running Docker (Cloud composer), but I don't get any explicit reason: [2018-12-05 20:20:57,138] {jobs.py:401} ERROR - Got an exception! Propagating... Traceback (most recent call last): File "/usr/local/lib/airflow/airflow/jobs.py", line 376, in helper settings.configure_orm() File "/usr/local/lib/airflow/airflow/settings.py", line 180, in configure_orm engine = create_engine(SQL_ALCHEMY_CONN, **engine_args) File "/usr/local/lib/python3.6/site-packages/sqlalchemy/engine/__init__.py", line 391, in create_engine return strategy.create(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/sqlalchemy/engine/strategies.py", line 162, in create engine = engineclass(pool, dialect, u, **engine_args) File "/usr/local/lib/python3.6/site-packages/sqlalchemy/engine/base.py", line 1772, in __init__ self.echo = echo File "/usr/local/lib/python3.6/site-packages/sqlalchemy/log.py", line 217, in __set__ instance_logger(instance, echoflag=value) File "/usr/local/lib/python3.6/site-packages/sqlalchemy/log.py", line 173, in instance_logger def instance_logger(instance, echoflag=None): File "/usr/local/lib/airflow/airflow/bin/cli.py", line 85, in sigint_handler sys.exit(0) So my initial thought was to just let the exception propagate. I think he behaviour without those signal handlers is that the “graceful” kill docker does won’t work and it will wait the 10s before sending a hard kill. Assuming airflow is pid 1 in the container. Might a different fix be to not catch SystemExit exception? @fokko in the case of Cloud Composer it is indeed the health check failing and hence a docker sends a sigterm. The actual exception which is in this case being propagated is the SystemExit exception being thrown by the sys.exit(0). It is a bit confusing, as the sigint_handler will be called on the main_thread and intermixed with other code. So the stacktrace isn't really a logical sequence of events anymore. Catching the sigterm and attempting a nice shutdown seems like a good idea, but I don't know if the current sys.exit(0) is the best implementation. Maybe setting a flag somewhere is a better option. Catching the sigint (which could be a ctrl+c) is also good idea, as otherwise the scheduler will probably not respond to ctrl+c anymore. Very good point @NielsZeilemaker Thanks for the elaboration. Setting a flag which will cause the scheduler to shutdown sounds like a better plan. In this case we should catch the SIGTERM, set a flag which will stop the scheduler loop, and shutdown the application in 30 seconds. Ref: https://pracucci.com/graceful-shutdown-of-kubernetes-pods.html If you're going to fix/mess with it, also have a look at this line: https://github.com/apache/incubator-airflow/blob/1d53f939669102cd0c8461ad9d756b3e0cf74dbe/airflow/jobs.py#L206 Doesn't seem like a good idea to mark a job as success upon a systemexit…. @NielsZeilemaker My suggestion would be to keep the PR's nice and small, so that might be something for you to pick up :-)
gharchive/pull-request
2018-12-05T20:56:23
2025-04-01T04:33:29.063968
{ "authors": [ "Fokko", "NielsZeilemaker", "ashb", "codecov-io" ], "repo": "apache/incubator-airflow", "url": "https://github.com/apache/incubator-airflow/pull/4283", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
125722608
APEXCORE-303 #resolve evaluate the launch properties at application l… …aunch @davidyan74 : Can you please add test case as well so that users know how to use this enhancement? @gauravgopi123 added unit test
gharchive/pull-request
2016-01-09T00:24:35
2025-04-01T04:33:29.065633
{ "authors": [ "davidyan74", "gauravgopi123" ], "repo": "apache/incubator-apex-core", "url": "https://github.com/apache/incubator-apex-core/pull/201", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
606663621
fixed some doc styles by markdownlint fixed some doc styles by markdownlint. related issue: #1273 Scan detected 1 error(s) in 1 file(s): [./doc/plugins/proxy-rewrite.md]: 95: line has trailing whitespace. @dayongi please take a look. too long... fixed Wen Ming notifications@github.com 于2020年4月25日周六 下午3:12写道: Scan detected 1 error(s) in 1 file(s): [./doc/plugins/proxy-rewrite.md]: 95: line has trailing whitespace. @dayongi https://github.com/dayongi please take a look. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/apache/incubator-apisix/pull/1501#issuecomment-619333940, or unsubscribe https://github.com/notifications/unsubscribe-auth/AJ62C5DCKRZG6PLMLPX4733ROKENTANCNFSM4MQRCCMQ . -- tel: 18336032131 email:tianxiaoyongcs@gmail.com address: 上海 @dayongi This PR is huge, and I recommend splitting it into smaller PRs. Each PR can contain only a few documents. This will make it easier to review and approve. @dayongi How about we close this PR first? You can then split it into multiple small PRs. ok i will do this YuanSheng Wang notifications@github.com 于2020年5月8日周五 下午1:57写道: @dayongi https://github.com/dayongi How about we close this PR first? You can then split it into multiple small PRs. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/apache/incubator-apisix/pull/1501#issuecomment-625645839, or unsubscribe https://github.com/notifications/unsubscribe-auth/AJ62C5B2WD4RHO63UFZ3I4DRQONLVANCNFSM4MQRCCMQ . -- tel: 18336032131 email:tianxiaoyongcs@gmail.com address: 上海 ok, we can close this PR first. waiting for your new small PRs.
gharchive/pull-request
2020-04-25T01:47:59
2025-04-01T04:33:29.074173
{ "authors": [ "dayongi", "juzhiyuan", "membphis", "moonming" ], "repo": "apache/incubator-apisix", "url": "https://github.com/apache/incubator-apisix/pull/1501", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1237324696
Add a plugin for Ones Description Collect data from Ones(Cloud). Pre-requisites Please make sure you have access to Ones API before applying to this issue; otherwise, it's nearly impossible to finish this task. Describe the solution you'd like Add a plugin for ones. Please refer to other data source plugins such as jira and tapd. This plugin performs 3 major tasks: Collect data from Ones API. The API response will be stored in the raw data layer. 1st ETL: transform the raw data and load it into the tool layer. 2nd ETL: transform the tool layer data and load it into the domain layer. As Ones is an issue tracking tool, the final data to be loaded into the domain layer are mainly entities from issue tracking domain How to start Please follow the doc how to add a plugin. Additional notes In DevLake, there're 3 layers of storage: raw data layer: tables starting with _raw_ tool layer: tables starting with _tool_ domain layer: tables with no prefix You can set up DevLake and see the tables in MySQL. I would like to tackle this one. Please approve @hezyin @Startrekzky and assign me to it : 最近几次Devlake会议都有进去参加学习 WeOpen Star I would like to tackle this one. Please approve @hezyin @Startrekzky and assign me to it : 最近几次Devlake会议都有进去参加学习 Thanks @yujng so much, I'm looking forward to your PRs. "Ones" only supports on-premise deployment, which means this issue is more suitable for those who already have the access to an "Ones" instance. We still need this. als thumb_up for this feature request. I would like to tackle this one. Please approve @hezyin @Startrekzky and assign me to it. I went to the recent community meeting for a couple of times. Any progress on this feature ? @hezyin @Startrekzky
gharchive/issue
2022-05-16T15:23:21
2025-04-01T04:33:29.082141
{ "authors": [ "Startrekzky", "morganchenjp", "yujng", "yumengwang03" ], "repo": "apache/incubator-devlake", "url": "https://github.com/apache/incubator-devlake/issues/1884", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
759951852
page of [TaskInstance] load data slow page of [TaskInstance] load data by interface "/dolphinscheduler/projects/stressTes/task-instance/list-paging" slow when the size of result more than 230000 Do you have any good solutions, welcome to submit PR implementation.
gharchive/issue
2020-12-09T03:09:35
2025-04-01T04:33:29.083364
{ "authors": [ "wangxj3", "zhuangchong" ], "repo": "apache/incubator-dolphinscheduler", "url": "https://github.com/apache/incubator-dolphinscheduler/issues/4184", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
957670999
Support bitmap(binary) type in hive to import doris Proposed changes Describe the big picture of your changes here to communicate to the maintainers why we should accept this pull request. If it fixes a bug or resolves a feature request, be sure to link to that issue. Types of changes What types of changes does your code introduce to Doris? Put an x in the boxes that apply [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Documentation Update (if none of the other choices apply) [ ] Code refactor (Modify the code structure, format the code, etc...) [ ] Optimization. Including functional usability improvements and performance improvements. [x] Dependency. Such as changes related to third-party components. [ ] Other. Checklist Put an x in the boxes that apply. You can also fill these out after creating the PR. If you're unsure about any of them, don't hesitate to ask. We're here to help! This is simply a reminder of what we are going to look for before merging your code. [ ] I have created an issue on (Fix #ISSUE) and described the bug/feature there in detail [ ] Compiling and unit tests pass locally with my changes [ ] I have added tests that prove my fix is effective or that my feature works [ ] If these changes need document changes, I have updated the document [ ] Any dependent changes have been merged Further comments If this is a relatively large or complex change, kick off the discussion at dev@doris.apache.org by explaining why you chose the solution you did and what alternatives you considered, etc... Please add User Manual for this feature This pr would be closed temporarily and will be re-pull request after internal test
gharchive/pull-request
2021-08-02T02:04:58
2025-04-01T04:33:29.089527
{ "authors": [ "rufeng97", "wangbo" ], "repo": "apache/incubator-doris", "url": "https://github.com/apache/incubator-doris/pull/6356", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1038175849
[Improve] Spark connector supports multiple spark versions:2.1.x/2.3.x/3.x Proposed changes Part of Spark Connector code is implemented using the "Spark DatasourcesV2" API . However , "Spark DatasourcesV2" API only exists in spark 2.3.x/2.4.x , not exists in spark 2.1.x/2.2.x/3.x. That API seems to have been removed after spark 3.x . For supporting multiple spark versions, part of the code is re-implemented using the "Spark DatasourcesV1" API. Types of changes What types of changes does your code introduce to Doris? Put an x in the boxes that apply [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Documentation Update (if none of the other choices apply) [ ] Code refactor (Modify the code structure, format the code, etc...) [ ] Optimization. Including functional usability improvements and performance improvements. [ ] Dependency. Such as changes related to third-party components. [x] Other. Checklist Put an x in the boxes that apply. You can also fill these out after creating the PR. If you're unsure about any of them, don't hesitate to ask. We're here to help! This is simply a reminder of what we are going to look for before merging your code. [ ] I have created an issue on (Fix #ISSUE) and described the bug/feature there in detail [x] Compiling and unit tests pass locally with my changes [ ] I have added tests that prove my fix is effective or that my feature works [ ] If these changes need document changes, I have updated the document [ ] Any dependent changes have been merged https://github.com/apache/incubator-doris/issues/7363
gharchive/pull-request
2021-10-28T07:29:48
2025-04-01T04:33:29.095900
{ "authors": [ "chovy-3012", "lordk911" ], "repo": "apache/incubator-doris", "url": "https://github.com/apache/incubator-doris/pull/6956", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
399084810
downtime when upgrading single node overlord+coordinator to 0.13.0 I noticed this behavior while upgrading from 0.12.3 to 0.13.0. We have a single node merged overlord+coordinator. The assumption is that since overlord/coordinator are not in the query path, upgrading that node shouldn't result in any downtime. However, it seems when the node restarted (after upgrade), it terminated all existing KIS tasks and restarted them. This tooks a few minutes to happen, and during these 5 minutes or so, no realtime data was available for querying. I saw a lot of messages like this during the node startup (after upgrade): o.a.d.i.c.IndexTaskClient [IndexTaskClient-pctile-hour-0] No TaskLocation available for task [index_kafka_pctile-hour_ec0a9bc8420bc02_ehjdgpfc], this task may not have been assigned to a worker yet or may have already completed Is this expected behavior? If so, it might be worth documenting. if KIS is being marked as stable (ref #6970) , this bug might be worth either fixing or atleast documenting.
gharchive/issue
2019-01-14T21:39:31
2025-04-01T04:33:29.098183
{ "authors": [ "pdeva" ], "repo": "apache/incubator-druid", "url": "https://github.com/apache/incubator-druid/issues/6854", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
452491144
First Time Launching Druid Tasks fail Hello, I set up an Ubuntu 19.04 Virtual Machine using VMWare Workstation. Java JRE build 11.0.3+7-Ubuntu-1ubuntu219.04.1 and Python 2.7 are installed. All starting Tasks fail, whenever I try to execute the single machine quicklaunch. Here's the log: WARNING: An illegal reflective access operation has occurred WARNING: Illegal reflective access by com.google.inject.internal.cglib.core.$ReflectUtils$1 (file:/home/myadmin/apache-druid-0.14.2-incubating/lib/guice-4.1.0.jar) to method java.lang.ClassLoader.defineClass(java.lang.String,byte[],int,int,java.security.ProtectionDomain) WARNING: Please consider reporting this to the maintainers of com.google.inject.internal.cglib.core.$ReflectUtils$1 WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations WARNING: All illegal access operations will be denied in a future release 2019-06-05T11:27:47,578 INFO [main] org.apache.druid.guice.PropertiesModule - Loading properties from common.runtime.properties 2019-06-05T11:27:47,583 INFO [main] org.apache.druid.guice.PropertiesModule - Loading properties from runtime.properties 2019-06-05T11:27:47,768 INFO [main] org.hibernate.validator.internal.util.Version - HV000001: Hibernate Validator 5.1.3.Final Exception in thread "main" java.lang.NoClassDefFoundError: javax/xml/bind/ValidationException at java.base/java.lang.ClassLoader.defineClass1(Native Method) at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:1016) at java.base/java.security.SecureClassLoader.defineClass(SecureClassLoader.java:174) at java.base/jdk.internal.loader.BuiltinClassLoader.defineClass(BuiltinClassLoader.java:802) at java.base/jdk.internal.loader.BuiltinClassLoader.findClassOnClassPathOrNull(BuiltinClassLoader.java:700) at java.base/jdk.internal.loader.BuiltinClassLoader.loadClassOrNull(BuiltinClassLoader.java:623) at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521) at org.hibernate.validator.internal.engine.ConfigurationImpl.<init>(ConfigurationImpl.java:110) at org.hibernate.validator.internal.engine.ConfigurationImpl.<init>(ConfigurationImpl.java:86) at org.hibernate.validator.HibernateValidator.createGenericConfiguration(HibernateValidator.java:41) at javax.validation.Validation$GenericBootstrapImpl.configure(Validation.java:276) at javax.validation.Validation.buildDefaultValidatorFactory(Validation.java:110) at org.apache.druid.guice.ConfigModule.configure(ConfigModule.java:39) at com.google.inject.spi.Elements$RecordingBinder.install(Elements.java:340) at com.google.inject.spi.Elements.getElements(Elements.java:110) at com.google.inject.internal.InjectorShell$Builder.build(InjectorShell.java:138) at com.google.inject.internal.InternalInjectorCreator.build(InternalInjectorCreator.java:104) at com.google.inject.Guice.createInjector(Guice.java:99) at com.google.inject.Guice.createInjector(Guice.java:73) at org.apache.druid.guice.GuiceInjectors.makeStartupInjector(GuiceInjectors.java:56) at org.apache.druid.cli.Main.main(Main.java:101) Caused by: java.lang.ClassNotFoundException: javax.xml.bind.ValidationException at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:583) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521) ... 23 more Is this a Java incompatibility? I use Druid 0.14.2 incubating and I tried Zookeeper 3.5.5. , 3.4.11 and 3.4.14 Best Regards, Calvin Hi Calvin, Afaik, Druid doesn't support Java 11; you'll have to use Java 8 (as mentioned in the quickstart doc). There's an initiative to support Java 11 compatibility (see #5589) but I don't think it's been completed. Thank you! Java 8 works.
gharchive/issue
2019-06-05T13:07:20
2025-04-01T04:33:29.102192
{ "authors": [ "CalvinSchulze", "justinborromeo" ], "repo": "apache/incubator-druid", "url": "https://github.com/apache/incubator-druid/issues/7835", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
388180919
Graceful shutdown enhancement in Spring What is the purpose of the change Graceful shutdown enhancement in Spring. AbstractConfig has added a DubboShutdownHook in its constructed function,but when we used spring, SpringExtensionFactory also added a ApplicationListener to invoke DubboShutdownHook when spring fired ContextClosedEvent. This seems to have a problem。 AbstractConfig‘s DubboShutdownHook will invoke when the application shutdown, at the same time spring boot shutDownhook will invoke to close the context。 when the application closing ,dubbo will make sure all the request has return ,if a request used bean in spring , this bean will close by context。 it will throw error,so in spring ,we should make sure dubbo shutdown before spring destory its beans。 我怕我英文没有表达清楚,用中文简单描述一下。正好也在解决这个问题,当我的应用在使用spring boot 时,spring boot 会注册一个shutdownhook 来关闭 applicationcontext ,同时 Dubbo 也会注册 shutdownhook, 但是我们没办法保证两个shutdownhook的执行顺序。因为dubbo 的优雅停机 会把已经进来的请求处理完,当这些请求依赖spring 的某些bean 时,这些bean 可能已经被spring 容器关闭了,产生了报错。我们应该确保dubbo完全停机了再关闭spring容器。 我目前的做法是取消了spring boot 注册的shutdownhook, 自己实现了一个shutdownhook ,运行时,先sleep(DEFAULT_SERVER_SHUTDOWN_TIMEOUT),再确保dubbo 停机后再关闭容器,如果 DEFAULT_SERVER_SHUTDOWN_TIMEOUT 过大的话,可能会拖延停机时间,这也是为什么我提了一个issue :https://github.com/apache/incubator-dubbo/issues/2877 。 Brief changelog when add ShutdownHookListener ,remove AbstractConfig‘s DubboShutdownHook Verifying this change XXXXX Follow this checklist to help us incorporate your contribution quickly and easily: [x] Make sure there is a GITHUB_issue field for the change (usually before you start working on it). Trivial changes like typos do not require a GITHUB issue. Your pull request should address just this issue, without pulling in other changes - one PR resolves one issue. [ ] Format the pull request title like [Dubbo-XXX] Fix UnknownException when host config not exist #XXX. Each commit in the pull request should have a meaningful subject line and body. [ ] Write a pull request description that is detailed enough to understand what the pull request does, how, and why. [ ] Write necessary unit-test to verify your logic correction, more mock a little better when cross module dependency exist. If the new feature or significant change is committed, please remember to add integration-test in test module. [ ] Run mvn clean install -DskipTests & mvn clean test-compile failsafe:integration-test to make sure unit-test and integration-test pass. [ ] If this contribution is large, please follow the Software Donation Guide. Codecov Report Merging #2901 into master will decrease coverage by 0.02%. The diff coverage is 100%. @@ Coverage Diff @@ ## master #2901 +/- ## ========================================== - Coverage 63.74% 63.71% -0.03% ========================================== Files 577 577 Lines 25930 25931 +1 Branches 4540 4540 ========================================== - Hits 16529 16522 -7 - Misses 7233 7235 +2 - Partials 2168 2174 +6 Impacted Files Coverage Δ ...ava/org/apache/dubbo/config/DubboShutdownHook.java 84% <100%> (ø) :arrow_up: ...onfig/spring/extension/SpringExtensionFactory.java 84.61% <100%> (+0.4%) :arrow_up: ...ache/dubbo/remoting/p2p/support/AbstractGroup.java 45.45% <0%> (-11.37%) :arrow_down: ...in/java/org/apache/dubbo/common/utils/JVMUtil.java 73.58% <0%> (-7.55%) :arrow_down: .../apache/dubbo/remoting/transport/AbstractPeer.java 58.69% <0%> (-4.35%) :arrow_down: ...dubbo/remoting/exchange/support/DefaultFuture.java 66.44% <0%> (-3.36%) :arrow_down: ...org/apache/dubbo/rpc/protocol/AbstractInvoker.java 62.9% <0%> (-3.23%) :arrow_down: ...dubbo/rpc/protocol/dubbo/CallbackServiceCodec.java 77.2% <0%> (-2.21%) :arrow_down: ...a/org/apache/dubbo/monitor/dubbo/DubboMonitor.java 87.85% <0%> (-1.87%) :arrow_down: ...apache/dubbo/rpc/protocol/dubbo/DubboProtocol.java 65.83% <0%> (-0.84%) :arrow_down: ... and 6 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 63bf28d...f6a3213. Read the comment docs.
gharchive/pull-request
2018-12-06T11:55:07
2025-04-01T04:33:29.122935
{ "authors": [ "LiZhenNet", "codecov-io" ], "repo": "apache/incubator-dubbo", "url": "https://github.com/apache/incubator-dubbo/pull/2901", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
499535808
toolTip的formatter属性里,用function自定义时,打印出params.seriesName不能跟配置的系列名对应 Version 4.3.0 Steps to reproduce 鼠标悬浮到地图,观察分组名,全都是第一个系列的名称 What is expected? 跟系列名对应起来 What is actually happening? 全部是第一个系列名称 Could you provide a reproducible demo? gallery.echartsjs.com 提交不了,我开了个repo作为demo:http://champyin.com/echarts-issues/ 你点点图例试试,把「分组1」隐藏了试试: 是不是都是「分组2」了? 你这两个 series 是罗在一起的地图啊~ 你点下图例试试: 把「分组1」隐藏,是不是都是「分组2」了? 你这两个 series 是两个罗在一起的中国地图啊= = 你点下图例试试: 把「分组1」隐藏,是不是都是「分组2」了? 你这两个 series 是两个罗在一起的两张中国地图啊= = 嗯,但是不使用formatter自定义的时候,是可以区分「分组1」和 「分组2」的,你可以拉一下我的demo源码,把formatter注释掉看看效果。 所以echarts应该是支持地图罗在一起的时候区分分组的。 我更新了问题复现demo:http://champyin.com/echarts-issues/ 加入了效果对比。 额,在原楼层改了好几回…… @zhouxin860114 因为我需要展示图例,所以需要使用多个series。如果像你建议的那样,我的图例没法展示了。 说得有点乱,总结一下,我估计是这么回事: 1、你的两个 map series 是重叠的,最上层的 series 是「分组1」 2、「分组1」显示时,鼠标事件被「分组1」的图形捕获,params 里的 series 是「分组1」,所以自定义的 formatter 只能显示「分组1」的属性 3、 默认的 formatter 似乎会遍历所有 series,把有数据的 series 都显示出来,并把数值相加,如下图: formatTooltip: function(dataIndex) { // FIXME orignalData and data is a bit confusing var data = this.getData(); var formattedValue = addCommas(this.getRawValue(dataIndex)); var name = data.getName(dataIndex); var seriesGroup = this.seriesGroup; var seriesNames = []; for (var i = 0; i < seriesGroup.length; i++) { var otherIndex = seriesGroup[i].originalData.indexOfName(name); var valueDim = data.mapDimension('value'); if (!isNaN(seriesGroup[i].originalData.get(valueDim, otherIndex))) { seriesNames.push( encodeHTML(seriesGroup[i].name) ); } } return seriesNames.join(', ') + '<br />' + encodeHTML(name + ' : ' + formattedValue); }, —————————————————— 图例可以自定义,第一个回答里,我后来贴上的那段代码,就是包含自定义图例的 还有一个思路,就是 formatter 的回调函数里,只通过 params.name 确定省份,然后通过省份去原始数据中查询分组信息、数值等; 这样你就可以用多个 series[i]-map 了 @zhouxin860114 分析的很有道理,按照你的方式可以实现我想要的效果,3Q~ @zhouxin860114 分析的很有道理,按照你的方式可以实现我想要的效果,3Q~ 不客气 :D
gharchive/issue
2019-09-27T16:23:24
2025-04-01T04:33:29.132437
{ "authors": [ "yc111", "yufeng04", "zhouxin860114" ], "repo": "apache/incubator-echarts", "url": "https://github.com/apache/incubator-echarts/issues/11347", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
333211508
Request: Cumulative Line Chart One-line summary The cumulative line chart doesn't seem to be included in ECharts. Request We're thinking of changing our charting library, but it will hard to do so if the new one does not support cumulative line charts. Example (NVD3) http://nvd3.org/examples/cumulativeLine.html Other comments If there is no time for this to be done by your team, could we develop it for you and have it added to the source code? Not sure if it meets your request, cumulative line chart can be created by processing the data and use tooltip.formatter and label.formatter to display the original value. Please tell me if I missed anything. Hi Pissang, Thank you for your quick reply and sorry for the delay. That's not exactly what we're looking for as we need a line (similar to the one in the example I've linked), that rebases the chart as you move it. Kind Regards, Meldanen @pissang can you please explan what do you mean?
gharchive/issue
2018-06-18T10:20:58
2025-04-01T04:33:29.136266
{ "authors": [ "Meldanen", "klukiyan", "pissang" ], "repo": "apache/incubator-echarts", "url": "https://github.com/apache/incubator-echarts/issues/8536", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
383051556
test [ ] I am using English in this issue. 在这个 Issue 中我使用了英文(强烈建议)。 General Questions PLEASE MAKE SURE OF ALL THE FOLLOWING OPTIONS IN REQUIRED FIELDS ARE TICKED (with x)! Otherwise, the issue will not be answered. And think before you tick. :) 请确保以下每项都打上勾了!并且确保都做了这些事哦~ 这将大量简化我们的工作流程,使你的问题更快速得到解答。 [ ] Required: I have read the document and examples and tried to solve it by myself. (必填)我读过了文档和教程,并且曾试图自己解决问题。 [ ] Required: I have searched for similar issues and that didn't help. (必填)我搜索过 issue 但是没有帮助。 [ ] Required: I have tried with the latest version of ECharts and still have this problem. (必填)我试过最新版本的 ECharts,还是存在这个问题。 In this issue, I have provided information with: 在这个 issue 中我提供了以下信息: [ ] Required: issue type;(必填)issue 类型 [ ] Required: one sentence description in issue details;(必填)一句话的问题描述 [ ] Required: demo;(必填)能反映问题的例子(如果你想提问或报 bug) [ ] Required: ECharts version;ECharts 版本 Issue Type [ ] I have a question to ask about how to use ECharts to ...;我想提问如何使用 ECharts 实现某功能 [ ] I have a bug to report;我想要报 bug [ ] I have a feature to request, e.g.: I'd like a new feature that ...;我需要一个新功能 [ ] I have a feature to enhance, e.g.: The current feature should be improved in the way that ...;我希望改进某个功能 [ ] There's something wrong with the documents;文档有些问题 [ ] Others, or I'm not sure which issue types to choose here;其他,或我不知道应该选什么类型 Issue Details {DESCRIBE_HERE} Expected Behavior {DESCRIBE_HERE} Current Behavior {DESCRIBE_HERE} Online Example THIS IS REQUIRED FOR ALL BUG REPORTS AND QUESTIONS!!! 如果是提问或报 bug 一定要提供这一项! {ADD_LINK_OF_DEMO_HERE} {PASTE_RELATED_CODE_HERE} Topics [ ] Legend [ ] Tooltip [ ] Event [ ] Performance [ ] SVG [ ] Map [ ] ECharts GL [ ] Third-party libraries, e.g.: Vue.js, React Anything Else We Need to Know {DESCRIBE_HERE} Environment ECharts version;ECharts 版本: {DESCRIBE_HERE} [ ] It happens only on certain browsers or operating systems. 对于特定浏览器或操作系统才会出现的问题,请提供相应环境信息:{BROWSER_VERSION_OR_OS_INFORMATION_HERE} 1
gharchive/issue
2018-11-21T10:27:55
2025-04-01T04:33:29.146495
{ "authors": [ "Ovilia" ], "repo": "apache/incubator-echarts", "url": "https://github.com/apache/incubator-echarts/issues/9450", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
387240905
tooltip不会自动消失 [x ] I am using English in this issue. 在这个 Issue 中我使用了英文(强烈建议)。 General Questions PLEASE MAKE SURE OF ALL THE FOLLOWING OPTIONS IN REQUIRED FIELDS ARE TICKED (with x)! Otherwise, the issue will not be answered. And think before you tick. :) 请确保以下每项都打上勾了!并且确保都做了这些事哦~ 这将大量简化我们的工作流程,使你的问题更快速得到解答。 [ x] Required: I have read the document and examples and tried to solve it by myself. (必填)我读过了文档和教程,并且曾试图自己解决问题。 [ x] Required: I have searched for similar issues and that didn't help. (必填)我搜索过 issue 但是没有帮助。 [x ] Required: I have tried with the latest version of ECharts and still have this problem. (必填)我试过最新版本的 ECharts,还是存在这个问题。 In this issue, I have provided information with: 在这个 issue 中我提供了以下信息: [ x] Required: issue type;(必填)issue 类型 [ x] Required: one sentence description in issue details;(必填)一句话的问题描述 [x ] Required: demo;(必填)能反映问题的例子(如果你想提问或报 bug) [x ] Required: ECharts version;ECharts 版本 Issue Type [ ] I have a question to ask about how to use ECharts to ...;我想提问如何使用 ECharts 实现某功能 [ x] I have a bug to report;我想要报 bug [ ] I have a feature to request, e.g.: I'd like a new feature that ...;我需要一个新功能 [ ] I have a feature to enhance, e.g.: The current feature should be improved in the way that ...;我希望改进某个功能 [ ] There's something wrong with the documents;文档有些问题 [ ] Others, or I'm not sure which issue types to choose here;其他,或我不知道应该选什么类型 Issue Details tooltip doesn't auto hide after pointer leave Expected Behavior tooltip auto hide Current Behavior {DESCRIBE_HERE} Online Example THIS IS REQUIRED FOR ALL BUG REPORTS AND QUESTIONS!!! 如果是提问或报 bug 一定要提供这一项! {ADD_LINK_OF_DEMO_HERE} tooltip: { trigger: 'axis', confine: true, enterable: true, backgroundColor: 'rgba(0,0,0,.8)' } Topics [ ] Legend [ x] Tooltip [ ] Event [ ] Performance [ ] SVG [ ] Map [ ] ECharts GL [ ] Third-party libraries, e.g.: Vue.js, React Anything Else We Need to Know {DESCRIBE_HERE} Environment ECharts version;ECharts 版本: 4.2.0-rc.2 [ ] It happens only on certain browsers or operating systems. 对于特定浏览器或操作系统才会出现的问题,请提供相应环境信息:{BROWSER_VERSION_OR_OS_INFORMATION_HERE} 请提供更完整的信息
gharchive/issue
2018-12-04T11:27:30
2025-04-01T04:33:29.157100
{ "authors": [ "node-rookie", "pissang" ], "repo": "apache/incubator-echarts", "url": "https://github.com/apache/incubator-echarts/issues/9522", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1569244913
[ISSUE #2973]Method passes a concatenated string to SLF4J's or Log4j2's format string[SubStreamHandler] fix #2973 @zhaoyansheng163 please solve the conflicts and the ci problems.
gharchive/pull-request
2023-02-03T05:27:40
2025-04-01T04:33:29.158166
{ "authors": [ "xwm1992", "zhaoyansheng163" ], "repo": "apache/incubator-eventmesh", "url": "https://github.com/apache/incubator-eventmesh/pull/3033", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
764754972
Allow localhost graceful shutdown #3581 Add gracefully shutdown In Apache Aurora, it is necessary to allow a graceful shutdown on localhost. Nice work.. ^^; [I 201214 14:47:11 web:1811] 200 GET /health (127.0.0.1) 0.64ms [I 201214 14:47:31 web:1811] 200 GET /health (127.0.0.1) 0.53ms [I 201214 14:47:46 killexecutorhandler:51] Received 'Killing process' request [I 201214 14:47:46 web:1811] 200 POST /quitquitquit (127.0.0.1) 1.08ms [I 201214 14:47:46 killexecutorhandler:44] Killing parent executor @thinker0 or @Code0x58 Is this PR ok to be further reviewed and merged?
gharchive/pull-request
2020-12-13T02:01:04
2025-04-01T04:33:29.159787
{ "authors": [ "joshfischer1108", "thinker0" ], "repo": "apache/incubator-heron", "url": "https://github.com/apache/incubator-heron/pull/3650", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
942382939
HOP-3013 Adapt ContextDialog position to monitor I would test tomorrow with multiple monitors, but if anyone can test on Linux / Max / WebHop It only works when you are using your primary monitor. If you work on your secondary monitor it trips (unless the resolutions of those monitors are the same) This still does not really give the desired result on mac, I'll take a look tomorrow if I can gather a bit more information on what is happening. I have a feeling it is still giving the size of the primary display. Works in Hop Web too, if you can confirm Windows it's ready for merge Works fine on Windows, but tested with only one monitor
gharchive/pull-request
2021-07-12T19:39:27
2025-04-01T04:33:29.161797
{ "authors": [ "hansva", "nadment" ], "repo": "apache/incubator-hop", "url": "https://github.com/apache/incubator-hop/pull/932", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1824206036
Setup lts patch Following update done on 1.40.x-prod https://github.com/kiegroup/kogito-images/pull/1611 Many thanks for submitting your Pull Request :heart:! Please make sure your PR meets the following requirements: [ ] You have read the contributors guide [ ] Pull Request title is properly formatted: [KOGITO|RHPAM-XYZ] Subject [ ] Pull Request contains link to the JIRA issue [ ] Pull Request contains description of the issue [ ] Pull Request does not include fixes for issues other than the main ticket [ ] Your feature/bug fix has a testcase that verifies it [ ] You've tested the new feature/bug fix in an actual OpenShift cluster [ ] You've added a RELEASE_NOTES.md entry regarding this change How to retest this PR or trigger a specific build: (Re)run Jenkins tests Please add comment: Jenkins [test|retest] this Prod tests Please add comment: Jenkins (re)run [prod|Prod|PROD] How to backport a pull request to a different branch? In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch). NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another. Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added. If something goes wrong, the author will be notified and at this point a manual backporting is needed. NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created. PR job #1019 was: FAILURE Possible explanation: Pipeline failure or project build failure Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images/1019/display/redirect See console log: Console Logs [Pipeline] }Build KIE » kogito » main » pullrequest » kogito-images.build-image PR #1653 - kogito-management-console: https://github.com/kiegroup/kogito-images/pull/1653 completed: UNSTABLE[Pipeline] unstableWARNING: Tests on kogito-management-console seems to have failed[Pipeline] }[Pipeline] // stage[Pipeline] }Build KIE » kogito » main » pullrequest » kogito-images.build-image PR #1653 - kogito-task-console: https://github.com/kiegroup/kogito-images/pull/1653 completed: SUCCESS[Pipeline] }[Pipeline] // stage[Pipeline] }Build KIE » kogito » main » pullrequest » kogito-images.build-image PR #1653 - kogito-s2i-builder: https://github.com/kiegroup/kogito-images/pull/1653 completed: UNSTABLE[Pipeline] unstableWARNING: Tests on kogito-s2i-builder seems to have failed[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // parallel[Pipeline] }[Pipeline] // script[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] echo[INFO] Clean workspace[Pipeline] cleanWs[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] Deferred wipeout is used...[WS-CLEANUP] done[Pipeline] echo[INFO] Workspace cleaned[Pipeline] echo[INFO] Cleanup Maven artifacts[Pipeline] sh+ rm -rf /home/jenkins/.m2/repository[Pipeline] echo[INFO] .m2/repository cleaned[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images/1019/api/json[Pipeline] readJSON[Pipeline] sh jenkins retest this Hey @radtriste the patch itself looks good to me but I fear it could hard to maintain in this way, I mean everytime we change QUARKUS_EXTENSIONS here modules/kogito-swf/devmode/build-config/module.yaml we should always update the patch. So why don't we fully specify all extensions? i.e., <group-id>:<artifact-id>:${KOGITO_VERSION}, is there any issue using this approach on community? if this works it will completely remove the need of having the patch Hey @radtriste the patch itself looks good to me but I fear it could be hard to maintain in this way, I mean everytime we change QUARKUS_EXTENSIONS here modules/kogito-swf/devmode/build-config/module.yaml we should always update the patch. So why don't we fully specify all extensions? i.e., <group-id>:<artifact-id>:${KOGITO_VERSION}, is there any issue using this approach on community? if this works it will completely remove the need of having the patch This is exactly what I propose to Filippe here: https://github.com/kiegroup/kogito-images/pull/1653#discussion_r1276200853 :) This is exactly what I propose to Filippe here: #1653 (comment) :) Ouch sorry, I completely missed you comment!! Then, I definitely agree on your proposal :) @lampajr hi, does this PR still valid? Hey @spolti , I think it depends on what you think about https://github.com/apache/incubator-kie-kogito-images/pull/1653#discussion_r1276200853. Because if you agree to fully specify all extensions in terms of <group-id>:<artifact-id>:${KOGITO_VERSION} I think we do NOT need this patch (therefore the PR can be closed) but we will need to create different PR to fully specify all extensions there. @lampajr yeah, I agree with that, the full GAV is better than a patch. feel free to merge :) Is the CI working for the kogito-images already? @lampajr yeah, I agree with that, the full GAV is better than a patch. feel free to merge :) Ok so we should create a different PR specifying the full GAV for all extensions and then we can close this one as no patch will be needed at that point. Is the CI working for the kogito-images already? TBH I do not know IIRC the issue was just when using Quarkus LTS
gharchive/pull-request
2023-07-27T11:50:06
2025-04-01T04:33:29.178902
{ "authors": [ "kie-ci1", "lampajr", "radtriste", "spolti" ], "repo": "apache/incubator-kie-kogito-images", "url": "https://github.com/apache/incubator-kie-kogito-images/pull/1653", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2138409762
kie-issues#777: Allow restricting jenkins agent labels for pipelines part of https://github.com/apache/incubator-kie-issues/issues/777 PR job #92 was: FAILURE Possible explanation: Pipeline failure or project build failure Please look here: https://ci-builds.apache.org/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images.build-and-test/92/display/redirect See console log: Console Logs Failed in branch kogito-base-builderBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1282 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stageBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1284 completed: FAILURE[Pipeline] }Failed in branch kogito-jobs-service-postgresql[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-jobs-service-ephemeralBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1285 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-swf-builderBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1273 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-jobs-service-allinoneBuild KIE » kogito » main » pullrequest » kogito-images.build-image #1281 completed: FAILURE[Pipeline] error[Pipeline] }[Pipeline] // stage[Pipeline] }Failed in branch kogito-swf-devmodeBuild KIE » kogito » main » pullrequest » kogito-images.build-image PR #1742 - kogito-data-index-ephemeral: https://github.com/apache/incubator-kie-kogito-images/pull/1742 completed: SUCCESS[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // parallel[Pipeline] }[Pipeline] // dir[Pipeline] }[Pipeline] // script[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - 'https://ci-builds.apache.org/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-images.build-and-test/92/api/json?depth=0'[Pipeline] readJSON[Pipeline] sh
gharchive/pull-request
2024-02-16T11:28:13
2025-04-01T04:33:29.183888
{ "authors": [ "cimbalek", "kie-ci3" ], "repo": "apache/incubator-kie-kogito-images", "url": "https://github.com/apache/incubator-kie-kogito-images/pull/1742", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1416402811
Flaky Test TestSlotMigrateFromSlave Search before asking [X] I had searched in the issues and found no similar issues. Motivation --- FAIL: TestSlotMigrateFromSlave (2.36s) server.go:106: Error Trace: /home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/server.go:106 /home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/server.go:112 /home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/server.go:101 /home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:47 /home/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:65 Error: Received unexpected error: signal: aborted (core dumped) Test: TestSlotMigrateFromSlave FAIL Solution Perhaps we should upload server output on test failures? cc @PragmaTwice Are you willing to submit a PR? [ ] I'm willing to submit a PR! https://github.com/apache/incubator-kvrocks/actions/runs/3288512905/jobs/5418990199 Looks crash after sending the terminate signal, but have no idea to reproduce it. https://github.com/apache/incubator-kvrocks/actions/runs/3314140570/jobs/5474097740 Another flaky case @git-hulk they're somehow different issues - not the exact failure cause, or not the exact test case: --- FAIL: TestSlotMigrateDataType (85.48s) --- FAIL: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL (60.85s) slotmigrate_test.go:703: Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:703 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:625 Error: Condition never satisfied Test: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL FAIL exit status 1 --- FAIL: TestPubSub (1.20s) server.go:106: Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:106 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:112 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:101 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/pubsub_test.go:369 Error: Received unexpected error: signal: abort trap Test: TestPubSub FAIL exit status 1 @git-hulk they're somehow different issues - not the exact failure cause, or not the exact test case: --- FAIL: TestSlotMigrateDataType (85.48s) --- FAIL: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL (60.85s) slotmigrate_test.go:703: Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:703 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/integration/slotmigrate/slotmigrate_test.go:625 Error: Condition never satisfied Test: TestSlotMigrateDataType/MIGRATE_-_Migrate_incremental_data_via_parsing_and_filtering_data_in_WAL FAIL exit status 1 --- FAIL: TestPubSub (1.20s) server.go:106: Error Trace: /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:106 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:112 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/server.go:101 /Users/runner/work/incubator-kvrocks/incubator-kvrocks/tests/gocase/unit/pubsub/pubsub_test.go:369 Error: Received unexpected error: signal: abort trap Test: TestPubSub FAIL exit status 1 Yes, for the TestPusSub test case, it should be the same with TestSlotMigrateFromSlave, there're core when exiting the server. @git-hulk I'm thinking of uploading the log files if there's a test failure...Don't have a concrete plan yet, though. I'm also thinking how to download the log file if it cored
gharchive/issue
2022-10-20T10:51:12
2025-04-01T04:33:29.190564
{ "authors": [ "git-hulk", "tisonkun" ], "repo": "apache/incubator-kvrocks", "url": "https://github.com/apache/incubator-kvrocks/issues/1016", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1512864489
error command for: georadius Search before asking [X] I had searched in the issues and found no similar issues. Version latest Minimal reproduce step example at : https://redis.io/commands/georadius/ redis> GEOADD Sicily 13.361389 38.115556 "Palermo" 15.087269 37.502669 "Catania" (integer) 2 What did you expect to see? redis> GEORADIUS Sicily 15 37 200 km WITHDIST WITHCOORD "Catania" "56.4413" What did you see instead? Expected :56.4413 Actual :3.479447370796909E15 Anything Else? No response Are you willing to submit a PR? [ ] I'm willing to submit a PR! @lt1946 Thanks for your report, but I can't reproduce the bug with your steps.
gharchive/issue
2022-12-28T14:43:30
2025-04-01T04:33:29.194792
{ "authors": [ "git-hulk", "lt1946" ], "repo": "apache/incubator-kvrocks", "url": "https://github.com/apache/incubator-kvrocks/issues/1213", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1091051758
Turn up maven options in mvn script Why are the changes needed? How was this patch tested? [ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible [ ] Add screenshots for manual tests if appropriate [ ] Run test locally before make a pull request Codecov Report Merging #1662 (4d31098) into master (8e9e8c3) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #1662 +/- ## ========================================= Coverage 59.57% 59.57% Complexity 193 193 ========================================= Files 265 265 Lines 13084 13084 Branches 1649 1649 ========================================= Hits 7795 7795 Misses 4631 4631 Partials 658 658 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 8e9e8c3...4d31098. Read the comment docs. thanks, merged to master
gharchive/pull-request
2021-12-30T11:40:13
2025-04-01T04:33:29.203196
{ "authors": [ "codecov-commenter", "yaooqinn" ], "repo": "apache/incubator-kyuubi", "url": "https://github.com/apache/incubator-kyuubi/pull/1662", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1215790424
[KYUUBI #2484] Add conf to SessionEvent and display it in EngineSessionPage Why are the changes needed? close #2484 How was this patch tested? [ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible [X] Add screenshots for manual tests if appropriate [ ] Run test locally before make a pull request Codecov Report Merging #2485 (5c3bb30) into master (6187321) will decrease coverage by 0.07%. The diff coverage is 7.69%. @@ Coverage Diff @@ ## master #2485 +/- ## ============================================ - Coverage 63.45% 63.37% -0.08% Complexity 69 69 ============================================ Files 371 371 Lines 17585 17608 +23 Branches 2348 2348 ============================================ + Hits 11158 11159 +1 - Misses 5392 5416 +24 + Partials 1035 1033 -2 Impacted Files Coverage Δ .../scala/org/apache/spark/ui/EngineSessionPage.scala 2.15% <0.00%> (-0.75%) :arrow_down: ...ache/kyuubi/engine/spark/events/SessionEvent.scala 84.00% <100.00%> (+1.39%) :arrow_up: ...rg/apache/kyuubi/engine/spark/SparkSQLEngine.scala 77.38% <0.00%> (+0.77%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 6187321...5c3bb30. Read the comment docs. need more unit tests Thanks, I'll add it later nice feature, thanks thanks, merged to master and branch-1.5
gharchive/pull-request
2022-04-26T11:07:47
2025-04-01T04:33:29.215566
{ "authors": [ "codecov-commenter", "turboFei", "wForget" ], "repo": "apache/incubator-kyuubi", "url": "https://github.com/apache/incubator-kyuubi/pull/2485", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1249179643
[DOC] Replace sphinx_rtd_theme with sphinx_book_theme Why are the changes needed? make documentation style to fit official web site How was this patch tested? [ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible [x] Add screenshots for manual tests if appropriate [x] Run test locally before make a pull request Codecov Report Merging #2751 (b812352) into master (b40bcbd) will decrease coverage by 2.49%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #2751 +/- ## ============================================ - Coverage 64.24% 61.74% -2.50% - Complexity 82 385 +303 ============================================ Files 385 433 +48 Lines 18673 20345 +1672 Branches 2531 2760 +229 ============================================ + Hits 11997 12563 +566 - Misses 5532 6544 +1012 - Partials 1144 1238 +94 Impacted Files Coverage Δ ...pache/kyuubi/sql/KyuubiSQLExtensionException.scala 0.00% <0.00%> (-100.00%) :arrow_down: ...rg/apache/kyuubi/sql/zorder/ZorderBytesUtils.scala 0.00% <0.00%> (-94.74%) :arrow_down: ...ubi/sql/zorder/InsertZorderBeforeWritingBase.scala 0.00% <0.00%> (-89.16%) :arrow_down: ...cala/org/apache/kyuubi/sql/zorder/ZorderBase.scala 0.00% <0.00%> (-87.50%) :arrow_down: .../kyuubi/sql/zorder/OptimizeZorderCommandBase.scala 0.00% <0.00%> (-81.82%) :arrow_down: ...yuubi/sql/zorder/OptimizeZorderStatementBase.scala 0.00% <0.00%> (-71.43%) :arrow_down: ...g/apache/kyuubi/sql/zorder/ResolveZorderBase.scala 0.00% <0.00%> (-69.57%) :arrow_down: ...n/spark/authz/ranger/FilteredShowObjectsExec.scala 42.10% <0.00%> (-37.90%) :arrow_down: ...uubi/engine/flink/operation/ExecuteStatement.scala 47.61% <0.00%> (-36.80%) :arrow_down: ...pache/kyuubi/sql/KyuubiQueryStagePreparation.scala 0.00% <0.00%> (-32.36%) :arrow_down: ... and 112 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update b40bcbd...b812352. Read the comment docs. +1 looks more structured! Thanks, merging to master
gharchive/pull-request
2022-05-26T06:58:19
2025-04-01T04:33:29.235785
{ "authors": [ "codecov-commenter", "pan3793", "yanghua", "yaooqinn" ], "repo": "apache/incubator-kyuubi", "url": "https://github.com/apache/incubator-kyuubi/pull/2751", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1484571320
Print the simple batch report if no log fetch to prevent misunderstand Why are the changes needed? With yarn-cluster mode, if spark.yarn.submit.waitAppCompletion is false and kyuubi-ctl wait completion is true, we need still print some simple batch report to prevent misunderstand. The batch report format refer spark-submit log. Application report for <ApplicationID> (state: RUNNING) How was this patch tested? [ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible [ ] Add screenshots for manual tests if appropriate [x] Run test locally before make a pull request Codecov Report Merging #3949 (9580277) into master (8305d80) will decrease coverage by 0.08%. The diff coverage is 0.00%. :exclamation: Current head 9580277 differs from pull request most recent head ea5597f. Consider uploading reports for the commit ea5597f to get more accurate results @@ Coverage Diff @@ ## master #3949 +/- ## ============================================ - Coverage 51.90% 51.82% -0.09% Complexity 13 13 ============================================ Files 521 521 Lines 28843 28807 -36 Branches 3864 3856 -8 ============================================ - Hits 14972 14930 -42 - Misses 12494 12505 +11 + Partials 1377 1372 -5 Impacted Files Coverage Δ ...rg/apache/kyuubi/ctl/cmd/log/LogBatchCommand.scala 60.93% <0.00%> (-3.00%) :arrow_down: ...g/apache/kyuubi/session/KyuubiSessionManager.scala 88.13% <0.00%> (-5.15%) :arrow_down: ...g/apache/kyuubi/operation/BatchJobSubmission.scala 74.64% <0.00%> (-3.79%) :arrow_down: ...apache/kyuubi/engine/JpsApplicationOperation.scala 77.41% <0.00%> (-3.23%) :arrow_down: .../apache/kyuubi/server/api/v1/BatchesResource.scala 69.00% <0.00%> (-3.00%) :arrow_down: ...ache/kyuubi/server/KyuubiRestFrontendService.scala 82.95% <0.00%> (-0.75%) :arrow_down: ...he/kyuubi/ha/client/etcd/EtcdDiscoveryClient.scala 68.50% <0.00%> (-0.56%) :arrow_down: ...apache/kyuubi/session/KyuubiBatchSessionImpl.scala 90.66% <0.00%> (-0.13%) :arrow_down: ...cala/org/apache/kyuubi/server/ui/JettyServer.scala 86.95% <0.00%> (ø) ...in/scala/org/apache/kyuubi/config/KyuubiConf.scala 97.50% <0.00%> (+0.06%) :arrow_up: ... and 4 more :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more thanks, merged to master and 1.6.2
gharchive/pull-request
2022-12-08T12:49:32
2025-04-01T04:33:29.253797
{ "authors": [ "codecov-commenter", "turboFei" ], "repo": "apache/incubator-kyuubi", "url": "https://github.com/apache/incubator-kyuubi/pull/3949", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1464545624
update how-to-email.md email-reply.png The path is wrongly written and submitted LGTM
gharchive/pull-request
2022-11-25T13:22:43
2025-04-01T04:33:29.255669
{ "authors": [ "GuoPhilipse", "binbinCheng" ], "repo": "apache/incubator-linkis-website", "url": "https://github.com/apache/incubator-linkis-website/pull/610", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1345746316
[Question] Before asking [X] I have read the How to ask for help (我已阅读如何寻求帮助指引文档). Your environment Linkis version used: 1.1.2 Environment name and version: hdp-3.1.0 hive-3.1.0 spark-2.4.3 scala-2.11.12 jdk 1.8.0_161-b14 .... Describe your questions Q1. ... 在dss中使用hive建表语法 create table test_78 as select * from *.; 时候报错,在后台用hive 登录执行以上语句正常执行。 Eureka service list eg: Some logs info or acctch file linkis-xxx.log: ERROR [Linkis-Default-Scheduler-Thread-19] org.apache.hadoop.hive.ql.Driver 1247 printError - FAILED: Error in acquiring locks: calling recordValidTxn() more than once in the same txnid:705363368 java.lang.IllegalStateException: calling recordValidTxn() more than once in the same txnid:705363368 at org.apache.hadoop.hive.ql.Driver.acquireLocks(Driver.java:1614) at org.apache.hadoop.hive.ql.Driver.lockAndRespond(Driver.java:1795) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1965) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.linkis.engineplugin.hive.executor.HiveDriverProxy.run(HiveEngineConnExecutor.scala:523) at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor.org$apache$linkis$engineplugin$hive$executor$HiveEngineConnExecutor$$executeHQL(HiveEngineConnExecutor.scala:195) at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor$$anon$1.run(HiveEngineConnExecutor.scala:146) at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor$$anon$1.run(HiveEngineConnExecutor.scala:139) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) at org.apache.linkis.engineplugin.hive.executor.HiveEngineConnExecutor.executeLine(HiveEngineConnExecutor.scala:139) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10$$anonfun$apply$11.apply(ComputationExecutor.scala:183) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10$$anonfun$apply$11.apply(ComputationExecutor.scala:182) at org.apache.linkis.common.utils.Utils$.tryCatch(Utils.scala:40) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10.apply(ComputationExecutor.scala:184) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2$$anonfun$apply$10.apply(ComputationExecutor.scala:178) at scala.collection.immutable.Range.foreach(Range.scala:160) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2.apply(ComputationExecutor.scala:177) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$toExecuteTask$2.apply(ComputationExecutor.scala:153) at org.apache.linkis.common.utils.Utils$.tryFinally(Utils.scala:61) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor.toExecuteTask(ComputationExecutor.scala:230) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$3.apply(ComputationExecutor.scala:245) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor$$anonfun$3.apply(ComputationExecutor.scala:245) at org.apache.linkis.common.utils.Utils$.tryFinally(Utils.scala:61) at org.apache.linkis.engineconn.acessible.executor.entity.AccessibleExecutor.ensureIdle(AccessibleExecutor.scala:55) at org.apache.linkis.engineconn.acessible.executor.entity.AccessibleExecutor.ensureIdle(AccessibleExecutor.scala:49) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor.ensureOp(ComputationExecutor.scala:137) at org.apache.linkis.engineconn.computation.executor.execute.ComputationExecutor.execute(ComputationExecutor.scala:244) at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl.org$apache$linkis$engineconn$computation$executor$service$TaskExecutionServiceImpl$$executeTask(TaskExecutionServiceImpl.scala:296) at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2$$anonfun$run$2.apply$mcV$sp(TaskExecutionServiceImpl.scala:229) at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2$$anonfun$run$2.apply(TaskExecutionServiceImpl.scala:227) at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2$$anonfun$run$2.apply(TaskExecutionServiceImpl.scala:227) at org.apache.linkis.common.utils.Utils$.tryCatch(Utils.scala:40) at org.apache.linkis.common.utils.Utils$.tryAndWarn(Utils.scala:69) at org.apache.linkis.engineconn.computation.executor.service.TaskExecutionServiceImpl$$anon$2.run(TaskExecutionServiceImpl.scala:227) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) log file: linkis-xxx.log err.log Does it happen every time? Is there any useful information I can get from here? https://www.mail-archive.com/issues@hive.apache.org/msg116947.html
gharchive/issue
2022-08-22T03:13:53
2025-04-01T04:33:29.264932
{ "authors": [ "casionone", "mowenfuhua123" ], "repo": "apache/incubator-linkis", "url": "https://github.com/apache/incubator-linkis/issues/2918", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
474848922
DMLC link removal Replaced v1.5.x Julia page with the content on master. Also, replaced dmlc.ml links with available content. @mxnet-label-bot Add [Doc] Preview: http://100.24.115.165/api/julia/index.html @aaronmarkham @marcoabreu Sorry about rushing that through without addressing your questions. We needed to be rid of the malware links ASAP. @IvyBazan Can you answer? We spoke about this offline and you had some reasoning why it might be ok to just kill the links with a #. @marcoabreu I removed those links because I couldn't find any replacement for the content that was hosted on dmlc.ml. It seems better to temporarily have replaced the links with a #, while waiting for that content to be recovered and rehosted, rather than having malware links still active on the site. The current experience is that the links redirect to the current page. Issues related to missing content: Missing nin content-https://github.com/apache/incubator-mxnet/issues/15729#issuecomment-518131446 Missing nnvm blog: https://github.com/apache/incubator-mxnet/issues/15730#issuecomment-518131577
gharchive/pull-request
2019-07-30T22:23:47
2025-04-01T04:33:29.269738
{ "authors": [ "IvyBazan", "aaronmarkham", "piyushghai" ], "repo": "apache/incubator-mxnet", "url": "https://github.com/apache/incubator-mxnet/pull/15708", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
607241504
Netdb multiple dns servers Summary PR adds capability to add multiple nameservers on run-time even when not using resolv.conf and reset the list of nameservers back to default setting (no nameserver at all or single predefined nameserver in Kconfig). This is useful for applications that change their cellular network frequently. Also make getaddrinfo re-entrant when doing service name query Impact Testing @juniskane there are two minor nxstyle, please fix them, thanks. @juniskane , since the 2nd patch generate the nxstyle issues, it's better to merge the 3rd patch into the 2nd one: git rebase --interactive change pick to squach for the 3rd patch git push -f origin netdb_multiple_dns_servers_pr
gharchive/pull-request
2020-04-27T05:17:36
2025-04-01T04:33:29.272283
{ "authors": [ "juniskane", "xiaoxiang781216" ], "repo": "apache/incubator-nuttx", "url": "https://github.com/apache/incubator-nuttx/pull/892", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
611887034
sched/semaphore/sem_holder.c: Remove DEBUGPANIC Summary There is a DEBUGPANIC in some logic. This happens if a a task exists at certain points with priority inheritance enabled. This event was not expected in the original design (although logic was provided to support it). Since, apparently, it does happen, the DEBUGPANIC must be removed. Impact Should eliminate a crash found by Brennan Ashton using the X86_64 Qemu port Testing This makes sense to me. Should we also remove the first line of the assert in nxsem_restorebaseprio that was disabled and re-enable it. I think this explains the behavior. I think the second half of the assert is correct. This makes sense to me. Should we also remove the first line of the assert in nxsem_restorebaseprio that was disabled and re-enable it. I think this explains the behavior. I think the second half of the assert is correct. I removed the code because I don't like commented out code. It just leaves questions for the next person who deals with the code (and no one to answer the questions). It is always better to be courageous and remove wrong code. I did re-enable the debug output because I think it still has valid information. I dropped from the error to a warning. I think the output has value if you are debugging with scheduler debug turned off.. although perhaps. It is not a error any more. A warning is for unusual events that you might want to know about while debugging. info() is just chit-char, flow of control stuff. So generally, my preference would be to leave the warning in place because it provides information for people who are silly enough to use scheduler debug output. I really doesn't work. If you enable the scheduler debug output, the system becomes very unstable. You can't really use the OS to monitor itself in the that way. syslog output also requires the scheduler. This makes sense to me. Should we also remove the first line of the assert in nxsem_restorebaseprio that was disabled and re-enable it. I think this explains the behavior. I think the second half of the assert is correct. I removed the code because I don't like commented out code. It just leaves questions for the next person who deals with the code (and no one to answer the questions). It is always better to be courageous and remove wrong code. I did re-enable the debug output because I think it still has valid information. I dropped from the error to a warning. I think the output has value if you are debugging with scheduler debug turned off.. although perhaps. It is not a error any more. A warning is for unusual events that you might want to know about while debugging. info() is just chit-char, flow of control stuff. So generally, my preference would be to leave the warning in place because it provides information for people who are silly enough to use scheduler debug output. I really doesn't work. If you enable the scheduler debug output, the system becomes very unstable. You can't really use the OS to monitor itself in the that way. syslog output also requires the scheduler. These are bogus build failures. The macOS builds fail a large percentage of the time when there is no problem with the changes. I have gotten so that I ignore these tests because the are not reliable indicators of anything. I will restart the checks. maybe we will be lucky next time.
gharchive/pull-request
2020-05-04T13:45:24
2025-04-01T04:33:29.278392
{ "authors": [ "btashton", "patacongo" ], "repo": "apache/incubator-nuttx", "url": "https://github.com/apache/incubator-nuttx/pull/967", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1989952643
RFC: Concurrent Stat in List Description relate: #3097 Inviting more committers to review the RFC.
gharchive/pull-request
2023-11-13T06:08:19
2025-04-01T04:33:29.279559
{ "authors": [ "Xuanwo", "morristai" ], "repo": "apache/incubator-opendal", "url": "https://github.com/apache/incubator-opendal/pull/3574", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
334605673
Allow for activation store to accept user and request information Other ActivationStore implementations, Elasticsearch for instance, will need user and request information. Description Related issue and scope [ ] I opened an issue to propose and discuss this change (#????) My changes affect the following components [ ] API [ ] Controller [ ] Message Bus (e.g., Kafka) [ ] Loadbalancer [ ] Invoker [ ] Intrinsic actions (e.g., sequences, conductors) [ ] Data stores (e.g., CouchDB) [ ] Tests [ ] Deployment [ ] CLI [ ] General tooling [ ] Documentation Types of changes [ ] Bug fix (generally a non-breaking change which closes an issue). [ ] Enhancement or new feature (adds new functionality). [ ] Breaking change (a bug fix or enhancement which changes existing behavior). Checklist: [x] I signed an Apache CLA. [x] I reviewed the style guides and followed the recommendations (Travis CI will check :). [ ] I added tests to cover my changes. [ ] My changes require further changes to the documentation. [ ] I updated the documentation where necessary. PG4 1863 ⏳ Codecov Report Merging #3798 into master will increase coverage by 0.01%. The diff coverage is 100%. @@ Coverage Diff @@ ## master #3798 +/- ## ========================================== + Coverage 75.15% 75.16% +0.01% ========================================== Files 132 132 Lines 6142 6145 +3 Branches 373 380 +7 ========================================== + Hits 4616 4619 +3 Misses 1526 1526 Impacted Files Coverage Δ ...la/whisk/core/entity/ArtifactActivationStore.scala 92.85% <ø> (ø) :arrow_up: ...main/scala/whisk/core/controller/Activations.scala 96.77% <100%> (+0.16%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 92a64c2...0140629. Read the comment docs. @markusthoemmes, very concerned about this polling behavior. https://github.com/apache/incubator-openwhisk/blob/1d79fa395de32c6a774efeb4f53563689e9cbb7e/core/controller/src/main/scala/whisk/core/controller/actions/PrimitiveActions.scala#L585 @dubee concerned in which way? I believe this is the database fallback polling in case an active ack goes missing. @dubee Can you share some more details on how Elasticsearch implementation would make use of user and request data. Would it like to use any aspect of request or we can extract explicit parts from request and pass it on. For now it appears that passing request param is making SPI usage more coupled to web requests @chetanmeh, certain Elasticsearch deployments require additional authentication which can be passed through the controller via HTTP headers. This allows the controller to communicate to Elasitcsearch for users making requests. Currently our Elasticsearch log store works in the same way. When required headers are specified via the OpenWhisk Ansible configuration for the log store, the log store will pass those headers from the client to Elasticsearch to perform a query. See the log store for example: https://github.com/apache/incubator-openwhisk/blob/92a64c291156a2cd3d6b304babc2a193a46d0699/common/scala/src/main/scala/whisk/core/containerpool/logging/ElasticSearchLogStore.scala#L103 @dubee Now its more clear. So we pass on some sort of "user context" to ES. May be we make that notion explicit via having a UserContext as parameter type case class UserContext(user:Option[Identity], request: Option[HttpRequest]) - def get(activationId: ActivationId, user: Option[Identity] = None, request: Option[HttpRequest] = None)( - def get(activationId: ActivationId, userContext: Option[UserContext] = None)( PG2 3475 🔵 @markusthoemmes, anything else on this one? @chetanmeh, where do you think the UserContext case class should live? Added UserContext in the third commit. The case class definition should probably be moved elsewhere though. PG4 2057 🔵 where do you think the UserContext case class should live Adding it to ActivationStore looks fine @markusthoemmes, any other comments on this one? @mdeuser, can you review this one?
gharchive/pull-request
2018-06-21T18:23:38
2025-04-01T04:33:29.298428
{ "authors": [ "chetanmeh", "codecov-io", "dubee", "markusthoemmes" ], "repo": "apache/incubator-openwhisk", "url": "https://github.com/apache/incubator-openwhisk/pull/3798", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2145734474
[Spark]Spark sql support get/list function Spark sql support get/list function CC @YannByron LGTM. @waywtdcc please rebase master and fix the spotless style. LGTM. @waywtdcc please rebase master and fix the spotless style. ok +1
gharchive/pull-request
2024-02-21T03:29:24
2025-04-01T04:33:29.300844
{ "authors": [ "JingsongLi", "YannByron", "waywtdcc" ], "repo": "apache/incubator-paimon", "url": "https://github.com/apache/incubator-paimon/pull/2880", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
811591448
Controller Crashing - repetition of same config causing multi value setting I have following settings in controller config where validation.frequencyInSeconds is accidentally repeated. This is causing controller to crash as the value is derived as multi value. it will be better to take either last defined value or avoid this crash. controller.realtime.segment.validation.frequencyInSeconds=300 controller.realtime.segment.validation.frequencyInSeconds=900 controller.broker.resource.validation.frequencyInSeconds=900 Error: 2021/02/18 14:46:44.389 ERROR [StartServiceManagerCommand] [main] Failed to start a Pinot [CONTROLLER] at 39.246 since launch java.lang.NumberFormatException: For input string: "[300, 900]" at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) ~[?:1.8.0_282] at java.lang.Integer.parseInt(Integer.java:580) ~[?:1.8.0_282] I feel we should try to validate those configures and provide more meaningful error msgs, e.g. the config controller.realtime.segment.validation.frequencyInSeconds is invalid.
gharchive/issue
2021-02-19T00:52:39
2025-04-01T04:33:29.303798
{ "authors": [ "fx19880617", "pabrahamusa" ], "repo": "apache/incubator-pinot", "url": "https://github.com/apache/incubator-pinot/issues/6595", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
624126880
Add UDFs for String Transformation Functions Added in Function Registry Generic Transform Function modified to handle more types Missing: The return type of Generic Transform function is still restricted to String. This is great! Could you add tests in: CalciteSqlCompilerTest to check sql parsing logic. @fx19880617 @siddharthteotia Should I add tests in CalciteSQL for all the functions? @fx19880617 @siddharthteotia Should I add tests in CalciteSQL for all the functions? @KKcorps , sorry missed seeing this. Yes, the query compilation tests should be in CalciteSqlCompilerTest. Here we can verify that PinotQuery is being built correctly and that gets converted to BrokerRequest correctly. Most other tests in this file do this. The other suggestion was to also add unit tests for exercising end-to-end execution path. Please consider adding these tests to an appropriate file in /incubator-pinot/pinot-core/src/test/java/org/apache/pinot/queries/. May be TransformQueriesTest Is the plan to use this wrapper solely for invoking scalar functions (like already done in this PR for StringFunctions) or are we expecting follow-ups to integrate it with rest of the transform functions. I think only the former? @sidd only for scalarfunctions.
gharchive/pull-request
2020-05-25T08:12:52
2025-04-01T04:33:29.307502
{ "authors": [ "KKcorps", "fx19880617", "kishoreg", "siddharthteotia" ], "repo": "apache/incubator-pinot", "url": "https://github.com/apache/incubator-pinot/pull/5440", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
311675274
Rest endpoint to query compaction status Currently returns whether compaction has not run, is running, is complete or has failed. If it has failed, the last error is returned also. retest this please // failing on the mock issue retest this please // flake in c++ AuthPluginTest.testAthenz retest this please // PersistentQueueE2ETest.testCancelReadRequestOnLastDisconnect retest this please // many failures this time org.apache.pulsar.broker.service.PersistentFailoverE2ETest.testSimpleConsumerEventsWithoutPartition org.apache.pulsar.broker.service.ReplicatorTest.testCloseReplicatorStartProducer org.apache.pulsar.client.impl.BrokerClientIntegrationTest.testUnsupportedBatchMessageConsumer
gharchive/pull-request
2018-04-05T16:10:04
2025-04-01T04:33:29.310350
{ "authors": [ "ivankelly" ], "repo": "apache/incubator-pulsar", "url": "https://github.com/apache/incubator-pulsar/pull/1501", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
721558295
RATIS-1097. DataStreamOutputImpl should use streamOffset instead of messageId See https://issues.apache.org/jira/browse/RATIS-1097 @runzhiwang , thanks for reviewing the change. I just have addressed your comments. Could you review it again? @szetszwo LGTM, +1. Just wait the fix of checkstyle. Reopen PR to trigger CI. @szetszwo The failed testAsyncConfiguration seems related to this PR. Oops, I have accidentally changed OrderedAsync.java. It should be OrderedStreamAsync.java @szetszwo Thanks the patch, @amaliujia Thanks for review. I have merged it.
gharchive/pull-request
2020-10-14T15:25:18
2025-04-01T04:33:29.313246
{ "authors": [ "runzhiwang", "szetszwo" ], "repo": "apache/incubator-ratis", "url": "https://github.com/apache/incubator-ratis/pull/224", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
577524266
Decouple sharding-proxy-transport's dependencies For #4594. Pull Request Test Coverage Report for Build 10119 2 of 2 (100.0%) changed or added relevant lines in 1 file are covered. 3 unchanged lines in 1 file lost coverage. Overall coverage decreased (-0.005%) to 59.047% Files with Coverage Reduction New Missed Lines % sharding-orchestration/sharding-orchestration-core/src/main/java/org/apache/shardingsphere/orchestration/internal/util/IpUtils.java 3 76.0% Totals Change from base Build 1021: -0.005% Covered Lines: 12113 Relevant Lines: 20514 💛 - Coveralls
gharchive/pull-request
2020-03-08T16:09:22
2025-04-01T04:33:29.319508
{ "authors": [ "coveralls", "terrymanu" ], "repo": "apache/incubator-shardingsphere", "url": "https://github.com/apache/incubator-shardingsphere/pull/4664", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1083051248
[BUG] abnormal return of file request. serious Is there an existing issue for this? [X] I have searched the existing issues Current Behavior Service A provides a file / picture download. shenyu-gateway return is an abnormal file character stream. Expected Behavior shenyu-gateway will return file stream. Steps To Reproduce No response Environment ShenYu version(s):2.4.2 Debug logs I don't think the file return should be converted to string. Anything else? No response I can fix this problem. Can you provide responseHeader?
gharchive/issue
2021-12-17T09:25:01
2025-04-01T04:33:29.323150
{ "authors": [ "MarsOu1995", "impactCn" ], "repo": "apache/incubator-shenyu", "url": "https://github.com/apache/incubator-shenyu/issues/2611", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1305539139
[type:refactor] refactor-spring-cloud-loadbalancer Make sure that: [ ] You have read the contribution guidelines. [ ] You submit test cases (unit or integration tests) that back your changes. [ ] Your local test passed ./mvnw clean install -Dmaven.javadoc.skip=true. Codecov Report Merging #3714 (9627a7a) into master (ec17f7e) will increase coverage by 0.07%. The diff coverage is 88.73%. @@ Coverage Diff @@ ## master #3714 +/- ## ============================================ + Coverage 63.42% 63.49% +0.07% + Complexity 5947 5938 -9 ============================================ Files 897 897 Lines 24681 24630 -51 Branches 2249 2238 -11 ============================================ - Hits 15653 15638 -15 + Misses 7607 7568 -39 - Partials 1421 1424 +3 Impacted Files Coverage Δ ...d/loadbalance/ShenyuSpringCloudServiceChooser.java 87.69% <87.69%> (ø) ...e/shenyu/plugin/springcloud/SpringCloudPlugin.java 94.11% <100.00%> (+23.52%) :arrow_up: ...in/springcloud/SpringCloudPluginConfiguration.java 100.00% <100.00%> (ø) ...yu/sync/data/http/refresh/AbstractDataRefresh.java 68.96% <0.00%> (-13.80%) :arrow_down: ...a/org/apache/shenyu/common/utils/VersionUtils.java 67.85% <0.00%> (-10.72%) :arrow_down: ...henyu/admin/service/impl/UpstreamCheckService.java 54.66% <0.00%> (-9.34%) :arrow_down: ...rg/apache/shenyu/plugin/rewrite/RewritePlugin.java 66.66% <0.00%> (-5.56%) :arrow_down: ...in/response/strategy/NettyClientMessageWriter.java 66.66% <0.00%> (-4.17%) :arrow_down: ...che/shenyu/sync/data/http/HttpSyncDataService.java 81.63% <0.00%> (-4.09%) :arrow_down: ...icsearch/client/ElasticSearchLogCollectClient.java 70.00% <0.00%> (-2.00%) :arrow_down: ... and 28 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update ec17f7e...9627a7a. Read the comment docs.
gharchive/pull-request
2022-07-15T03:50:19
2025-04-01T04:33:29.341763
{ "authors": [ "codecov-commenter", "moremind" ], "repo": "apache/incubator-shenyu", "url": "https://github.com/apache/incubator-shenyu/pull/3714", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
389696941
请问什么时候可以兼容spring cloud gateway呢? Please answer these questions before submitting your issue. Why do you submit this issue? [ ] Question or discussion [ ] Bug [ ] Requirement [ ] Feature or performance improvement Question What do you want to know? Bug Which version of SkyWalking, OS and JRE? Which company or project? What happen? If possible, provide a way for reproducing the error. e.g. demo application, component version. Requirement or improvement Please describe about your requirements or improvement suggestions. English only. Requirement only accept when some contributor wants it. spring cloud gateway 现在不兼容
gharchive/issue
2018-12-11T10:35:55
2025-04-01T04:33:29.345857
{ "authors": [ "github24518", "monkeyAlwaysWin", "wu-sheng" ], "repo": "apache/incubator-skywalking", "url": "https://github.com/apache/incubator-skywalking/issues/2034", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
428581362
Increase the test coverage rate Today, SkyWalking's tests are mostly around integration tests, which are benifit for end users. Now we hope through this, people from community could help on Unit Tests. JUnit and Mockito are provided. This is a good start for everyone wants to contribute codes. When you submit test cases PR, please ref to this issue. Also, in this kind of pull request, we don't limit how many test cases you provide. Even one new test case, which could effectively test one more line, it will be acceptable. Just try your best to make tests more sense. Where can I find the code coverage report? The coverage is here, https://coveralls.io/github/apache/skywalking Update here, we are only just increasing the test case coverage, we are doing more. [x] @kezhenxu94 set up the Jenkins of ASF for power test env. [ ] @ascrutae lead the agent test in Jenkins jobs of every PR when necessary [ ] @kezhenxu94 lead the e2e test, make agent/backend integration test done in every PR. We have e2e tests fro single node w/ H2 and cluster mode w/ ES. We are trying to build more. Including MySQL storage tests, alarm tests, and aggregation accuracy tests. After #3561, we have 30%+ coverage now. @ascrutae lead the agent test in Jenkins jobs of every PR when necessary This task is lead by @dmsolr on #3528. Ownership replaced. Agent autotest framework has been provided, and 300+ test cases have been added. Seems working well. Just working for @dmsolr to provide the official documents, rather than guide users to do that one by one.
gharchive/issue
2019-04-03T06:04:35
2025-04-01T04:33:29.350512
{ "authors": [ "flycash", "wu-sheng" ], "repo": "apache/incubator-skywalking", "url": "https://github.com/apache/incubator-skywalking/issues/2438", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
742904233
docs: fix docker config paths SUMMARY Just pointin' people in the right direction BEFORE/AFTER SCREENSHOTS OR ANIMATED GIF TEST PLAN ADDITIONAL INFORMATION [ ] Has associated issue: [ ] Changes UI [ ] Requires DB Migration. [ ] Confirm DB Migration upgrade and downgrade tested. [ ] Introduces new feature or API [ ] Removes existing feature or API this is a docs change and e2e has failed 3 times :( xit! Codecov Report Merging #11703 (942b6f6) into master (68693c7) will decrease coverage by 0.04%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #11703 +/- ## ========================================== - Coverage 62.86% 62.81% -0.05% ========================================== Files 889 454 -435 Lines 43055 16185 -26870 Branches 4017 4017 ========================================== - Hits 27065 10167 -16898 + Misses 15811 5839 -9972 Partials 179 179 Flag Coverage Δ javascript 62.81% <ø> (ø) python ? Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ ...t/annotation_layers/annotations/commands/create.py superset/examples/multi_line.py superset/queries/filters.py superset/datasets/commands/refresh.py ...uperset/migrations/versions/1d2ddd543133_log_dt.py ...s/versions/3c3ffe173e4f_add_sql_string_to_table.py superset/stats_logger.py superset/db_engine_specs/redshift.py superset/examples/misc_dashboard.py superset/db_engine_specs/bigquery.py ... and 425 more Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 68693c7...942b6f6. Read the comment docs.
gharchive/pull-request
2020-11-14T02:19:37
2025-04-01T04:33:29.367002
{ "authors": [ "codecov-io", "mistercrunch", "suddjian" ], "repo": "apache/incubator-superset", "url": "https://github.com/apache/incubator-superset/pull/11703", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
342021375
Migrated to click Fixed merge conflicts and linting errors from https://github.com/apache/incubator-superset/pull/4078. Please fix the merge conflicts and build errors Codecov Report Merging #5410 into master will increase coverage by <.01%. The diff coverage is 92.85%. @@ Coverage Diff @@ ## master #5410 +/- ## ========================================= + Coverage 59.09% 59.1% +<.01% ========================================= Files 372 372 Lines 23747 23759 +12 Branches 2758 2758 ========================================= + Hits 14033 14042 +9 - Misses 9699 9702 +3 Partials 15 15 Impacted Files Coverage Δ superset/cli.py 46.59% <92.85%> (+1.9%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 99ce7b7...aa6d503. Read the comment docs. @mistercrunch Fixed. Can we now remove flask-script from our setup.py & requeirments.txt ? @mistercrunch Oh, yeah good call. Sorry about that.
gharchive/pull-request
2018-07-17T17:51:35
2025-04-01T04:33:29.375524
{ "authors": [ "codecov-io", "gbates101", "mistercrunch" ], "repo": "apache/incubator-superset", "url": "https://github.com/apache/incubator-superset/pull/5410", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
440256903
feat: Scheduling queries from SQL Lab (#7416) merged from commit 90eef51 on lyft-release-sp8 @betodealmeida @xtinec Codecov Report Merging #7446 into master will decrease coverage by 0.03%. The diff coverage is 52%. @@ Coverage Diff @@ ## master #7446 +/- ## ========================================== - Coverage 65.26% 65.22% -0.04% ========================================== Files 430 431 +1 Lines 21078 21147 +69 Branches 2338 2343 +5 ========================================== + Hits 13757 13794 +37 - Misses 7205 7237 +32 Partials 116 116 Impacted Files Coverage Δ ...src/components/FilterableTable/FilterableTable.jsx 91.01% <ø> (ø) :arrow_up: .../assets/src/SqlLab/components/QueryAutoRefresh.jsx 11.9% <0%> (-0.92%) :arrow_down: superset/views/sql_lab.py 94.64% <100%> (ø) :arrow_up: superset/assets/src/featureFlags.ts 88.88% <100%> (+1.38%) :arrow_up: ...sets/src/SqlLab/components/ScheduleQueryButton.jsx 13.33% <13.33%> (ø) ...uperset/assets/src/SqlLab/components/SqlEditor.jsx 59.8% <50%> (-0.2%) :arrow_down: superset/db_engine_specs.py 61.45% <86.11%> (+1.22%) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update c8bb7e0...4b5e54f. Read the comment docs.
gharchive/pull-request
2019-05-03T23:25:13
2025-04-01T04:33:29.387073
{ "authors": [ "DiggidyDave", "codecov-io" ], "repo": "apache/incubator-superset", "url": "https://github.com/apache/incubator-superset/pull/7446", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
613817538
Fusion base directory is not mounted I see u guys made a lot of changes. And I cloned ur the newest code to give it a shot. I have to say this is much better than the old version. But when I was try to launch Teaclave Services with docker-compose command, I got a situation here. Like below. teaclave-execution-service | [2020-05-07T04:25:14Z ERROR teaclave_execution_service_enclave] Failed to start the service: Fusion base directory is not mounted: /tmp/fusion_data teaclave-execution-service | Service exit with error: ServiceError teaclave-execution-service exited with code 0 It seems the execution-service is down because of a mount thing. Waiting for any help. BTW, if I run it successfully, what would I see when I hit the link(https://api.trustedservices.intel.com:443). I don not see any doc about it. BTW AGAIN, the new docs ur uploaded explain my lots of questions. "Fusion base directory" is a directory for store temporary (encrypted) fusion data and can be set in the runtime config (https://github.com/apache/incubator-teaclave/blob/master/config/runtime.config.toml#L34). The issue is caused by a check in the execution service (https://github.com/apache/incubator-teaclave/blob/master/services/execution/enclave/src/lib.rs#L66). You can create a directory by your self in /tmp/fusion_data to start the service. Thanks for your report. We will propose a better implementation for this case. At last, the link you mentioned (https://api.trustedservices.intel.com:443) is the URL of attestation service. API specifications can be found here: https://software.intel.com/content/dam/develop/public/us/en/documents/sgx-attestation-api-spec.pdf.
gharchive/issue
2020-05-07T06:49:56
2025-04-01T04:33:29.392006
{ "authors": [ "czzmmc", "mssun" ], "repo": "apache/incubator-teaclave", "url": "https://github.com/apache/incubator-teaclave/issues/287", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
189362585
Fix TC-43: TR prints 'No certificate data for https' for a ds with cert @dneuman64 @trevorackerman Could you help to review if this is needed? Thanks https://issues.apache.org/jira/browse/TC-43 Looks good to me. I don't think we need a debug log entry but it's not a huge deal.
gharchive/pull-request
2016-11-15T11:06:35
2025-04-01T04:33:29.393796
{ "authors": [ "trevorackerman", "weifensh" ], "repo": "apache/incubator-trafficcontrol", "url": "https://github.com/apache/incubator-trafficcontrol/pull/81", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
811860997
alter doc for check style alter doc for check style @qiaojialin @jixuan1989 @HTHou hi, could you please review this? @qiaojialin @jixuan1989 @HTHou hi, could you please review this? Nice catch!
gharchive/pull-request
2021-02-19T09:33:20
2025-04-01T04:33:29.395086
{ "authors": [ "HTHou", "wangchao316" ], "repo": "apache/iotdb", "url": "https://github.com/apache/iotdb/pull/2697", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1525336773
JAMES-3756 JMAP endpoint (upload/download) should support being called with accountIds of delegated accounts resolve https://github.com/linagora/james-project/issues/4679 Please rebase @vttranlina rebased green
gharchive/pull-request
2023-01-09T10:49:28
2025-04-01T04:33:29.397468
{ "authors": [ "chibenwa", "vttranlina" ], "repo": "apache/james-project", "url": "https://github.com/apache/james-project/pull/1382", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
841164374
KAFKA-12556: Add --under-preferred-replica-partitions option to describe topics command Whether the preferred replica is the partition leader directly affects the external output traffic of the broker. When the preferred replica of all partitions becomes the leader, the external output traffic of the broker will be in a balanced state. When there are a large number of partition leaders that are not preferred replicas, it will be destroyed this state of balance. Currently, the controller will periodically check the unbalanced ratio of the partition preferred replicas (if enabled) to trigger the preferred replica election, or manually trigger the election through the kafka-leader-election tool. However, if we want to know which partition leader is in the non-preferred replica, we need to look it up in the controller log or judge ourselves from the topic details list. We can add the --under-preferred-replica-partitions configuration option in TopicCommand describe topics to query the list of partitions in the current cluster that are in non-preferred replicas. Committer Checklist (excluded from commit message) [ ] Verify design and implementation [ ] Verify test coverage and CI build status [ ] Verify documentation (including upgrade notes) @chia7712 Can you take a look at this pr?Thanks. :) @wenbingshen Thanks for the PR. We need a KIP for adding a new flag to a command line tool as it is considered as part of the public API. @wenbingshen Thanks for the PR. We need a KIP for adding a new flag to a command line tool as it is considered as part of the public API. Thank you for your comment, I will create a KIP for this. :)
gharchive/pull-request
2021-03-25T17:43:07
2025-04-01T04:33:29.421520
{ "authors": [ "dajac", "wenbingshen" ], "repo": "apache/kafka", "url": "https://github.com/apache/kafka/pull/10403", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1497675098
KAFKA-14492: Extract a method to create LogManager, in order to be overrided by subclass of KafkaServer KAFKA-14492 In our scene, we want to implement a subclass of LogManager, add do some interception like encrypting/auditing. Even more, we want enhance the write log performance in the future. Split alone a create method can provide an extension point to allow user define the action from outside kafka core. Base kafka, entend it, rather than modification. hi, those failed checks are not caused by my modification in this time. what can i do next ? Thanks for the PR. It is important to note that KafkaServer is not considered as a public interface in Kafka. Hence, there is not guarantee that the interface will remain like this in the future. It can basically change anytime. Therefore, I am personally not comfortable with this change. Have you considered alternative approaches? @ijuma What do you think? Yes, a KIP would be required to turn this into a public API. For non public APIs, the code can change at any time. Thanks for your reply. I agree that the KafkaServer is not designed for extend, attempt to replace the LogManager by override a createLogManager, maybe not a permanently method. As your said, a KIP is needed to propose a complete approache that specially focus on which components can be extended in KafkaServer. Design an interface describe which components support extend is obviously a better way. Before an api or interface be provided, allow to extend an object's method properly can be a temporary way with less drawback. At least is better than that fork a standalone repository, then modify with the code. The extend way is just add a kafka-core.jar, then override inside classes. Except the extend side, provide a createLogManager method is coincide with createReplicaManager below those code. It's also a good style only by the readability. Further on, only this modification is negligible, we could meet more problems in the way to extend LogManager. In that time, i am willing to provide a comprehensive plan as a KIP to resolve the extension problem. Thank you both again.
gharchive/pull-request
2022-12-15T02:10:23
2025-04-01T04:33:29.426670
{ "authors": [ "PhantomMaa", "dajac", "ijuma" ], "repo": "apache/kafka", "url": "https://github.com/apache/kafka/pull/12997", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2062313351
KAFKA-14412: Decouple RocksDB access from CF To support future use-cases that use different strategies for accessing RocksDB, we need to de-couple the RocksDB access strategy from the Column Family access strategy. To do this, we now have two separate accessors: DBAccessor: dictates how we access RocksDB. Currently only one strategy is supported: DirectDBAccessor, which access RocksDB directly, via the RocksDB class for all operations. In the future, a BatchedDBAccessor will be added, which enables transactions via WriteBatch. ColumnFamilyAccessor: maps StateStore operations to operations on one or more column families. This is a rename of the old RocksDBDBAccessor. @cadonna @mjsax @ableegoldman @lucasbru @wcarlson5 @bbejeck @vvcephei @guozhangwang This is part of KIP-892, and has been broken out into a separate PR to reduce the review burden on the main KIP-892 implementation, since it can be merged independently. There are no tests, because there are no behavioural changes, just a refactoring. The existing test suite should ensure no regressions. @nicktelford Seems like all build jobs timed out. Could you take a look? @lucasbru I think this is just the CI causing trouble again. The build passes locally. Is there a way to rerun the build without pushing more commits? @nicktelford I'll rerun it. That being said, the last jobs on trunk have all finished within 3-5 hours, so this must be caused by either infrastructure or the code in this PR. @lucasbru OK, my bad. It turns out I did a minor refactoring after I ran the test suite yesterday that was so insignificant I didn't think I needed to run the tests again... Turns out I was wrong :see_no_evil: I've fixed the bug now and the tests pass locally for real now.
gharchive/pull-request
2024-01-02T11:52:50
2025-04-01T04:33:29.431227
{ "authors": [ "lucasbru", "nicktelford" ], "repo": "apache/kafka", "url": "https://github.com/apache/kafka/pull/15105", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
110295179
KAFKA-2620: Introduce Scalariform This is a WIP. There will likely need to be discussion around a rule set we want to use and if we actually want to include this in the build. This could also be used as a one time rebase. This commit does not include the actual formatting changes. To see what effect the patch has run gradle formatScala @granthenke Would you like to continue working on this JIRA and PR? I think we may want to use scalafmt instead as it supports a maxColumn setting: http://scalameta.org/scalafmt/#FAQ/Troubleshooting Closing as this has been inactive for a while. This is a WIP. There will likely need to be discussion around a rule set we want to use and if we actually want to include this in the build. This could also be used as a one time rebase. This commit does not include the actual formatting changes. To see what effect the patch has run gradle formatScala
gharchive/pull-request
2015-10-07T19:06:30
2025-04-01T04:33:29.434572
{ "authors": [ "b-goyal", "granthenke", "guozhangwang", "ijuma" ], "repo": "apache/kafka", "url": "https://github.com/apache/kafka/pull/284", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
588005308
Suppress detailed responses for security-sensitive (PCI-DSS) environments Before this PR, Kafka Connect's exception mapper would print an unchecked messages from unhandled stack traces. I am contributing this code change so that the community has the option to change this behavior based on a configuration option, "error.rest.response.message.detail.enabled" As this functionality is an API-level change, as it changes the response of handled errors, it should not be enabled by default. I have provided a unit test for handling the expected response if the option is enabled. Issue: https://issues.apache.org/jira/browse/KAFKA-9766
gharchive/pull-request
2020-03-25T21:34:51
2025-04-01T04:33:29.436247
{ "authors": [ "cwpenhale" ], "repo": "apache/kafka", "url": "https://github.com/apache/kafka/pull/8355", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
672717996
KAFKA-10314: KafkaStorageException on reassignment when offline log d… …irectories exist Make sure that we set the isNew field in LeaderAndIsrRequest correctly for brokers that gets added to the replica set on reassignment. This is tested by creating a variant of ControllerIntergationTest.testPartitionReassignment() that makes one of the log directories on the target broker offline before initiating the reassignment. Without the change to the way isNew is set, this fails after a timeout. With the change, it succeeds. To facilitate calling causeLogDirFailure() both from ControllerIntegrationTest and LogDirFailureTest, the method was moved to TestUtils along with the other helper methods that deals with interacting with KafkaServer instances for test cases. More detailed description of your change, if necessary. The PR title and PR message become the squashed commit message, so use a separate comment to ping reviewers. Summary of testing strategy (including rationale) for the feature or bug fix. Unit and/or integration tests are expected for any behaviour change and system tests should be considered for larger changes. Committer Checklist (excluded from commit message) [ ] Verify design and implementation [ ] Verify test coverage and CI build status [ ] Verify documentation (including upgrade notes) Ok to test @nresare Good catch! The change looks good. Before merging I'd like to get this change tested in the CI pipeline. The Apache Jenkins setup just changed a few days ago and it needs 241e1447fae8b85b3bb491f371357b8f92d2da72 to run. So can you rebase on trunk? Trying to navigate the tests, it seems they are related to flakiness in other parts of the overall test suite. Is it possible to re-trigger tests?
gharchive/pull-request
2020-08-04T11:26:09
2025-04-01T04:33:29.440980
{ "authors": [ "mimaison", "nresare" ], "repo": "apache/kafka", "url": "https://github.com/apache/kafka/pull/9122", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1733886208
Extract table from ResolvedIdentifier for DropTable in Spark 3.4 Why are the changes needed? adapting changes in logical plan of DropTable in Spark 3.4 by extracting table object from ResolvedIdntifier, to fix test w/ Spark 3.4 ut "DropTable" How was this patch tested? [ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible [ ] Add screenshots for manual tests if appropriate [x] Run test locally before make a pull request cc @zhouyifan279 Codecov Report Merging #4910 (53c76f6) into master (8f61835) will not change coverage. The diff coverage is 0.00%. @@ Coverage Diff @@ ## master #4910 +/- ## ====================================== Coverage 0.00% 0.00% ====================================== Files 557 557 Lines 30700 30703 +3 Branches 3995 3996 +1 ====================================== - Misses 30700 30703 +3 Impacted Files Coverage Δ ...ubi/plugin/spark/authz/serde/tableExtractors.scala 0.00% <0.00%> (ø) :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more Thanks, merged to master.
gharchive/pull-request
2023-05-31T10:26:31
2025-04-01T04:33:29.448507
{ "authors": [ "bowenliang123", "codecov-commenter" ], "repo": "apache/kyuubi", "url": "https://github.com/apache/kyuubi/pull/4910", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2083636373
[KYUUBI #5964][BUG] Avoid check not fully optimized query for InsertIntoDataSourceDirCommand and InsertIntoDataSourceCommand :mag: Description Issue References 🔗 This pull request fixes #5964 Describe Your Solution 🔧 InsertIntoDataSourceDirCommand and InsertIntoDataSourceCommand‘s query is not fully optimized, we direct check it's query will cause request privilege that we haven't used. We can directly ignore the query's check. Since we will check it's generated plan. Still will request the correct privilege of the SQL Types of changes :bookmark: [x] Bugfix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) Test Plan 🧪 Behavior Without This Pull Request :coffin: Behavior With This Pull Request :tada: Related Unit Tests Checklist 📝 [ ] This patch was not authored or co-authored using Generative Tooling Be nice. Be informative. Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Comparison is base (3915fe8) 61.21% compared to head (1adcf8d) 61.06%. Report is 1 commits behind head on master. Additional details and impacted files @@ Coverage Diff @@ ## master #5983 +/- ## ============================================ - Coverage 61.21% 61.06% -0.15% Complexity 23 23 ============================================ Files 622 622 Lines 36897 37036 +139 Branches 5016 5023 +7 ============================================ + Hits 22585 22617 +32 - Misses 11877 11973 +96 - Partials 2435 2446 +11 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here. ping @yaooqinn
gharchive/pull-request
2024-01-16T10:54:02
2025-04-01T04:33:29.456010
{ "authors": [ "AngersZhuuuu", "codecov-commenter" ], "repo": "apache/kyuubi", "url": "https://github.com/apache/kyuubi/pull/5983", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2286801315
Improve closeEngine method names and logs due to closeSession :mag: Description Issue References 🔗 This pull request aims to improve closeEngine method names and logs due to closeSession. when the share level is CONNECTION, closeSession triggers the closeEngine, there are some minor issues with logging and method name. Describe Your Solution 🔧 Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change. Types of changes :bookmark: [ ] Bugfix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) Test Plan 🧪 Behavior Without This Pull Request :coffin: Behavior With This Pull Request :tada: Related Unit Tests Checklist 📝 [x] This patch was not authored or co-authored using Generative Tooling Be nice. Be informative. Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Project coverage is 58.44%. Comparing base (12c5568) to head (1b5a061). Additional details and impacted files @@ Coverage Diff @@ ## master #6376 +/- ## ============================================ - Coverage 58.51% 58.44% -0.08% Complexity 24 24 ============================================ Files 653 653 Lines 39895 39895 Branches 5482 5482 ============================================ - Hits 23345 23315 -30 - Misses 14055 14081 +26 - Partials 2495 2499 +4 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here. @pan3793 Please take a look if you find a time. thanks, merged to master.
gharchive/pull-request
2024-05-09T03:01:58
2025-04-01T04:33:29.464245
{ "authors": [ "codecov-commenter", "yikf" ], "repo": "apache/kyuubi", "url": "https://github.com/apache/kyuubi/pull/6376", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
473539609
Fix hash validation with stream uploads Fix hash validation with stream uploads Description Bug fix for cloudfiles upload_object_via_stream() erroneously throwing a hash mismatch when using a file handle as the iterator. Example: from libcloud.storage.providers import get_driver from libcloud.storage.types import ObjectDoesNotExistError driver = get_driver('cloudfiles') driver = driver('user', 'key') container = driver.get_container(container_name='test') with open('/tmp/test.file', 'w') as fh: fh.write('test') with open('/tmp/test.file') as fh: driver.upload_object_via_stream(iterator=fh, container=container, object_name='test.file') This would result in the exception: libcloud.storage.types.ObjectHashMismatchError: <ObjectHashMismatchError in <libcloud.storage.drivers.cloudfiles.CloudFilesStorageDriver object at 0x103f43d68>, value=MD5 hash checksum does not match (expected=d41d8cd98f00b204e9800998ecf8427e, actual=098f6bcd4621d373cade4e832627b4f6), object = test.file> d41d8cd98f00b204e9800998ecf8427e is the hash for an empty file. The problem is the upload happens before the hash is calculated and the file handle is read to the EOF. The fix is to seek the file handle to the beginning before feeding it into the hash function. Status done, ready for review Checklist (tick everything that applies) [x] Code linting (required, can be done after the PR checks) [ ] Documentation [ ] Tests [ ] ICLA (required for bigger changes) Thanks for the contribution. Can you please also add a corresponding test case? I've added a test case (858d9e6e16d0f1a2160537d9acac1ca0f897ef70) and will go ahead and merge this into trunk. It's also worth pointing out that you made change in the base driver class so this change potentially affects all the drivers. I think the change is the correct one, but we should still keep an eye on potential issues once a new version is released. Sorry, I've been busy lately, but did intend to add tests. Thanks for doing it for me, @Kami!
gharchive/pull-request
2019-07-26T21:17:49
2025-04-01T04:33:29.469222
{ "authors": [ "Kami", "gvengel" ], "repo": "apache/libcloud", "url": "https://github.com/apache/libcloud/pull/1326", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }