2024-11-12 15:31:13,314 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-12 15:31:13,327 main DEBUG Took 0.010821 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-11-12 15:31:13,327 main DEBUG PluginManager 'Core' found 129 plugins 2024-11-12 15:31:13,328 main DEBUG PluginManager 'Level' found 0 plugins 2024-11-12 15:31:13,329 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-11-12 15:31:13,330 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,345 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-11-12 15:31:13,356 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,358 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,358 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,359 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,359 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,360 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,361 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,361 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,361 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,362 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,363 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,363 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,364 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,364 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,364 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,365 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,365 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,365 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,366 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,366 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,367 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,367 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,367 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,368 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-12 15:31:13,368 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,369 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-11-12 15:31:13,370 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-12 15:31:13,371 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-11-12 15:31:13,373 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-11-12 15:31:13,373 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-11-12 15:31:13,375 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-11-12 15:31:13,375 main DEBUG PluginManager 'Converter' found 47 plugins 2024-11-12 15:31:13,384 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-11-12 15:31:13,386 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-11-12 15:31:13,388 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-11-12 15:31:13,388 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-11-12 15:31:13,389 main DEBUG createAppenders(={Console}) 2024-11-12 15:31:13,390 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-11-12 15:31:13,390 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-12 15:31:13,391 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-11-12 15:31:13,391 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-11-12 15:31:13,391 main DEBUG OutputStream closed 2024-11-12 15:31:13,392 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-11-12 15:31:13,392 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-11-12 15:31:13,392 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-11-12 15:31:13,460 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-11-12 15:31:13,462 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-11-12 15:31:13,464 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-11-12 15:31:13,465 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-11-12 15:31:13,465 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-11-12 15:31:13,466 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-11-12 15:31:13,466 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-11-12 15:31:13,467 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-11-12 15:31:13,467 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-11-12 15:31:13,467 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-11-12 15:31:13,468 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-11-12 15:31:13,468 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-11-12 15:31:13,468 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-11-12 15:31:13,469 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-11-12 15:31:13,469 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-11-12 15:31:13,469 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-11-12 15:31:13,470 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-11-12 15:31:13,471 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-11-12 15:31:13,473 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-12 15:31:13,474 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-logging/target/hbase-logging-3.0.0-beta-2-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-11-12 15:31:13,474 main DEBUG Shutdown hook enabled. Registering a new one. 2024-11-12 15:31:13,475 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-11-12T15:31:13,489 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-11-12 15:31:13,491 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-11-12 15:31:13,492 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-12T15:31:13,711 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6 2024-11-12T15:31:13,737 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896, deleteOnExit=true 2024-11-12T15:31:13,739 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/test.cache.data in system properties and HBase conf 2024-11-12T15:31:13,740 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.tmp.dir in system properties and HBase conf 2024-11-12T15:31:13,741 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir in system properties and HBase conf 2024-11-12T15:31:13,742 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-12T15:31:13,742 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-12T15:31:13,743 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-12T15:31:13,827 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-11-12T15:31:13,909 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-12T15:31:13,913 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-12T15:31:13,914 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-12T15:31:13,915 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-12T15:31:13,915 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-12T15:31:13,916 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-12T15:31:13,916 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-12T15:31:13,917 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-12T15:31:13,917 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-12T15:31:13,918 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-12T15:31:13,918 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/nfs.dump.dir in system properties and HBase conf 2024-11-12T15:31:13,918 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/java.io.tmpdir in system properties and HBase conf 2024-11-12T15:31:13,919 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-12T15:31:13,919 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-12T15:31:13,920 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-12T15:31:14,968 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-11-12T15:31:15,043 INFO [Time-limited test {}] log.Log(170): Logging initialized @2345ms to org.eclipse.jetty.util.log.Slf4jLog 2024-11-12T15:31:15,109 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:15,165 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:15,183 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:15,184 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:15,185 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-12T15:31:15,195 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:15,198 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@b03c34d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:15,199 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3cbd6fd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:15,365 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5cb83937{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/java.io.tmpdir/jetty-localhost-39289-hadoop-hdfs-3_4_1-tests_jar-_-any-10648769834685247033/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-12T15:31:15,372 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@69b5b273{HTTP/1.1, (http/1.1)}{localhost:39289} 2024-11-12T15:31:15,372 INFO [Time-limited test {}] server.Server(415): Started @2674ms 2024-11-12T15:31:15,914 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:15,921 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:15,922 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:15,922 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:15,922 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-12T15:31:15,923 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@179ed6d0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:15,924 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4f02cc61{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:16,021 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3990ff75{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/java.io.tmpdir/jetty-localhost-40091-hadoop-hdfs-3_4_1-tests_jar-_-any-14178218629057200151/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:16,022 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@48731e1b{HTTP/1.1, (http/1.1)}{localhost:40091} 2024-11-12T15:31:16,022 INFO [Time-limited test {}] server.Server(415): Started @3324ms 2024-11-12T15:31:16,073 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-12T15:31:16,175 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:16,182 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:16,185 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:16,185 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:16,185 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-12T15:31:16,186 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@50510811{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:16,187 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@125705fb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:16,291 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@133f1bad{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/java.io.tmpdir/jetty-localhost-46029-hadoop-hdfs-3_4_1-tests_jar-_-any-5604729638911061529/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:16,292 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@39860596{HTTP/1.1, (http/1.1)}{localhost:46029} 2024-11-12T15:31:16,292 INFO [Time-limited test {}] server.Server(415): Started @3594ms 2024-11-12T15:31:16,294 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-12T15:31:16,333 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:16,337 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:16,340 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:16,341 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:16,341 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-12T15:31:16,344 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@28d0ee11{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:16,345 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@14c1b227{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:16,464 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@17f8e572{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/java.io.tmpdir/jetty-localhost-36901-hadoop-hdfs-3_4_1-tests_jar-_-any-13270551521745096574/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:16,465 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11aaae40{HTTP/1.1, (http/1.1)}{localhost:36901} 2024-11-12T15:31:16,465 INFO [Time-limited test {}] server.Server(415): Started @3767ms 2024-11-12T15:31:16,467 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-12T15:31:17,549 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data4/current/BP-934991163-172.17.0.2-1731425474470/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:17,549 WARN [Thread-122 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data1/current/BP-934991163-172.17.0.2-1731425474470/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:17,549 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data3/current/BP-934991163-172.17.0.2-1731425474470/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:17,549 WARN [Thread-123 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data2/current/BP-934991163-172.17.0.2-1731425474470/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:17,581 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-12T15:31:17,581 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-12T15:31:17,624 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data5/current/BP-934991163-172.17.0.2-1731425474470/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:17,624 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data6/current/BP-934991163-172.17.0.2-1731425474470/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:17,625 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5e8699a55e7a85bd with lease ID 0x497cc9360a3ce7cf: Processing first storage report for DS-8cd4ebbe-e241-4236-81a5-1900de5094c9 from datanode DatanodeRegistration(127.0.0.1:37019, datanodeUuid=6492489c-4b63-46ed-ad64-7d0260599c6f, infoPort=37755, infoSecurePort=0, ipcPort=33319, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470) 2024-11-12T15:31:17,626 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5e8699a55e7a85bd with lease ID 0x497cc9360a3ce7cf: from storage DS-8cd4ebbe-e241-4236-81a5-1900de5094c9 node DatanodeRegistration(127.0.0.1:37019, datanodeUuid=6492489c-4b63-46ed-ad64-7d0260599c6f, infoPort=37755, infoSecurePort=0, ipcPort=33319, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-12T15:31:17,627 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x26c8f584b57c3277 with lease ID 0x497cc9360a3ce7ce: Processing first storage report for DS-77fbb761-241d-485a-acf0-7881db3c0972 from datanode DatanodeRegistration(127.0.0.1:42999, datanodeUuid=4cb0ffa7-4ac3-40a3-b6d3-1556f7f964e1, infoPort=45141, infoSecurePort=0, ipcPort=46817, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470) 2024-11-12T15:31:17,627 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x26c8f584b57c3277 with lease ID 0x497cc9360a3ce7ce: from storage DS-77fbb761-241d-485a-acf0-7881db3c0972 node DatanodeRegistration(127.0.0.1:42999, datanodeUuid=4cb0ffa7-4ac3-40a3-b6d3-1556f7f964e1, infoPort=45141, infoSecurePort=0, ipcPort=46817, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-12T15:31:17,627 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5e8699a55e7a85bd with lease ID 0x497cc9360a3ce7cf: Processing first storage report for DS-6db7cae4-b76c-4a23-8cd9-08a04551c65a from datanode DatanodeRegistration(127.0.0.1:37019, datanodeUuid=6492489c-4b63-46ed-ad64-7d0260599c6f, infoPort=37755, infoSecurePort=0, ipcPort=33319, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470) 2024-11-12T15:31:17,627 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5e8699a55e7a85bd with lease ID 0x497cc9360a3ce7cf: from storage DS-6db7cae4-b76c-4a23-8cd9-08a04551c65a node DatanodeRegistration(127.0.0.1:37019, datanodeUuid=6492489c-4b63-46ed-ad64-7d0260599c6f, infoPort=37755, infoSecurePort=0, ipcPort=33319, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:17,628 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x26c8f584b57c3277 with lease ID 0x497cc9360a3ce7ce: Processing first storage report for DS-bd15cfd9-127f-4dd3-a786-79377c9b3000 from datanode DatanodeRegistration(127.0.0.1:42999, datanodeUuid=4cb0ffa7-4ac3-40a3-b6d3-1556f7f964e1, infoPort=45141, infoSecurePort=0, ipcPort=46817, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470) 2024-11-12T15:31:17,628 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x26c8f584b57c3277 with lease ID 0x497cc9360a3ce7ce: from storage DS-bd15cfd9-127f-4dd3-a786-79377c9b3000 node DatanodeRegistration(127.0.0.1:42999, datanodeUuid=4cb0ffa7-4ac3-40a3-b6d3-1556f7f964e1, infoPort=45141, infoSecurePort=0, ipcPort=46817, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:17,646 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-12T15:31:17,652 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xca4fb043183e1d67 with lease ID 0x497cc9360a3ce7d0: Processing first storage report for DS-9a38f403-ac87-4dad-bce7-330500803097 from datanode DatanodeRegistration(127.0.0.1:35073, datanodeUuid=3beb0c95-28b9-4eb1-ada9-486403015f38, infoPort=42313, infoSecurePort=0, ipcPort=38911, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470) 2024-11-12T15:31:17,652 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xca4fb043183e1d67 with lease ID 0x497cc9360a3ce7d0: from storage DS-9a38f403-ac87-4dad-bce7-330500803097 node DatanodeRegistration(127.0.0.1:35073, datanodeUuid=3beb0c95-28b9-4eb1-ada9-486403015f38, infoPort=42313, infoSecurePort=0, ipcPort=38911, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:17,652 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xca4fb043183e1d67 with lease ID 0x497cc9360a3ce7d0: Processing first storage report for DS-669d2b81-d883-476f-945e-5b64c2744543 from datanode DatanodeRegistration(127.0.0.1:35073, datanodeUuid=3beb0c95-28b9-4eb1-ada9-486403015f38, infoPort=42313, infoSecurePort=0, ipcPort=38911, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470) 2024-11-12T15:31:17,652 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xca4fb043183e1d67 with lease ID 0x497cc9360a3ce7d0: from storage DS-669d2b81-d883-476f-945e-5b64c2744543 node DatanodeRegistration(127.0.0.1:35073, datanodeUuid=3beb0c95-28b9-4eb1-ada9-486403015f38, infoPort=42313, infoSecurePort=0, ipcPort=38911, storageInfo=lv=-57;cid=testClusterID;nsid=1603442158;c=1731425474470), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:17,686 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6 2024-11-12T15:31:17,749 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-11-12T15:31:17,806 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=158, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=207, ProcessCount=11, AvailableMemoryMB=8619 2024-11-12T15:31:17,808 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-12T15:31:17,815 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-11-12T15:31:17,896 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/zookeeper_0, clientPort=58397, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-12T15:31:17,906 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=58397 2024-11-12T15:31:17,915 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:17,918 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:17,990 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:17,991 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:18,030 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:33322 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:35073:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33322 dst: /127.0.0.1:35073 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:18,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775792_1002 (size=7) 2024-11-12T15:31:18,449 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:18,461 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349 with version=8 2024-11-12T15:31:18,462 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/hbase-staging 2024-11-12T15:31:18,545 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-11-12T15:31:18,789 INFO [Time-limited test {}] client.ConnectionUtils(128): master/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:18,798 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:18,799 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:18,804 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:18,804 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:18,805 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:18,968 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-12T15:31:19,030 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-11-12T15:31:19,040 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-11-12T15:31:19,044 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:19,066 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 92793 (auto-detected) 2024-11-12T15:31:19,068 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-11-12T15:31:19,085 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:40305 2024-11-12T15:31:19,104 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:40305 connecting to ZooKeeper ensemble=127.0.0.1:58397 2024-11-12T15:31:19,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:403050x0, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:19,224 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:40305-0x1012fac209b0000 connected 2024-11-12T15:31:19,326 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,331 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,342 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:19,346 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349, hbase.cluster.distributed=false 2024-11-12T15:31:19,367 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:19,372 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40305 2024-11-12T15:31:19,372 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40305 2024-11-12T15:31:19,373 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40305 2024-11-12T15:31:19,373 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40305 2024-11-12T15:31:19,374 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40305 2024-11-12T15:31:19,466 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:19,467 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,468 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,468 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:19,468 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,468 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:19,471 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-12T15:31:19,473 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:19,474 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36945 2024-11-12T15:31:19,476 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36945 connecting to ZooKeeper ensemble=127.0.0.1:58397 2024-11-12T15:31:19,477 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,479 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,493 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:369450x0, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:19,494 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:19,494 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36945-0x1012fac209b0001 connected 2024-11-12T15:31:19,499 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-12T15:31:19,507 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-12T15:31:19,510 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-12T15:31:19,515 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:19,515 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36945 2024-11-12T15:31:19,516 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36945 2024-11-12T15:31:19,516 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36945 2024-11-12T15:31:19,517 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36945 2024-11-12T15:31:19,517 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36945 2024-11-12T15:31:19,532 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:19,532 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,533 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,533 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:19,533 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,533 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:19,534 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-12T15:31:19,534 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:19,535 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:45955 2024-11-12T15:31:19,537 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:45955 connecting to ZooKeeper ensemble=127.0.0.1:58397 2024-11-12T15:31:19,538 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,540 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,557 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:459550x0, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:19,557 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:459550x0, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:19,558 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:45955-0x1012fac209b0002 connected 2024-11-12T15:31:19,558 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-12T15:31:19,559 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-12T15:31:19,560 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-12T15:31:19,562 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:19,563 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=45955 2024-11-12T15:31:19,563 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=45955 2024-11-12T15:31:19,564 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=45955 2024-11-12T15:31:19,565 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=45955 2024-11-12T15:31:19,566 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=45955 2024-11-12T15:31:19,581 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:19,582 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,582 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,582 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:19,582 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:19,582 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:19,582 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-12T15:31:19,583 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:19,584 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41281 2024-11-12T15:31:19,585 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:41281 connecting to ZooKeeper ensemble=127.0.0.1:58397 2024-11-12T15:31:19,587 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,589 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,604 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:412810x0, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:19,605 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:41281-0x1012fac209b0003 connected 2024-11-12T15:31:19,605 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:19,605 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-12T15:31:19,606 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-12T15:31:19,607 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-12T15:31:19,609 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:19,610 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41281 2024-11-12T15:31:19,610 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41281 2024-11-12T15:31:19,611 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41281 2024-11-12T15:31:19,612 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41281 2024-11-12T15:31:19,612 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41281 2024-11-12T15:31:19,630 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a6f201fa5419:40305 2024-11-12T15:31:19,631 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/a6f201fa5419,40305,1731425478640 2024-11-12T15:31:19,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,649 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a6f201fa5419,40305,1731425478640 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:19,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,680 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-12T15:31:19,682 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a6f201fa5419,40305,1731425478640 from backup master directory 2024-11-12T15:31:19,693 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,693 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,693 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a6f201fa5419,40305,1731425478640 2024-11-12T15:31:19,693 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,693 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:19,694 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:19,694 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a6f201fa5419,40305,1731425478640 2024-11-12T15:31:19,696 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-11-12T15:31:19,697 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-11-12T15:31:19,756 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/hbase.id] with ID: 7a65a283-bcb2-4966-8aca-7e30e20bb4ab 2024-11-12T15:31:19,756 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/.tmp/hbase.id 2024-11-12T15:31:19,763 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,763 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,766 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:33340 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:35073:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33340 dst: /127.0.0.1:35073 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:19,772 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775776_1004 (size=42) 2024-11-12T15:31:19,773 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:19,773 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/.tmp/hbase.id]:[hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/hbase.id] 2024-11-12T15:31:19,821 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:19,827 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-12T15:31:19,847 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 17ms. 2024-11-12T15:31:19,862 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,862 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,862 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,862 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:19,876 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,876 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,880 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:37908 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:42999:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37908 dst: /127.0.0.1:42999 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:19,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775760_1006 (size=196) 2024-11-12T15:31:19,887 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:19,900 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-12T15:31:19,903 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-12T15:31:19,908 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-12T15:31:19,936 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,936 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,939 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:50998 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50998 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:19,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775744_1008 (size=1189) 2024-11-12T15:31:19,946 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:19,961 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store 2024-11-12T15:31:19,978 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,978 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:19,981 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:37928 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:42999:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37928 dst: /127.0.0.1:42999 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:19,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775728_1010 (size=34) 2024-11-12T15:31:19,987 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:19,991 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-11-12T15:31:19,994 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:19,995 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-12T15:31:19,995 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:19,995 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:19,996 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-12T15:31:19,996 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:19,996 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:19,997 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731425479994Disabling compacts and flushes for region at 1731425479995 (+1 ms)Disabling writes for close at 1731425479996 (+1 ms)Writing region close event to WAL at 1731425479996Closed at 1731425479996 2024-11-12T15:31:19,999 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/.initializing 2024-11-12T15:31:19,999 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/WALs/a6f201fa5419,40305,1731425478640 2024-11-12T15:31:20,007 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-12T15:31:20,020 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C40305%2C1731425478640, suffix=, logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/WALs/a6f201fa5419,40305,1731425478640, archiveDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/oldWALs, maxLogs=10 2024-11-12T15:31:20,047 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/WALs/a6f201fa5419,40305,1731425478640/a6f201fa5419%2C40305%2C1731425478640.1731425480025, exclude list is [], retry=0 2024-11-12T15:31:20,064 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:20,066 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37019,DS-8cd4ebbe-e241-4236-81a5-1900de5094c9,DISK] 2024-11-12T15:31:20,066 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42999,DS-77fbb761-241d-485a-acf0-7881db3c0972,DISK] 2024-11-12T15:31:20,066 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35073,DS-9a38f403-ac87-4dad-bce7-330500803097,DISK] 2024-11-12T15:31:20,068 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-11-12T15:31:20,108 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/WALs/a6f201fa5419,40305,1731425478640/a6f201fa5419%2C40305%2C1731425478640.1731425480025 2024-11-12T15:31:20,109 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:37755:37755),(127.0.0.1/127.0.0.1:42313:42313),(127.0.0.1/127.0.0.1:45141:45141)] 2024-11-12T15:31:20,109 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-12T15:31:20,110 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:20,113 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,114 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,149 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,170 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-12T15:31:20,173 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,175 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:20,176 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,179 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-12T15:31:20,180 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,181 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:20,181 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,184 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-12T15:31:20,184 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,185 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:20,185 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,188 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-12T15:31:20,188 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,189 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:20,189 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,193 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,194 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,198 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,199 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,203 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-12T15:31:20,207 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:20,214 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-12T15:31:20,215 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64184331, jitterRate=-0.04357893764972687}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-12T15:31:20,224 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1731425480125Initializing all the Stores at 1731425480127 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425480128 (+1 ms)Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425480128Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425480129 (+1 ms)Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425480129Cleaning up temporary data from old regions at 1731425480199 (+70 ms)Region opened successfully at 1731425480224 (+25 ms) 2024-11-12T15:31:20,225 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-12T15:31:20,255 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@33c7e611, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:20,282 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-12T15:31:20,292 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-12T15:31:20,292 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-12T15:31:20,294 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-12T15:31:20,296 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-11-12T15:31:20,302 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 5 msec 2024-11-12T15:31:20,302 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-12T15:31:20,331 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-12T15:31:20,340 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-12T15:31:20,393 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-12T15:31:20,397 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-12T15:31:20,399 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-12T15:31:20,409 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-12T15:31:20,411 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-12T15:31:20,414 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-12T15:31:20,425 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-12T15:31:20,426 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-12T15:31:20,435 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-12T15:31:20,454 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-12T15:31:20,461 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-12T15:31:20,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:20,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:20,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:20,472 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:20,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,473 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,477 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=a6f201fa5419,40305,1731425478640, sessionid=0x1012fac209b0000, setting cluster-up flag (Was=false) 2024-11-12T15:31:20,509 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,509 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,509 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,509 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,541 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-12T15:31:20,547 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a6f201fa5419,40305,1731425478640 2024-11-12T15:31:20,573 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,573 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,573 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,573 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:20,604 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-12T15:31:20,608 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a6f201fa5419,40305,1731425478640 2024-11-12T15:31:20,615 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-12T15:31:20,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775788_1002 (size=7) 2024-11-12T15:31:20,631 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775789_1002 (size=7) 2024-11-12T15:31:20,685 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:20,696 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-12T15:31:20,702 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-12T15:31:20,707 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a6f201fa5419,40305,1731425478640 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-12T15:31:20,713 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:20,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:20,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:20,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:20,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a6f201fa5419:0, corePoolSize=10, maxPoolSize=10 2024-11-12T15:31:20,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:20,715 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,716 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1731425510716 2024-11-12T15:31:20,717 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(746): ClusterId : 7a65a283-bcb2-4966-8aca-7e30e20bb4ab 2024-11-12T15:31:20,717 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(746): ClusterId : 7a65a283-bcb2-4966-8aca-7e30e20bb4ab 2024-11-12T15:31:20,718 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-12T15:31:20,719 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-12T15:31:20,720 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(746): ClusterId : 7a65a283-bcb2-4966-8aca-7e30e20bb4ab 2024-11-12T15:31:20,720 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-12T15:31:20,720 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-12T15:31:20,720 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-12T15:31:20,722 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:20,722 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-12T15:31:20,723 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-12T15:31:20,723 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-12T15:31:20,723 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-12T15:31:20,723 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-12T15:31:20,724 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,727 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-12T15:31:20,728 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-12T15:31:20,728 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-12T15:31:20,730 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-12T15:31:20,730 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,730 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-12T15:31:20,730 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-12T15:31:20,732 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.large.0-1731425480731,5,FailOnTimeoutGroup] 2024-11-12T15:31:20,732 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.small.0-1731425480732,5,FailOnTimeoutGroup] 2024-11-12T15:31:20,733 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,733 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-12T15:31:20,734 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,734 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,737 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:20,738 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:20,742 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:51030 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:51030 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:20,743 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-12T15:31:20,743 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-12T15:31:20,743 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-12T15:31:20,743 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-12T15:31:20,743 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-12T15:31:20,743 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-12T15:31:20,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775712_1013 (size=1321) 2024-11-12T15:31:20,747 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:20,748 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-12T15:31:20,749 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349 2024-11-12T15:31:20,755 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:20,755 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:20,758 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:33386 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:35073:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33386 dst: /127.0.0.1:35073 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:20,763 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-12T15:31:20,763 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-12T15:31:20,763 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-12T15:31:20,763 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775696_1015 (size=32) 2024-11-12T15:31:20,764 DEBUG [RS:1;a6f201fa5419:45955 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6b1b4fcd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:20,764 DEBUG [RS:0;a6f201fa5419:36945 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@673cb913, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:20,764 DEBUG [RS:2;a6f201fa5419:41281 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1cdba65, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:20,765 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:20,766 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:20,769 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-12T15:31:20,773 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-12T15:31:20,773 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,774 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:20,775 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-12T15:31:20,778 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-12T15:31:20,778 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,780 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:20,780 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-12T15:31:20,783 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-12T15:31:20,783 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;a6f201fa5419:45955 2024-11-12T15:31:20,783 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,784 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;a6f201fa5419:41281 2024-11-12T15:31:20,784 DEBUG [RS:0;a6f201fa5419:36945 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a6f201fa5419:36945 2024-11-12T15:31:20,784 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:20,785 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-12T15:31:20,787 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-12T15:31:20,787 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-12T15:31:20,787 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-12T15:31:20,787 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-12T15:31:20,787 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-12T15:31:20,787 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-12T15:31:20,787 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-12T15:31:20,787 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-12T15:31:20,787 DEBUG [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-12T15:31:20,788 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-12T15:31:20,788 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:20,789 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:20,790 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-12T15:31:20,791 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(2659): reportForDuty to master=a6f201fa5419,40305,1731425478640 with port=36945, startcode=1731425479435 2024-11-12T15:31:20,791 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(2659): reportForDuty to master=a6f201fa5419,40305,1731425478640 with port=45955, startcode=1731425479532 2024-11-12T15:31:20,791 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(2659): reportForDuty to master=a6f201fa5419,40305,1731425478640 with port=41281, startcode=1731425479581 2024-11-12T15:31:20,791 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740 2024-11-12T15:31:20,792 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740 2024-11-12T15:31:20,795 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-12T15:31:20,796 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-12T15:31:20,797 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-12T15:31:20,800 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-12T15:31:20,806 DEBUG [RS:1;a6f201fa5419:45955 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-12T15:31:20,806 DEBUG [RS:0;a6f201fa5419:36945 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-12T15:31:20,806 DEBUG [RS:2;a6f201fa5419:41281 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-12T15:31:20,814 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-12T15:31:20,815 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73416215, jitterRate=0.09398685395717621}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-12T15:31:20,820 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1731425480767Initializing all the Stores at 1731425480769 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425480769Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425480769Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425480769Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425480769Cleaning up temporary data from old regions at 1731425480796 (+27 ms)Region opened successfully at 1731425480820 (+24 ms) 2024-11-12T15:31:20,820 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-12T15:31:20,821 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-12T15:31:20,821 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-12T15:31:20,821 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-12T15:31:20,821 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-12T15:31:20,823 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-12T15:31:20,824 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731425480820Disabling compacts and flushes for region at 1731425480820Disabling writes for close at 1731425480821 (+1 ms)Writing region close event to WAL at 1731425480823 (+2 ms)Closed at 1731425480823 2024-11-12T15:31:20,835 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:20,836 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-12T15:31:20,842 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43497, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-11-12T15:31:20,842 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36689, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-11-12T15:31:20,842 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45659, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-11-12T15:31:20,845 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-12T15:31:20,850 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40305 {}] master.ServerManager(363): Checking decommissioned status of RegionServer a6f201fa5419,41281,1731425479581 2024-11-12T15:31:20,852 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40305 {}] master.ServerManager(517): Registering regionserver=a6f201fa5419,41281,1731425479581 2024-11-12T15:31:20,855 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-12T15:31:20,859 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-12T15:31:20,864 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40305 {}] master.ServerManager(363): Checking decommissioned status of RegionServer a6f201fa5419,36945,1731425479435 2024-11-12T15:31:20,864 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40305 {}] master.ServerManager(517): Registering regionserver=a6f201fa5419,36945,1731425479435 2024-11-12T15:31:20,867 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40305 {}] master.ServerManager(363): Checking decommissioned status of RegionServer a6f201fa5419,45955,1731425479532 2024-11-12T15:31:20,867 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40305 {}] master.ServerManager(517): Registering regionserver=a6f201fa5419,45955,1731425479532 2024-11-12T15:31:20,868 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349 2024-11-12T15:31:20,868 DEBUG [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349 2024-11-12T15:31:20,868 DEBUG [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35895 2024-11-12T15:31:20,868 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35895 2024-11-12T15:31:20,868 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-12T15:31:20,868 DEBUG [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-12T15:31:20,871 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349 2024-11-12T15:31:20,871 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35895 2024-11-12T15:31:20,871 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-12T15:31:20,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-12T15:31:20,925 DEBUG [RS:2;a6f201fa5419:41281 {}] zookeeper.ZKUtil(111): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a6f201fa5419,41281,1731425479581 2024-11-12T15:31:20,925 WARN [RS:2;a6f201fa5419:41281 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:20,925 INFO [RS:2;a6f201fa5419:41281 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-12T15:31:20,925 DEBUG [RS:1;a6f201fa5419:45955 {}] zookeeper.ZKUtil(111): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a6f201fa5419,45955,1731425479532 2024-11-12T15:31:20,925 WARN [RS:1;a6f201fa5419:45955 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:20,925 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,41281,1731425479581 2024-11-12T15:31:20,925 DEBUG [RS:0;a6f201fa5419:36945 {}] zookeeper.ZKUtil(111): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a6f201fa5419,36945,1731425479435 2024-11-12T15:31:20,925 INFO [RS:1;a6f201fa5419:45955 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-12T15:31:20,925 WARN [RS:0;a6f201fa5419:36945 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:20,926 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532 2024-11-12T15:31:20,926 INFO [RS:0;a6f201fa5419:36945 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-12T15:31:20,926 DEBUG [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,36945,1731425479435 2024-11-12T15:31:20,926 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a6f201fa5419,41281,1731425479581] 2024-11-12T15:31:20,926 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a6f201fa5419,36945,1731425479435] 2024-11-12T15:31:20,926 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a6f201fa5419,45955,1731425479532] 2024-11-12T15:31:20,952 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-12T15:31:20,952 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-12T15:31:20,952 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-12T15:31:20,970 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-12T15:31:20,970 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-12T15:31:20,971 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-12T15:31:20,976 INFO [RS:1;a6f201fa5419:45955 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-12T15:31:20,976 INFO [RS:2;a6f201fa5419:41281 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-12T15:31:20,976 INFO [RS:0;a6f201fa5419:36945 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-12T15:31:20,976 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,976 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,976 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,977 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-12T15:31:20,978 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-12T15:31:20,978 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-12T15:31:20,983 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-12T15:31:20,983 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-12T15:31:20,983 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-12T15:31:20,985 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,985 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,985 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:20,985 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,985 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,985 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,986 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:20,987 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:0;a6f201fa5419:36945 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:20,987 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:20,987 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:20,987 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:20,987 DEBUG [RS:2;a6f201fa5419:41281 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:20,987 DEBUG [RS:1;a6f201fa5419:45955 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:21,005 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,004 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41281,1731425479581-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,45955,1731425479532-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:21,005 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,006 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,006 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,006 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,36945,1731425479435-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:21,010 WARN [a6f201fa5419:40305 {}] assignment.AssignmentManager(2443): No servers available; cannot place 1 unassigned regions. 2024-11-12T15:31:21,027 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-12T15:31:21,027 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-12T15:31:21,029 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41281,1731425479581-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,029 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,36945,1731425479435-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,029 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,029 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,030 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.Replication(171): a6f201fa5419,36945,1731425479435 started 2024-11-12T15:31:21,030 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.Replication(171): a6f201fa5419,41281,1731425479581 started 2024-11-12T15:31:21,030 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-12T15:31:21,031 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,45955,1731425479532-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,031 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,031 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.Replication(171): a6f201fa5419,45955,1731425479532 started 2024-11-12T15:31:21,047 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,047 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,047 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1482): Serving as a6f201fa5419,41281,1731425479581, RpcServer on a6f201fa5419/172.17.0.2:41281, sessionid=0x1012fac209b0003 2024-11-12T15:31:21,047 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(1482): Serving as a6f201fa5419,36945,1731425479435, RpcServer on a6f201fa5419/172.17.0.2:36945, sessionid=0x1012fac209b0001 2024-11-12T15:31:21,048 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-12T15:31:21,048 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-12T15:31:21,048 DEBUG [RS:2;a6f201fa5419:41281 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a6f201fa5419,41281,1731425479581 2024-11-12T15:31:21,048 DEBUG [RS:0;a6f201fa5419:36945 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a6f201fa5419,36945,1731425479435 2024-11-12T15:31:21,048 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,36945,1731425479435' 2024-11-12T15:31:21,048 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,41281,1731425479581' 2024-11-12T15:31:21,048 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-12T15:31:21,048 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-12T15:31:21,049 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-12T15:31:21,049 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-12T15:31:21,050 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-12T15:31:21,050 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-12T15:31:21,050 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-12T15:31:21,050 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-12T15:31:21,050 DEBUG [RS:0;a6f201fa5419:36945 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a6f201fa5419,36945,1731425479435 2024-11-12T15:31:21,050 DEBUG [RS:2;a6f201fa5419:41281 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a6f201fa5419,41281,1731425479581 2024-11-12T15:31:21,050 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,36945,1731425479435' 2024-11-12T15:31:21,050 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,41281,1731425479581' 2024-11-12T15:31:21,050 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-12T15:31:21,050 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-12T15:31:21,050 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-12T15:31:21,050 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-12T15:31:21,051 DEBUG [RS:0;a6f201fa5419:36945 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-12T15:31:21,051 DEBUG [RS:2;a6f201fa5419:41281 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-12T15:31:21,051 INFO [RS:0;a6f201fa5419:36945 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-12T15:31:21,051 INFO [RS:2;a6f201fa5419:41281 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-12T15:31:21,051 INFO [RS:0;a6f201fa5419:36945 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-12T15:31:21,051 INFO [RS:2;a6f201fa5419:41281 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-12T15:31:21,053 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,053 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1482): Serving as a6f201fa5419,45955,1731425479532, RpcServer on a6f201fa5419/172.17.0.2:45955, sessionid=0x1012fac209b0002 2024-11-12T15:31:21,053 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-12T15:31:21,053 DEBUG [RS:1;a6f201fa5419:45955 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a6f201fa5419,45955,1731425479532 2024-11-12T15:31:21,053 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,45955,1731425479532' 2024-11-12T15:31:21,053 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-12T15:31:21,054 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-12T15:31:21,055 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-12T15:31:21,055 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-12T15:31:21,055 DEBUG [RS:1;a6f201fa5419:45955 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a6f201fa5419,45955,1731425479532 2024-11-12T15:31:21,055 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,45955,1731425479532' 2024-11-12T15:31:21,055 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-12T15:31:21,056 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-12T15:31:21,056 DEBUG [RS:1;a6f201fa5419:45955 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-12T15:31:21,057 INFO [RS:1;a6f201fa5419:45955 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-12T15:31:21,057 INFO [RS:1;a6f201fa5419:45955 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-12T15:31:21,156 INFO [RS:2;a6f201fa5419:41281 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-12T15:31:21,156 INFO [RS:0;a6f201fa5419:36945 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-12T15:31:21,157 INFO [RS:1;a6f201fa5419:45955 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-12T15:31:21,159 INFO [RS:0;a6f201fa5419:36945 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C36945%2C1731425479435, suffix=, logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,36945,1731425479435, archiveDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs, maxLogs=32 2024-11-12T15:31:21,159 INFO [RS:2;a6f201fa5419:41281 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C41281%2C1731425479581, suffix=, logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,41281,1731425479581, archiveDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs, maxLogs=32 2024-11-12T15:31:21,160 INFO [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C45955%2C1731425479532, suffix=, logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532, archiveDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs, maxLogs=32 2024-11-12T15:31:21,176 DEBUG [RS:0;a6f201fa5419:36945 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,36945,1731425479435/a6f201fa5419%2C36945%2C1731425479435.1731425481163, exclude list is [], retry=0 2024-11-12T15:31:21,181 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42999,DS-77fbb761-241d-485a-acf0-7881db3c0972,DISK] 2024-11-12T15:31:21,181 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35073,DS-9a38f403-ac87-4dad-bce7-330500803097,DISK] 2024-11-12T15:31:21,181 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37019,DS-8cd4ebbe-e241-4236-81a5-1900de5094c9,DISK] 2024-11-12T15:31:21,182 DEBUG [RS:2;a6f201fa5419:41281 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,41281,1731425479581/a6f201fa5419%2C41281%2C1731425479581.1731425481163, exclude list is [], retry=0 2024-11-12T15:31:21,182 DEBUG [RS:1;a6f201fa5419:45955 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532/a6f201fa5419%2C45955%2C1731425479532.1731425481164, exclude list is [], retry=0 2024-11-12T15:31:21,190 INFO [RS:0;a6f201fa5419:36945 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,36945,1731425479435/a6f201fa5419%2C36945%2C1731425479435.1731425481163 2024-11-12T15:31:21,191 DEBUG [RS:0;a6f201fa5419:36945 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42313:42313),(127.0.0.1/127.0.0.1:45141:45141),(127.0.0.1/127.0.0.1:37755:37755)] 2024-11-12T15:31:21,205 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42999,DS-77fbb761-241d-485a-acf0-7881db3c0972,DISK] 2024-11-12T15:31:21,205 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35073,DS-9a38f403-ac87-4dad-bce7-330500803097,DISK] 2024-11-12T15:31:21,206 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42999,DS-77fbb761-241d-485a-acf0-7881db3c0972,DISK] 2024-11-12T15:31:21,206 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35073,DS-9a38f403-ac87-4dad-bce7-330500803097,DISK] 2024-11-12T15:31:21,206 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37019,DS-8cd4ebbe-e241-4236-81a5-1900de5094c9,DISK] 2024-11-12T15:31:21,207 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37019,DS-8cd4ebbe-e241-4236-81a5-1900de5094c9,DISK] 2024-11-12T15:31:21,212 INFO [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532/a6f201fa5419%2C45955%2C1731425479532.1731425481164 2024-11-12T15:31:21,213 DEBUG [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42313:42313),(127.0.0.1/127.0.0.1:45141:45141),(127.0.0.1/127.0.0.1:37755:37755)] 2024-11-12T15:31:21,213 INFO [RS:2;a6f201fa5419:41281 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,41281,1731425479581/a6f201fa5419%2C41281%2C1731425479581.1731425481163 2024-11-12T15:31:21,214 DEBUG [RS:2;a6f201fa5419:41281 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:45141:45141),(127.0.0.1/127.0.0.1:42313:42313),(127.0.0.1/127.0.0.1:37755:37755)] 2024-11-12T15:31:21,264 DEBUG [a6f201fa5419:40305 {}] assignment.AssignmentManager(2464): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-12T15:31:21,276 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(204): Hosts are {a6f201fa5419=0} racks are {/default-rack=0} 2024-11-12T15:31:21,283 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-12T15:31:21,283 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-12T15:31:21,283 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-12T15:31:21,283 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-12T15:31:21,283 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-12T15:31:21,283 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-12T15:31:21,283 INFO [a6f201fa5419:40305 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-12T15:31:21,283 INFO [a6f201fa5419:40305 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-12T15:31:21,284 INFO [a6f201fa5419:40305 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-12T15:31:21,284 DEBUG [a6f201fa5419:40305 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-12T15:31:21,290 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a6f201fa5419,45955,1731425479532 2024-11-12T15:31:21,296 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a6f201fa5419,45955,1731425479532, state=OPENING 2024-11-12T15:31:21,362 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-12T15:31:21,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:21,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:21,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:21,373 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:21,375 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,375 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,376 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,376 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,379 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-12T15:31:21,382 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=a6f201fa5419,45955,1731425479532}] 2024-11-12T15:31:21,558 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-12T15:31:21,561 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56457, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-12T15:31:21,573 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-12T15:31:21,574 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-12T15:31:21,574 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-11-12T15:31:21,578 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C45955%2C1731425479532.meta, suffix=.meta, logDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532, archiveDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs, maxLogs=32 2024-11-12T15:31:21,594 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532/a6f201fa5419%2C45955%2C1731425479532.meta.1731425481580.meta, exclude list is [], retry=0 2024-11-12T15:31:21,598 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35073,DS-9a38f403-ac87-4dad-bce7-330500803097,DISK] 2024-11-12T15:31:21,598 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:42999,DS-77fbb761-241d-485a-acf0-7881db3c0972,DISK] 2024-11-12T15:31:21,598 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37019,DS-8cd4ebbe-e241-4236-81a5-1900de5094c9,DISK] 2024-11-12T15:31:21,601 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/WALs/a6f201fa5419,45955,1731425479532/a6f201fa5419%2C45955%2C1731425479532.meta.1731425481580.meta 2024-11-12T15:31:21,601 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:42313:42313),(127.0.0.1/127.0.0.1:45141:45141),(127.0.0.1/127.0.0.1:37755:37755)] 2024-11-12T15:31:21,601 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-12T15:31:21,603 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-12T15:31:21,605 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-12T15:31:21,609 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-12T15:31:21,613 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-12T15:31:21,613 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:21,614 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-12T15:31:21,614 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-12T15:31:21,617 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-12T15:31:21,618 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-12T15:31:21,618 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:21,619 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:21,619 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-12T15:31:21,621 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-12T15:31:21,621 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:21,622 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:21,622 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-12T15:31:21,623 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-12T15:31:21,623 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:21,624 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:21,624 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-12T15:31:21,625 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-12T15:31:21,626 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:21,626 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:21,626 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-12T15:31:21,628 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740 2024-11-12T15:31:21,630 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740 2024-11-12T15:31:21,632 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-12T15:31:21,632 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-12T15:31:21,634 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-12T15:31:21,636 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-12T15:31:21,638 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75359094, jitterRate=0.12293800711631775}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-12T15:31:21,638 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-12T15:31:21,639 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1731425481614Writing region info on filesystem at 1731425481614Initializing all the Stores at 1731425481616 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425481616Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425481617 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425481617Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425481617Cleaning up temporary data from old regions at 1731425481633 (+16 ms)Running coprocessor post-open hooks at 1731425481638 (+5 ms)Region opened successfully at 1731425481639 (+1 ms) 2024-11-12T15:31:21,646 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1731425481548 2024-11-12T15:31:21,674 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-12T15:31:21,674 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-12T15:31:21,676 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a6f201fa5419,45955,1731425479532 2024-11-12T15:31:21,679 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a6f201fa5419,45955,1731425479532, state=OPEN 2024-11-12T15:31:21,762 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:21,762 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:21,762 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:21,762 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:21,762 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,762 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,762 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,762 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:21,763 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=a6f201fa5419,45955,1731425479532 2024-11-12T15:31:21,770 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-12T15:31:21,770 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=a6f201fa5419,45955,1731425479532 in 382 msec 2024-11-12T15:31:21,777 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-12T15:31:21,777 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 928 msec 2024-11-12T15:31:21,778 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:21,778 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-12T15:31:21,796 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-12T15:31:21,797 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=a6f201fa5419,45955,1731425479532, seqNum=-1] 2024-11-12T15:31:21,813 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-12T15:31:21,815 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38257, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-12T15:31:21,834 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1850 sec 2024-11-12T15:31:21,834 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1731425481834, completionTime=-1 2024-11-12T15:31:21,837 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-12T15:31:21,837 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] assignment.AssignmentManager(1756): Joining cluster... 2024-11-12T15:31:21,861 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] assignment.AssignmentManager(1768): Number of RegionServers=3 2024-11-12T15:31:21,861 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1731425541861 2024-11-12T15:31:21,861 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1731425601861 2024-11-12T15:31:21,861 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] assignment.AssignmentManager(1775): Joined the cluster in 23 msec 2024-11-12T15:31:21,862 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-11-12T15:31:21,868 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,40305,1731425478640-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,869 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,40305,1731425478640-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,869 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,40305,1731425478640-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,870 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a6f201fa5419:40305, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,871 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,871 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,877 DEBUG [master/a6f201fa5419:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-12T15:31:21,896 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.201sec 2024-11-12T15:31:21,898 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-12T15:31:21,899 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-12T15:31:21,899 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-12T15:31:21,900 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-12T15:31:21,900 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-12T15:31:21,900 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,40305,1731425478640-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:21,901 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,40305,1731425478640-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-12T15:31:21,905 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-12T15:31:21,906 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-12T15:31:21,907 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,40305,1731425478640-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:21,926 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@674ba637, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-12T15:31:21,931 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-11-12T15:31:21,931 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-11-12T15:31:21,935 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request a6f201fa5419,40305,-1 for getting cluster id 2024-11-12T15:31:21,937 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-12T15:31:21,946 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '7a65a283-bcb2-4966-8aca-7e30e20bb4ab' 2024-11-12T15:31:21,948 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-12T15:31:21,948 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "7a65a283-bcb2-4966-8aca-7e30e20bb4ab" 2024-11-12T15:31:21,948 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3ad4d101, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-12T15:31:21,948 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [a6f201fa5419,40305,-1] 2024-11-12T15:31:21,950 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-12T15:31:21,952 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:21,953 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52912, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-12T15:31:21,956 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4dd3b555, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-12T15:31:21,956 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-12T15:31:21,962 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=a6f201fa5419,45955,1731425479532, seqNum=-1] 2024-11-12T15:31:21,963 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-12T15:31:21,965 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35760, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-12T15:31:21,988 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=a6f201fa5419,40305,1731425478640 2024-11-12T15:31:21,993 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-12T15:31:21,998 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is a6f201fa5419,40305,1731425478640 2024-11-12T15:31:22,001 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@7f05b924 2024-11-12T15:31:22,002 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-12T15:31:22,004 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52920, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-12T15:31:22,010 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-12T15:31:22,017 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-12T15:31:22,019 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-12T15:31:22,021 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-12T15:31:22,021 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:22,024 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-12T15:31:22,026 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:22,031 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:22,032 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:22,035 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:50524 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50524 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:22,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775680_1021 (size=392) 2024-11-12T15:31:22,140 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:22,351 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:22,443 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:22,449 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 6f386d04b52cd24a5cdbce5c4e8bcd90, NAME => 'TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349 2024-11-12T15:31:22,457 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:22,457 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:22,460 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:50538 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50538 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:22,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775664_1023 (size=51) 2024-11-12T15:31:22,465 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:22,465 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:22,465 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 6f386d04b52cd24a5cdbce5c4e8bcd90, disabling compactions & flushes 2024-11-12T15:31:22,465 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,465 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,466 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. after waiting 0 ms 2024-11-12T15:31:22,466 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,466 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,466 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 6f386d04b52cd24a5cdbce5c4e8bcd90: Waiting for close lock at 1731425482465Disabling compacts and flushes for region at 1731425482465Disabling writes for close at 1731425482466 (+1 ms)Writing region close event to WAL at 1731425482466Closed at 1731425482466 2024-11-12T15:31:22,468 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-12T15:31:22,473 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1731425482468"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1731425482468"}]},"ts":"1731425482468"} 2024-11-12T15:31:22,478 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-12T15:31:22,480 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-12T15:31:22,483 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731425482480"}]},"ts":"1731425482480"} 2024-11-12T15:31:22,487 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-12T15:31:22,488 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {a6f201fa5419=0} racks are {/default-rack=0} 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-12T15:31:22,489 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-12T15:31:22,489 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-12T15:31:22,489 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-12T15:31:22,489 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-12T15:31:22,490 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6f386d04b52cd24a5cdbce5c4e8bcd90, ASSIGN}] 2024-11-12T15:31:22,493 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6f386d04b52cd24a5cdbce5c4e8bcd90, ASSIGN 2024-11-12T15:31:22,494 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6f386d04b52cd24a5cdbce5c4e8bcd90, ASSIGN; state=OFFLINE, location=a6f201fa5419,41281,1731425479581; forceNewPlan=false, retain=false 2024-11-12T15:31:22,648 INFO [a6f201fa5419:40305 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-12T15:31:22,650 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=6f386d04b52cd24a5cdbce5c4e8bcd90, regionState=OPENING, regionLocation=a6f201fa5419,41281,1731425479581 2024-11-12T15:31:22,657 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6f386d04b52cd24a5cdbce5c4e8bcd90, ASSIGN because future has completed 2024-11-12T15:31:22,658 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 6f386d04b52cd24a5cdbce5c4e8bcd90, server=a6f201fa5419,41281,1731425479581}] 2024-11-12T15:31:22,660 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:22,813 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-12T15:31:22,818 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55525, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-12T15:31:22,827 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,827 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 6f386d04b52cd24a5cdbce5c4e8bcd90, NAME => 'TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90.', STARTKEY => '', ENDKEY => ''} 2024-11-12T15:31:22,828 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,828 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:22,828 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,828 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,830 INFO [StoreOpener-6f386d04b52cd24a5cdbce5c4e8bcd90-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,832 INFO [StoreOpener-6f386d04b52cd24a5cdbce5c4e8bcd90-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6f386d04b52cd24a5cdbce5c4e8bcd90 columnFamilyName cf 2024-11-12T15:31:22,832 DEBUG [StoreOpener-6f386d04b52cd24a5cdbce5c4e8bcd90-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:22,833 INFO [StoreOpener-6f386d04b52cd24a5cdbce5c4e8bcd90-1 {}] regionserver.HStore(327): Store=6f386d04b52cd24a5cdbce5c4e8bcd90/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:22,834 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,835 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,835 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,836 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,836 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,838 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,843 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-12T15:31:22,844 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 6f386d04b52cd24a5cdbce5c4e8bcd90; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61898427, jitterRate=-0.077641561627388}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-12T15:31:22,844 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:22,845 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 6f386d04b52cd24a5cdbce5c4e8bcd90: Running coprocessor pre-open hook at 1731425482828Writing region info on filesystem at 1731425482828Initializing all the Stores at 1731425482830 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425482830Cleaning up temporary data from old regions at 1731425482836 (+6 ms)Running coprocessor post-open hooks at 1731425482844 (+8 ms)Region opened successfully at 1731425482845 (+1 ms) 2024-11-12T15:31:22,847 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90., pid=6, masterSystemTime=1731425482812 2024-11-12T15:31:22,850 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,851 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:22,852 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=6f386d04b52cd24a5cdbce5c4e8bcd90, regionState=OPEN, openSeqNum=2, regionLocation=a6f201fa5419,41281,1731425479581 2024-11-12T15:31:22,855 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 6f386d04b52cd24a5cdbce5c4e8bcd90, server=a6f201fa5419,41281,1731425479581 because future has completed 2024-11-12T15:31:22,862 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-12T15:31:22,862 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 6f386d04b52cd24a5cdbce5c4e8bcd90, server=a6f201fa5419,41281,1731425479581 in 200 msec 2024-11-12T15:31:22,867 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-12T15:31:22,868 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6f386d04b52cd24a5cdbce5c4e8bcd90, ASSIGN in 372 msec 2024-11-12T15:31:22,869 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-12T15:31:22,869 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731425482869"}]},"ts":"1731425482869"} 2024-11-12T15:31:22,872 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-12T15:31:22,874 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-12T15:31:22,877 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 862 msec 2024-11-12T15:31:23,170 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:23,171 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-12T15:31:23,171 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-12T15:31:23,174 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-12T15:31:23,181 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-12T15:31:23,182 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-12T15:31:23,182 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-12T15:31:23,189 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90., hostname=a6f201fa5419,41281,1731425479581, seqNum=2] 2024-11-12T15:31:23,190 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-12T15:31:23,193 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44606, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-12T15:31:23,202 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-11-12T15:31:23,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-12T15:31:23,208 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-12T15:31:23,208 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:23,210 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-12T15:31:23,211 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-12T15:31:23,319 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:23,377 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41281 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-12T15:31:23,378 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,384 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 6f386d04b52cd24a5cdbce5c4e8bcd90 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-12T15:31:23,442 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/.tmp/cf/a264ad6aebfc438289de0c37d8c13be7 is 36, key is row/cf:cq/1731425483194/Put/seqid=0 2024-11-12T15:31:23,448 WARN [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:23,448 WARN [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:23,452 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_967676176_22 at /127.0.0.1:50558 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50558 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:23,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775648_1025 (size=4787) 2024-11-12T15:31:23,457 WARN [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:23,457 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/.tmp/cf/a264ad6aebfc438289de0c37d8c13be7 2024-11-12T15:31:23,498 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/.tmp/cf/a264ad6aebfc438289de0c37d8c13be7 as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/cf/a264ad6aebfc438289de0c37d8c13be7 2024-11-12T15:31:23,509 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/cf/a264ad6aebfc438289de0c37d8c13be7, entries=1, sequenceid=5, filesize=4.7 K 2024-11-12T15:31:23,515 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 6f386d04b52cd24a5cdbce5c4e8bcd90 in 132ms, sequenceid=5, compaction requested=false 2024-11-12T15:31:23,516 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-11-12T15:31:23,519 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 6f386d04b52cd24a5cdbce5c4e8bcd90: 2024-11-12T15:31:23,519 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,520 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-12T15:31:23,522 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-12T15:31:23,527 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-12T15:31:23,527 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 313 msec 2024-11-12T15:31:23,529 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:23,530 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 325 msec 2024-11-12T15:31:23,609 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775773_1004 (size=42) 2024-11-12T15:31:23,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775725_1010 (size=34) 2024-11-12T15:31:23,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775724_1010 (size=34) 2024-11-12T15:31:23,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775772_1004 (size=42) 2024-11-12T15:31:23,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775756_1006 (size=196) 2024-11-12T15:31:23,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775757_1006 (size=196) 2024-11-12T15:31:23,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775740_1008 (size=1189) 2024-11-12T15:31:23,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775709_1013 (size=1321) 2024-11-12T15:31:23,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775708_1013 (size=1321) 2024-11-12T15:31:23,661 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775741_1008 (size=1189) 2024-11-12T15:31:23,840 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=40305 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:23,840 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-12T15:31:23,854 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-12T15:31:23,855 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-12T15:31:23,855 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:23,861 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,862 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,862 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-12T15:31:23,862 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-12T15:31:23,862 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1575943754, stopped=false 2024-11-12T15:31:23,862 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=a6f201fa5419,40305,1731425478640 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:23,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:23,932 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:23,932 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-12T15:31:23,933 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-12T15:31:23,933 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:23,934 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:23,934 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:23,934 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,934 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:23,934 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:23,936 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'a6f201fa5419,36945,1731425479435' ***** 2024-11-12T15:31:23,936 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-12T15:31:23,936 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'a6f201fa5419,45955,1731425479532' ***** 2024-11-12T15:31:23,936 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-12T15:31:23,936 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'a6f201fa5419,41281,1731425479581' ***** 2024-11-12T15:31:23,937 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-12T15:31:23,937 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-12T15:31:23,937 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-12T15:31:23,937 INFO [RS:0;a6f201fa5419:36945 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-12T15:31:23,937 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-12T15:31:23,938 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-12T15:31:23,938 INFO [RS:0;a6f201fa5419:36945 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-12T15:31:23,938 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-12T15:31:23,938 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(959): stopping server a6f201fa5419,36945,1731425479435 2024-11-12T15:31:23,938 INFO [RS:1;a6f201fa5419:45955 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-12T15:31:23,938 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-12T15:31:23,938 INFO [RS:0;a6f201fa5419:36945 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:23,938 INFO [RS:1;a6f201fa5419:45955 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-12T15:31:23,939 INFO [RS:0;a6f201fa5419:36945 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;a6f201fa5419:36945. 2024-11-12T15:31:23,939 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(959): stopping server a6f201fa5419,45955,1731425479532 2024-11-12T15:31:23,939 INFO [RS:1;a6f201fa5419:45955 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:23,939 INFO [RS:2;a6f201fa5419:41281 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-12T15:31:23,939 DEBUG [RS:0;a6f201fa5419:36945 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:23,939 INFO [RS:2;a6f201fa5419:41281 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-12T15:31:23,939 INFO [RS:1;a6f201fa5419:45955 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;a6f201fa5419:45955. 2024-11-12T15:31:23,939 DEBUG [RS:0;a6f201fa5419:36945 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,940 DEBUG [RS:1;a6f201fa5419:45955 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:23,940 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(3091): Received CLOSE for 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:23,940 DEBUG [RS:1;a6f201fa5419:45955 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,940 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(976): stopping server a6f201fa5419,36945,1731425479435; all regions closed. 2024-11-12T15:31:23,940 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-12T15:31:23,940 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-12T15:31:23,940 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-12T15:31:23,940 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-12T15:31:23,940 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(959): stopping server a6f201fa5419,41281,1731425479581 2024-11-12T15:31:23,940 INFO [RS:2;a6f201fa5419:41281 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:23,941 INFO [RS:2;a6f201fa5419:41281 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;a6f201fa5419:41281. 2024-11-12T15:31:23,941 DEBUG [RS:2;a6f201fa5419:41281 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:23,941 DEBUG [RS:2;a6f201fa5419:41281 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,941 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-12T15:31:23,941 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-12T15:31:23,941 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 6f386d04b52cd24a5cdbce5c4e8bcd90, disabling compactions & flushes 2024-11-12T15:31:23,941 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1325): Online Regions={6f386d04b52cd24a5cdbce5c4e8bcd90=TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90.} 2024-11-12T15:31:23,941 INFO [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-12T15:31:23,941 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,941 DEBUG [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. after waiting 0 ms 2024-11-12T15:31:23,941 DEBUG [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1351): Waiting on 6f386d04b52cd24a5cdbce5c4e8bcd90 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-12T15:31:23,941 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-12T15:31:23,942 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-12T15:31:23,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_1073741826_1016 (size=93) 2024-11-12T15:31:23,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_1073741826_1016 (size=93) 2024-11-12T15:31:23,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_1073741826_1016 (size=93) 2024-11-12T15:31:23,952 DEBUG [RS:0;a6f201fa5419:36945 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs 2024-11-12T15:31:23,952 INFO [RS:0;a6f201fa5419:36945 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL a6f201fa5419%2C36945%2C1731425479435:(num 1731425481163) 2024-11-12T15:31:23,953 DEBUG [RS:0;a6f201fa5419:36945 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] hbase.ChoreService(370): Chore service for: regionserver/a6f201fa5419:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-12T15:31:23,953 INFO [regionserver/a6f201fa5419:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:23,953 INFO [RS:0;a6f201fa5419:36945 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:23,954 INFO [RS:0;a6f201fa5419:36945 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36945 2024-11-12T15:31:23,962 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/default/TestHBaseWalOnEC/6f386d04b52cd24a5cdbce5c4e8bcd90/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-12T15:31:23,965 INFO [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,965 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 6f386d04b52cd24a5cdbce5c4e8bcd90: Waiting for close lock at 1731425483941Running coprocessor pre-close hooks at 1731425483941Disabling compacts and flushes for region at 1731425483941Disabling writes for close at 1731425483941Writing region close event to WAL at 1731425483942 (+1 ms)Running coprocessor post-close hooks at 1731425483963 (+21 ms)Closed at 1731425483965 (+2 ms) 2024-11-12T15:31:23,965 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90. 2024-11-12T15:31:23,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-12T15:31:23,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a6f201fa5419,36945,1731425479435 2024-11-12T15:31:23,967 INFO [RS:0;a6f201fa5419:36945 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:23,967 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a6f201fa5419,36945,1731425479435] 2024-11-12T15:31:23,972 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/info/fa9e14200f92481c9b5454512a0bbd23 is 153, key is TestHBaseWalOnEC,,1731425482006.6f386d04b52cd24a5cdbce5c4e8bcd90./info:regioninfo/1731425482851/Put/seqid=0 2024-11-12T15:31:23,975 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:23,975 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:23,980 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1902369634_22 at /127.0.0.1:52452 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:42999:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52452 dst: /127.0.0.1:42999 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:23,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775632_1027 (size=6637) 2024-11-12T15:31:23,985 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:23,985 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/info/fa9e14200f92481c9b5454512a0bbd23 2024-11-12T15:31:23,988 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/a6f201fa5419,36945,1731425479435 already deleted, retry=false 2024-11-12T15:31:23,988 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; a6f201fa5419,36945,1731425479435 expired; onlineServers=2 2024-11-12T15:31:24,005 INFO [regionserver/a6f201fa5419:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-11-12T15:31:24,005 INFO [regionserver/a6f201fa5419:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-11-12T15:31:24,006 INFO [regionserver/a6f201fa5419:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-11-12T15:31:24,006 INFO [regionserver/a6f201fa5419:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-11-12T15:31:24,009 INFO [regionserver/a6f201fa5419:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:24,010 INFO [regionserver/a6f201fa5419:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:24,012 INFO [regionserver/a6f201fa5419:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:24,014 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/ns/66f7a55c43b6409c9a84d6a9a8495573 is 43, key is default/ns:d/1731425481819/Put/seqid=0 2024-11-12T15:31:24,016 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,016 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,020 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1902369634_22 at /127.0.0.1:50628 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50628 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:24,023 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775616_1029 (size=5153) 2024-11-12T15:31:24,024 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:24,024 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/ns/66f7a55c43b6409c9a84d6a9a8495573 2024-11-12T15:31:24,050 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/table/1fbf352355644c5b86062eb9673b4e20 is 52, key is TestHBaseWalOnEC/table:state/1731425482869/Put/seqid=0 2024-11-12T15:31:24,052 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,053 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,056 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1902369634_22 at /127.0.0.1:50642 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50642 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:24,059 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775600_1031 (size=5249) 2024-11-12T15:31:24,060 WARN [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:24,060 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/table/1fbf352355644c5b86062eb9673b4e20 2024-11-12T15:31:24,070 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/info/fa9e14200f92481c9b5454512a0bbd23 as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/info/fa9e14200f92481c9b5454512a0bbd23 2024-11-12T15:31:24,077 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,078 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36945-0x1012fac209b0001, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,078 INFO [RS:0;a6f201fa5419:36945 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:24,078 INFO [RS:0;a6f201fa5419:36945 {}] regionserver.HRegionServer(1031): Exiting; stopping=a6f201fa5419,36945,1731425479435; zookeeper connection closed. 2024-11-12T15:31:24,078 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@234e9b17 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@234e9b17 2024-11-12T15:31:24,080 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/info/fa9e14200f92481c9b5454512a0bbd23, entries=10, sequenceid=11, filesize=6.5 K 2024-11-12T15:31:24,082 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/ns/66f7a55c43b6409c9a84d6a9a8495573 as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/ns/66f7a55c43b6409c9a84d6a9a8495573 2024-11-12T15:31:24,092 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/ns/66f7a55c43b6409c9a84d6a9a8495573, entries=2, sequenceid=11, filesize=5.0 K 2024-11-12T15:31:24,093 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/.tmp/table/1fbf352355644c5b86062eb9673b4e20 as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/table/1fbf352355644c5b86062eb9673b4e20 2024-11-12T15:31:24,103 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/table/1fbf352355644c5b86062eb9673b4e20, entries=2, sequenceid=11, filesize=5.1 K 2024-11-12T15:31:24,104 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 163ms, sequenceid=11, compaction requested=false 2024-11-12T15:31:24,105 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-11-12T15:31:24,118 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-12T15:31:24,118 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-12T15:31:24,119 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-12T15:31:24,119 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731425483941Running coprocessor pre-close hooks at 1731425483941Disabling compacts and flushes for region at 1731425483941Disabling writes for close at 1731425483941Obtaining lock to block concurrent updates at 1731425483942 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1731425483942Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1731425483942Flushing stores of hbase:meta,,1.1588230740 at 1731425483943 (+1 ms)Flushing 1588230740/info: creating writer at 1731425483943Flushing 1588230740/info: appending metadata at 1731425483967 (+24 ms)Flushing 1588230740/info: closing flushed file at 1731425483967Flushing 1588230740/ns: creating writer at 1731425483997 (+30 ms)Flushing 1588230740/ns: appending metadata at 1731425484012 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1731425484012Flushing 1588230740/table: creating writer at 1731425484033 (+21 ms)Flushing 1588230740/table: appending metadata at 1731425484049 (+16 ms)Flushing 1588230740/table: closing flushed file at 1731425484049Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@58bcfeb9: reopening flushed file at 1731425484069 (+20 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@4c0def6: reopening flushed file at 1731425484081 (+12 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@73b21929: reopening flushed file at 1731425484092 (+11 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 163ms, sequenceid=11, compaction requested=false at 1731425484104 (+12 ms)Writing region close event to WAL at 1731425484106 (+2 ms)Running coprocessor post-close hooks at 1731425484118 (+12 ms)Closed at 1731425484119 (+1 ms) 2024-11-12T15:31:24,119 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-12T15:31:24,141 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(976): stopping server a6f201fa5419,45955,1731425479532; all regions closed. 2024-11-12T15:31:24,141 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(976): stopping server a6f201fa5419,41281,1731425479581; all regions closed. 2024-11-12T15:31:24,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_1073741829_1019 (size=2751) 2024-11-12T15:31:24,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_1073741829_1019 (size=2751) 2024-11-12T15:31:24,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_1073741829_1019 (size=2751) 2024-11-12T15:31:24,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_1073741827_1017 (size=1298) 2024-11-12T15:31:24,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_1073741827_1017 (size=1298) 2024-11-12T15:31:24,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_1073741827_1017 (size=1298) 2024-11-12T15:31:24,150 DEBUG [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs 2024-11-12T15:31:24,150 INFO [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL a6f201fa5419%2C45955%2C1731425479532.meta:.meta(num 1731425481580) 2024-11-12T15:31:24,152 DEBUG [RS:2;a6f201fa5419:41281 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs 2024-11-12T15:31:24,152 INFO [RS:2;a6f201fa5419:41281 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL a6f201fa5419%2C41281%2C1731425479581:(num 1731425481163) 2024-11-12T15:31:24,152 DEBUG [RS:2;a6f201fa5419:41281 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:24,152 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:24,152 INFO [RS:2;a6f201fa5419:41281 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:24,152 INFO [RS:2;a6f201fa5419:41281 {}] hbase.ChoreService(370): Chore service for: regionserver/a6f201fa5419:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:24,152 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-12T15:31:24,152 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-12T15:31:24,153 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-12T15:31:24,153 INFO [RS:2;a6f201fa5419:41281 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:24,153 INFO [RS:2;a6f201fa5419:41281 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41281 2024-11-12T15:31:24,153 INFO [regionserver/a6f201fa5419:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:24,153 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_1073741828_1018 (size=93) 2024-11-12T15:31:24,153 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_1073741828_1018 (size=93) 2024-11-12T15:31:24,153 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_1073741828_1018 (size=93) 2024-11-12T15:31:24,157 DEBUG [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/oldWALs 2024-11-12T15:31:24,157 INFO [RS:1;a6f201fa5419:45955 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL a6f201fa5419%2C45955%2C1731425479532:(num 1731425481164) 2024-11-12T15:31:24,157 DEBUG [RS:1;a6f201fa5419:45955 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:24,157 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:24,157 INFO [RS:1;a6f201fa5419:45955 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:24,158 INFO [RS:1;a6f201fa5419:45955 {}] hbase.ChoreService(370): Chore service for: regionserver/a6f201fa5419:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:24,158 INFO [RS:1;a6f201fa5419:45955 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:24,158 INFO [regionserver/a6f201fa5419:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:24,158 INFO [RS:1;a6f201fa5419:45955 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:45955 2024-11-12T15:31:24,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-12T15:31:24,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a6f201fa5419,41281,1731425479581 2024-11-12T15:31:24,166 INFO [RS:2;a6f201fa5419:41281 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:24,177 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a6f201fa5419,45955,1731425479532 2024-11-12T15:31:24,177 INFO [RS:1;a6f201fa5419:45955 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:24,177 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a6f201fa5419,41281,1731425479581] 2024-11-12T15:31:24,198 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/a6f201fa5419,41281,1731425479581 already deleted, retry=false 2024-11-12T15:31:24,198 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; a6f201fa5419,41281,1731425479581 expired; onlineServers=1 2024-11-12T15:31:24,198 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a6f201fa5419,45955,1731425479532] 2024-11-12T15:31:24,208 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/a6f201fa5419,45955,1731425479532 already deleted, retry=false 2024-11-12T15:31:24,209 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; a6f201fa5419,45955,1731425479532 expired; onlineServers=0 2024-11-12T15:31:24,209 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'a6f201fa5419,40305,1731425478640' ***** 2024-11-12T15:31:24,209 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-12T15:31:24,209 INFO [M:0;a6f201fa5419:40305 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:24,210 INFO [M:0;a6f201fa5419:40305 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:24,210 DEBUG [M:0;a6f201fa5419:40305 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-12T15:31:24,210 DEBUG [M:0;a6f201fa5419:40305 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-12T15:31:24,210 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-12T15:31:24,210 DEBUG [master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.large.0-1731425480731 {}] cleaner.HFileCleaner(306): Exit Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.large.0-1731425480731,5,FailOnTimeoutGroup] 2024-11-12T15:31:24,210 DEBUG [master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.small.0-1731425480732 {}] cleaner.HFileCleaner(306): Exit Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.small.0-1731425480732,5,FailOnTimeoutGroup] 2024-11-12T15:31:24,211 INFO [M:0;a6f201fa5419:40305 {}] hbase.ChoreService(370): Chore service for: master/a6f201fa5419:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:24,211 INFO [M:0;a6f201fa5419:40305 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:24,211 DEBUG [M:0;a6f201fa5419:40305 {}] master.HMaster(1795): Stopping service threads 2024-11-12T15:31:24,212 INFO [M:0;a6f201fa5419:40305 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-12T15:31:24,212 INFO [M:0;a6f201fa5419:40305 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-12T15:31:24,213 INFO [M:0;a6f201fa5419:40305 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-12T15:31:24,213 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-12T15:31:24,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:24,219 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:24,220 DEBUG [M:0;a6f201fa5419:40305 {}] zookeeper.ZKUtil(347): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-12T15:31:24,220 WARN [M:0;a6f201fa5419:40305 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-12T15:31:24,221 INFO [M:0;a6f201fa5419:40305 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/.lastflushedseqids 2024-11-12T15:31:24,234 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,234 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,236 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:59602 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:35073:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59602 dst: /127.0.0.1:35073 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:24,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775584_1033 (size=127) 2024-11-12T15:31:24,240 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:24,240 INFO [M:0;a6f201fa5419:40305 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-12T15:31:24,241 INFO [M:0;a6f201fa5419:40305 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-12T15:31:24,241 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-12T15:31:24,241 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:24,241 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:24,241 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-12T15:31:24,241 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:24,241 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.85 KB heapSize=34.13 KB 2024-11-12T15:31:24,264 DEBUG [M:0;a6f201fa5419:40305 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/5032f296aaf544fcbfad00276ad997af is 82, key is hbase:meta,,1/info:regioninfo/1731425481675/Put/seqid=0 2024-11-12T15:31:24,267 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,267 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,269 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:52474 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:42999:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52474 dst: /127.0.0.1:42999 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:24,275 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_-9223372036854775568_1035 (size=5672) 2024-11-12T15:31:24,276 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:24,276 INFO [M:0;a6f201fa5419:40305 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/5032f296aaf544fcbfad00276ad997af 2024-11-12T15:31:24,288 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,288 INFO [RS:2;a6f201fa5419:41281 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:24,288 INFO [RS:2;a6f201fa5419:41281 {}] regionserver.HRegionServer(1031): Exiting; stopping=a6f201fa5419,41281,1731425479581; zookeeper connection closed. 2024-11-12T15:31:24,288 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:41281-0x1012fac209b0003, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,288 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@24386b8a {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@24386b8a 2024-11-12T15:31:24,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,298 INFO [RS:1;a6f201fa5419:45955 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:24,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:45955-0x1012fac209b0002, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,298 INFO [RS:1;a6f201fa5419:45955 {}] regionserver.HRegionServer(1031): Exiting; stopping=a6f201fa5419,45955,1731425479532; zookeeper connection closed. 2024-11-12T15:31:24,299 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@25805e3d {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@25805e3d 2024-11-12T15:31:24,299 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-12T15:31:24,305 DEBUG [M:0;a6f201fa5419:40305 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/f010bc9189874438aeb940f8937b6cf2 is 749, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1731425482876/Put/seqid=0 2024-11-12T15:31:24,307 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,308 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,310 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:50654 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:37019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50654 dst: /127.0.0.1:37019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:24,314 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_-9223372036854775552_1037 (size=6441) 2024-11-12T15:31:24,315 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:24,315 INFO [M:0;a6f201fa5419:40305 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.17 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/f010bc9189874438aeb940f8937b6cf2 2024-11-12T15:31:24,338 DEBUG [M:0;a6f201fa5419:40305 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7d4b69ec1ec84b06987341080f6c86ec is 69, key is a6f201fa5419,36945,1731425479435/rs:state/1731425480865/Put/seqid=0 2024-11-12T15:31:24,341 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,341 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-12T15:31:24,343 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_2039910080_22 at /127.0.0.1:59612 [Receiving block BP-934991163-172.17.0.2-1731425474470:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:35073:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59612 dst: /127.0.0.1:35073 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-12T15:31:24,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_-9223372036854775536_1039 (size=5294) 2024-11-12T15:31:24,347 WARN [M:0;a6f201fa5419:40305 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-12T15:31:24,348 INFO [M:0;a6f201fa5419:40305 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7d4b69ec1ec84b06987341080f6c86ec 2024-11-12T15:31:24,356 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/5032f296aaf544fcbfad00276ad997af as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/5032f296aaf544fcbfad00276ad997af 2024-11-12T15:31:24,363 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/5032f296aaf544fcbfad00276ad997af, entries=8, sequenceid=72, filesize=5.5 K 2024-11-12T15:31:24,364 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/f010bc9189874438aeb940f8937b6cf2 as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/f010bc9189874438aeb940f8937b6cf2 2024-11-12T15:31:24,372 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/f010bc9189874438aeb940f8937b6cf2, entries=8, sequenceid=72, filesize=6.3 K 2024-11-12T15:31:24,373 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/7d4b69ec1ec84b06987341080f6c86ec as hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/7d4b69ec1ec84b06987341080f6c86ec 2024-11-12T15:31:24,381 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/7d4b69ec1ec84b06987341080f6c86ec, entries=3, sequenceid=72, filesize=5.2 K 2024-11-12T15:31:24,382 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 141ms, sequenceid=72, compaction requested=false 2024-11-12T15:31:24,383 INFO [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:24,383 DEBUG [M:0;a6f201fa5419:40305 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731425484241Disabling compacts and flushes for region at 1731425484241Disabling writes for close at 1731425484241Obtaining lock to block concurrent updates at 1731425484241Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1731425484241Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27492, getHeapSize=34888, getOffHeapSize=0, getCellsCount=85 at 1731425484242 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1731425484243 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1731425484243Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1731425484263 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1731425484263Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1731425484283 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1731425484304 (+21 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1731425484305 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1731425484323 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1731425484338 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1731425484338Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6dc904e8: reopening flushed file at 1731425484355 (+17 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@710cb272: reopening flushed file at 1731425484363 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@343883e7: reopening flushed file at 1731425484372 (+9 ms)Finished flush of dataSize ~26.85 KB/27492, heapSize ~33.84 KB/34648, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 141ms, sequenceid=72, compaction requested=false at 1731425484382 (+10 ms)Writing region close event to WAL at 1731425484383 (+1 ms)Closed at 1731425484383 2024-11-12T15:31:24,386 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35073 is added to blk_1073741825_1011 (size=32695) 2024-11-12T15:31:24,387 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37019 is added to blk_1073741825_1011 (size=32695) 2024-11-12T15:31:24,387 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42999 is added to blk_1073741825_1011 (size=32695) 2024-11-12T15:31:24,387 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:24,388 INFO [M:0;a6f201fa5419:40305 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-12T15:31:24,388 INFO [M:0;a6f201fa5419:40305 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:40305 2024-11-12T15:31:24,388 INFO [M:0;a6f201fa5419:40305 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:24,498 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,498 INFO [M:0;a6f201fa5419:40305 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:24,499 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40305-0x1012fac209b0000, quorum=127.0.0.1:58397, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:24,536 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@17f8e572{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:24,539 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11aaae40{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:24,539 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:24,540 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@14c1b227{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:24,540 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@28d0ee11{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:24,544 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-12T15:31:24,544 WARN [BP-934991163-172.17.0.2-1731425474470 heartbeating to localhost/127.0.0.1:35895 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-12T15:31:24,544 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-12T15:31:24,544 WARN [BP-934991163-172.17.0.2-1731425474470 heartbeating to localhost/127.0.0.1:35895 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-934991163-172.17.0.2-1731425474470 (Datanode Uuid 3beb0c95-28b9-4eb1-ada9-486403015f38) service to localhost/127.0.0.1:35895 2024-11-12T15:31:24,546 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data5/current/BP-934991163-172.17.0.2-1731425474470 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:24,546 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data6/current/BP-934991163-172.17.0.2-1731425474470 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:24,547 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-12T15:31:24,549 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@133f1bad{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:24,549 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@39860596{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:24,549 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:24,549 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@125705fb{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:24,550 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@50510811{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:24,551 WARN [BP-934991163-172.17.0.2-1731425474470 heartbeating to localhost/127.0.0.1:35895 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-12T15:31:24,551 WARN [BP-934991163-172.17.0.2-1731425474470 heartbeating to localhost/127.0.0.1:35895 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-934991163-172.17.0.2-1731425474470 (Datanode Uuid 6492489c-4b63-46ed-ad64-7d0260599c6f) service to localhost/127.0.0.1:35895 2024-11-12T15:31:24,551 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data3/current/BP-934991163-172.17.0.2-1731425474470 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:24,551 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data4/current/BP-934991163-172.17.0.2-1731425474470 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:24,552 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-12T15:31:24,552 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-12T15:31:24,552 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-12T15:31:24,557 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3990ff75{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:24,557 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@48731e1b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:24,557 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:24,557 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4f02cc61{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:24,558 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@179ed6d0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:24,559 WARN [BP-934991163-172.17.0.2-1731425474470 heartbeating to localhost/127.0.0.1:35895 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-12T15:31:24,559 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-12T15:31:24,559 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-12T15:31:24,559 WARN [BP-934991163-172.17.0.2-1731425474470 heartbeating to localhost/127.0.0.1:35895 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-934991163-172.17.0.2-1731425474470 (Datanode Uuid 4cb0ffa7-4ac3-40a3-b6d3-1556f7f964e1) service to localhost/127.0.0.1:35895 2024-11-12T15:31:24,559 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data1/current/BP-934991163-172.17.0.2-1731425474470 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:24,560 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/cluster_74ba54c1-41f8-4908-d583-1b69be4fe896/data/data2/current/BP-934991163-172.17.0.2-1731425474470 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:24,560 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-12T15:31:24,567 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5cb83937{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-12T15:31:24,568 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@69b5b273{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:24,568 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:24,568 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3cbd6fd{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:24,568 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@b03c34d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:24,576 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-12T15:31:24,603 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-12T15:31:24,609 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=88 (was 158), OpenFileDescriptor=445 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=190 (was 207), ProcessCount=11 (was 11), AvailableMemoryMB=8332 (was 8619) 2024-11-12T15:31:24,614 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=88, OpenFileDescriptor=445, MaxFileDescriptor=1048576, SystemLoadAverage=190, ProcessCount=11, AvailableMemoryMB=8332 2024-11-12T15:31:24,614 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.log.dir so I do NOT create it in target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/6e79b496-9ea0-58e0-32b9-eb7ced150ec6/hadoop.tmp.dir so I do NOT create it in target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf, deleteOnExit=true 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/test.cache.data in system properties and HBase conf 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.tmp.dir in system properties and HBase conf 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir in system properties and HBase conf 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-12T15:31:24,615 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-12T15:31:24,616 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-12T15:31:24,616 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-12T15:31:24,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/nfs.dump.dir in system properties and HBase conf 2024-11-12T15:31:24,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/java.io.tmpdir in system properties and HBase conf 2024-11-12T15:31:24,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-12T15:31:24,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-12T15:31:24,617 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-12T15:31:25,037 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:25,042 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:25,044 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:25,044 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:25,044 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-12T15:31:25,049 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:25,049 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@25cf301b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:25,049 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4ad7e349{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:25,143 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@248097e{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/java.io.tmpdir/jetty-localhost-45021-hadoop-hdfs-3_4_1-tests_jar-_-any-13812390205265910373/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-12T15:31:25,144 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@69db7ce7{HTTP/1.1, (http/1.1)}{localhost:45021} 2024-11-12T15:31:25,144 INFO [Time-limited test {}] server.Server(415): Started @12446ms 2024-11-12T15:31:25,404 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:25,408 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:25,409 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:25,409 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:25,409 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-12T15:31:25,409 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@75cf72b6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:25,410 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3d8004b1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:25,504 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@10c16ff9{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/java.io.tmpdir/jetty-localhost-42699-hadoop-hdfs-3_4_1-tests_jar-_-any-4947849203407476743/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:25,504 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2cb7157f{HTTP/1.1, (http/1.1)}{localhost:42699} 2024-11-12T15:31:25,504 INFO [Time-limited test {}] server.Server(415): Started @12806ms 2024-11-12T15:31:25,506 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-12T15:31:25,537 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:25,541 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:25,543 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:25,543 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:25,543 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-12T15:31:25,543 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7bcbe8bf{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:25,544 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@59bfc6ef{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:25,636 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3c51f4f3{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/java.io.tmpdir/jetty-localhost-43931-hadoop-hdfs-3_4_1-tests_jar-_-any-12114974474860379641/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:25,636 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4cd93a34{HTTP/1.1, (http/1.1)}{localhost:43931} 2024-11-12T15:31:25,636 INFO [Time-limited test {}] server.Server(415): Started @12938ms 2024-11-12T15:31:25,638 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-12T15:31:25,671 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-12T15:31:25,673 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-12T15:31:25,674 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-12T15:31:25,674 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-12T15:31:25,674 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-12T15:31:25,675 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6b33dae7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,AVAILABLE} 2024-11-12T15:31:25,675 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6ead761f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-12T15:31:25,769 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@17c6a72a{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/java.io.tmpdir/jetty-localhost-46467-hadoop-hdfs-3_4_1-tests_jar-_-any-11351362853096095713/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:25,770 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@51f8341d{HTTP/1.1, (http/1.1)}{localhost:46467} 2024-11-12T15:31:25,770 INFO [Time-limited test {}] server.Server(415): Started @13072ms 2024-11-12T15:31:25,771 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-12T15:31:26,640 WARN [Thread-566 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data2/current/BP-1078066073-172.17.0.2-1731425484641/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:26,640 WARN [Thread-565 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data1/current/BP-1078066073-172.17.0.2-1731425484641/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:26,660 WARN [Thread-506 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-12T15:31:26,663 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x63e680cee0128202 with lease ID 0xb14c1ae40361e807: Processing first storage report for DS-e3f571dc-8eb7-4ca1-b2b8-0251e5a62f73 from datanode DatanodeRegistration(127.0.0.1:40349, datanodeUuid=d11c1117-47ce-495d-9708-db9e0910f77a, infoPort=38265, infoSecurePort=0, ipcPort=42781, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641) 2024-11-12T15:31:26,663 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x63e680cee0128202 with lease ID 0xb14c1ae40361e807: from storage DS-e3f571dc-8eb7-4ca1-b2b8-0251e5a62f73 node DatanodeRegistration(127.0.0.1:40349, datanodeUuid=d11c1117-47ce-495d-9708-db9e0910f77a, infoPort=38265, infoSecurePort=0, ipcPort=42781, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:26,663 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x63e680cee0128202 with lease ID 0xb14c1ae40361e807: Processing first storage report for DS-f1deb3f2-049e-414e-8f2d-b94b6649ee5f from datanode DatanodeRegistration(127.0.0.1:40349, datanodeUuid=d11c1117-47ce-495d-9708-db9e0910f77a, infoPort=38265, infoSecurePort=0, ipcPort=42781, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641) 2024-11-12T15:31:26,663 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x63e680cee0128202 with lease ID 0xb14c1ae40361e807: from storage DS-f1deb3f2-049e-414e-8f2d-b94b6649ee5f node DatanodeRegistration(127.0.0.1:40349, datanodeUuid=d11c1117-47ce-495d-9708-db9e0910f77a, infoPort=38265, infoSecurePort=0, ipcPort=42781, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:26,996 WARN [Thread-577 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data3/current/BP-1078066073-172.17.0.2-1731425484641/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:26,996 WARN [Thread-578 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data4/current/BP-1078066073-172.17.0.2-1731425484641/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:27,014 WARN [Thread-529 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-12T15:31:27,017 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x274b9b9a696be8bc with lease ID 0xb14c1ae40361e808: Processing first storage report for DS-11c48ca5-5db5-44c1-ab30-5a5c8ca01eff from datanode DatanodeRegistration(127.0.0.1:42159, datanodeUuid=5acbd09d-044b-46bb-b65c-8899da80485b, infoPort=46373, infoSecurePort=0, ipcPort=46647, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641) 2024-11-12T15:31:27,017 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x274b9b9a696be8bc with lease ID 0xb14c1ae40361e808: from storage DS-11c48ca5-5db5-44c1-ab30-5a5c8ca01eff node DatanodeRegistration(127.0.0.1:42159, datanodeUuid=5acbd09d-044b-46bb-b65c-8899da80485b, infoPort=46373, infoSecurePort=0, ipcPort=46647, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:27,017 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x274b9b9a696be8bc with lease ID 0xb14c1ae40361e808: Processing first storage report for DS-325b6ffb-e29d-40ca-baf2-6bb70b950ba6 from datanode DatanodeRegistration(127.0.0.1:42159, datanodeUuid=5acbd09d-044b-46bb-b65c-8899da80485b, infoPort=46373, infoSecurePort=0, ipcPort=46647, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641) 2024-11-12T15:31:27,017 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x274b9b9a696be8bc with lease ID 0xb14c1ae40361e808: from storage DS-325b6ffb-e29d-40ca-baf2-6bb70b950ba6 node DatanodeRegistration(127.0.0.1:42159, datanodeUuid=5acbd09d-044b-46bb-b65c-8899da80485b, infoPort=46373, infoSecurePort=0, ipcPort=46647, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:27,110 WARN [Thread-589 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data6/current/BP-1078066073-172.17.0.2-1731425484641/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:27,110 WARN [Thread-588 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data5/current/BP-1078066073-172.17.0.2-1731425484641/current, will proceed with Du for space computation calculation, 2024-11-12T15:31:27,121 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-11-12T15:31:27,134 WARN [Thread-551 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-12T15:31:27,136 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc72ad1789a520f25 with lease ID 0xb14c1ae40361e809: Processing first storage report for DS-9c37b1d0-15ee-4973-a763-7ec06dc7ff1e from datanode DatanodeRegistration(127.0.0.1:34531, datanodeUuid=962cc0fc-f52e-4129-b24e-a4666e0ccdc2, infoPort=38599, infoSecurePort=0, ipcPort=41359, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641) 2024-11-12T15:31:27,137 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc72ad1789a520f25 with lease ID 0xb14c1ae40361e809: from storage DS-9c37b1d0-15ee-4973-a763-7ec06dc7ff1e node DatanodeRegistration(127.0.0.1:34531, datanodeUuid=962cc0fc-f52e-4129-b24e-a4666e0ccdc2, infoPort=38599, infoSecurePort=0, ipcPort=41359, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:27,137 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc72ad1789a520f25 with lease ID 0xb14c1ae40361e809: Processing first storage report for DS-849c4094-9880-4b35-8422-7c575192867f from datanode DatanodeRegistration(127.0.0.1:34531, datanodeUuid=962cc0fc-f52e-4129-b24e-a4666e0ccdc2, infoPort=38599, infoSecurePort=0, ipcPort=41359, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641) 2024-11-12T15:31:27,137 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc72ad1789a520f25 with lease ID 0xb14c1ae40361e809: from storage DS-849c4094-9880-4b35-8422-7c575192867f node DatanodeRegistration(127.0.0.1:34531, datanodeUuid=962cc0fc-f52e-4129-b24e-a4666e0ccdc2, infoPort=38599, infoSecurePort=0, ipcPort=41359, storageInfo=lv=-57;cid=testClusterID;nsid=1325246846;c=1731425484641), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-12T15:31:27,179 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-12T15:31:27,179 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-11-12T15:31:27,214 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6 2024-11-12T15:31:27,218 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/zookeeper_0, clientPort=59933, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-12T15:31:27,220 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=59933 2024-11-12T15:31:27,220 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,223 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741825_1001 (size=7) 2024-11-12T15:31:27,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741825_1001 (size=7) 2024-11-12T15:31:27,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741825_1001 (size=7) 2024-11-12T15:31:27,242 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587 with version=8 2024-11-12T15:31:27,242 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:35895/user/jenkins/test-data/85e69137-82b4-fc0e-3872-324d35bc3349/hbase-staging 2024-11-12T15:31:27,244 INFO [Time-limited test {}] client.ConnectionUtils(128): master/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:27,244 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,244 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,244 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:27,244 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,244 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:27,244 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-12T15:31:27,245 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:27,245 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:41733 2024-11-12T15:31:27,247 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:41733 connecting to ZooKeeper ensemble=127.0.0.1:59933 2024-11-12T15:31:27,307 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:417330x0, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:27,308 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:41733-0x1012fac451d0000 connected 2024-11-12T15:31:27,409 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,410 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,412 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:27,412 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587, hbase.cluster.distributed=false 2024-11-12T15:31:27,414 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:27,415 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41733 2024-11-12T15:31:27,415 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41733 2024-11-12T15:31:27,415 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41733 2024-11-12T15:31:27,415 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41733 2024-11-12T15:31:27,415 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41733 2024-11-12T15:31:27,430 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:27,430 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,430 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,430 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:27,431 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,431 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:27,431 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-12T15:31:27,431 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:27,431 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:35231 2024-11-12T15:31:27,433 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:35231 connecting to ZooKeeper ensemble=127.0.0.1:59933 2024-11-12T15:31:27,434 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,435 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,445 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:352310x0, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:27,446 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35231-0x1012fac451d0001 connected 2024-11-12T15:31:27,446 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:27,446 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-12T15:31:27,447 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-12T15:31:27,448 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-12T15:31:27,449 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:27,450 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35231 2024-11-12T15:31:27,450 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35231 2024-11-12T15:31:27,450 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35231 2024-11-12T15:31:27,451 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35231 2024-11-12T15:31:27,451 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35231 2024-11-12T15:31:27,468 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:27,468 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,468 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,468 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:27,468 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,469 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:27,469 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-12T15:31:27,469 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:27,469 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36601 2024-11-12T15:31:27,471 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:36601 connecting to ZooKeeper ensemble=127.0.0.1:59933 2024-11-12T15:31:27,471 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,473 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,488 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:366010x0, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:27,488 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:27,488 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36601-0x1012fac451d0002 connected 2024-11-12T15:31:27,489 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-12T15:31:27,490 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-12T15:31:27,490 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-12T15:31:27,492 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:27,493 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36601 2024-11-12T15:31:27,493 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36601 2024-11-12T15:31:27,494 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36601 2024-11-12T15:31:27,495 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36601 2024-11-12T15:31:27,495 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36601 2024-11-12T15:31:27,514 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/a6f201fa5419:0 server-side Connection retries=45 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-12T15:31:27,514 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-12T15:31:27,515 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39817 2024-11-12T15:31:27,516 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39817 connecting to ZooKeeper ensemble=127.0.0.1:59933 2024-11-12T15:31:27,517 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,518 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,530 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:398170x0, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-12T15:31:27,530 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:27,530 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39817-0x1012fac451d0003 connected 2024-11-12T15:31:27,531 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-12T15:31:27,531 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-12T15:31:27,532 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-12T15:31:27,533 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-12T15:31:27,534 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39817 2024-11-12T15:31:27,534 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39817 2024-11-12T15:31:27,534 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39817 2024-11-12T15:31:27,535 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39817 2024-11-12T15:31:27,535 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39817 2024-11-12T15:31:27,549 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a6f201fa5419:41733 2024-11-12T15:31:27,549 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,562 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,573 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-12T15:31:27,574 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a6f201fa5419,41733,1731425487244 from backup master directory 2024-11-12T15:31:27,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,587 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:27,587 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-12T15:31:27,587 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,593 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/hbase.id] with ID: 7c8fa3e3-b30e-40e6-880d-848e204d9f72 2024-11-12T15:31:27,593 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/.tmp/hbase.id 2024-11-12T15:31:27,601 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741826_1002 (size=42) 2024-11-12T15:31:27,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741826_1002 (size=42) 2024-11-12T15:31:27,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741826_1002 (size=42) 2024-11-12T15:31:27,603 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/.tmp/hbase.id]:[hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/hbase.id] 2024-11-12T15:31:27,619 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-12T15:31:27,619 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-12T15:31:27,621 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-11-12T15:31:27,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,629 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741827_1003 (size=196) 2024-11-12T15:31:27,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741827_1003 (size=196) 2024-11-12T15:31:27,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741827_1003 (size=196) 2024-11-12T15:31:27,640 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-12T15:31:27,641 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-12T15:31:27,641 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-12T15:31:27,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741828_1004 (size=1189) 2024-11-12T15:31:27,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741828_1004 (size=1189) 2024-11-12T15:31:27,659 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741828_1004 (size=1189) 2024-11-12T15:31:27,660 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store 2024-11-12T15:31:27,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741829_1005 (size=34) 2024-11-12T15:31:27,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741829_1005 (size=34) 2024-11-12T15:31:27,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741829_1005 (size=34) 2024-11-12T15:31:27,670 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:27,670 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-12T15:31:27,670 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:27,670 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:27,670 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-12T15:31:27,670 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:27,670 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:27,671 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731425487670Disabling compacts and flushes for region at 1731425487670Disabling writes for close at 1731425487670Writing region close event to WAL at 1731425487670Closed at 1731425487670 2024-11-12T15:31:27,672 WARN [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/.initializing 2024-11-12T15:31:27,672 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/WALs/a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,675 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C41733%2C1731425487244, suffix=, logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/WALs/a6f201fa5419,41733,1731425487244, archiveDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/oldWALs, maxLogs=10 2024-11-12T15:31:27,676 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a6f201fa5419%2C41733%2C1731425487244.1731425487676 2024-11-12T15:31:27,686 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/WALs/a6f201fa5419,41733,1731425487244/a6f201fa5419%2C41733%2C1731425487244.1731425487676 2024-11-12T15:31:27,688 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46373:46373),(127.0.0.1/127.0.0.1:38599:38599),(127.0.0.1/127.0.0.1:38265:38265)] 2024-11-12T15:31:27,691 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-12T15:31:27,691 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:27,691 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,691 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,693 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,695 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-12T15:31:27,695 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,695 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:27,696 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,697 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-12T15:31:27,697 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,698 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:27,698 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,700 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-12T15:31:27,700 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,701 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:27,701 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,703 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-12T15:31:27,703 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,703 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:27,703 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,704 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,705 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,707 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,707 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,708 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-12T15:31:27,709 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-12T15:31:27,713 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-12T15:31:27,713 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72290753, jitterRate=0.07721616327762604}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-12T15:31:27,714 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1731425487691Initializing all the Stores at 1731425487692 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425487692Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425487693 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425487693Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425487693Cleaning up temporary data from old regions at 1731425487707 (+14 ms)Region opened successfully at 1731425487714 (+7 ms) 2024-11-12T15:31:27,715 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-12T15:31:27,720 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@28ed2b64, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:27,722 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-12T15:31:27,722 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-12T15:31:27,722 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-12T15:31:27,722 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-12T15:31:27,723 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-11-12T15:31:27,723 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-11-12T15:31:27,723 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-12T15:31:27,726 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-12T15:31:27,727 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-12T15:31:27,734 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-12T15:31:27,735 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-12T15:31:27,736 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-12T15:31:27,745 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-12T15:31:27,746 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-12T15:31:27,747 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-12T15:31:27,755 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-12T15:31:27,756 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-12T15:31:27,766 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-12T15:31:27,769 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-12T15:31:27,776 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,787 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,788 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,788 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=a6f201fa5419,41733,1731425487244, sessionid=0x1012fac451d0000, setting cluster-up flag (Was=false) 2024-11-12T15:31:27,808 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,808 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,808 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,840 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-12T15:31:27,844 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:27,898 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-12T15:31:27,899 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a6f201fa5419,41733,1731425487244 2024-11-12T15:31:27,901 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-12T15:31:27,904 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:27,904 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-12T15:31:27,904 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-12T15:31:27,904 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a6f201fa5419,41733,1731425487244 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-12T15:31:27,906 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a6f201fa5419:0, corePoolSize=5, maxPoolSize=5 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a6f201fa5419:0, corePoolSize=10, maxPoolSize=10 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:27,907 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1731425517909 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-12T15:31:27,909 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-12T15:31:27,910 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:27,910 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-12T15:31:27,910 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:27,910 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-12T15:31:27,910 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-12T15:31:27,910 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-12T15:31:27,911 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-12T15:31:27,911 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-12T15:31:27,911 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.large.0-1731425487911,5,FailOnTimeoutGroup] 2024-11-12T15:31:27,911 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.small.0-1731425487911,5,FailOnTimeoutGroup] 2024-11-12T15:31:27,911 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:27,912 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-12T15:31:27,912 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:27,912 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:27,912 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,912 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-12T15:31:27,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741831_1007 (size=1321) 2024-11-12T15:31:27,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741831_1007 (size=1321) 2024-11-12T15:31:27,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741831_1007 (size=1321) 2024-11-12T15:31:27,923 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-12T15:31:27,924 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587 2024-11-12T15:31:27,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741832_1008 (size=32) 2024-11-12T15:31:27,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741832_1008 (size=32) 2024-11-12T15:31:27,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741832_1008 (size=32) 2024-11-12T15:31:27,934 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:27,936 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-12T15:31:27,937 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(746): ClusterId : 7c8fa3e3-b30e-40e6-880d-848e204d9f72 2024-11-12T15:31:27,937 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(746): ClusterId : 7c8fa3e3-b30e-40e6-880d-848e204d9f72 2024-11-12T15:31:27,937 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(746): ClusterId : 7c8fa3e3-b30e-40e6-880d-848e204d9f72 2024-11-12T15:31:27,937 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-12T15:31:27,937 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-12T15:31:27,937 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-12T15:31:27,937 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-12T15:31:27,937 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,938 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:27,938 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-12T15:31:27,939 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-12T15:31:27,940 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,940 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:27,940 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-12T15:31:27,941 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-12T15:31:27,941 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,942 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:27,942 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-12T15:31:27,944 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-12T15:31:27,944 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:27,944 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:27,945 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-12T15:31:27,946 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740 2024-11-12T15:31:27,946 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740 2024-11-12T15:31:27,948 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-12T15:31:27,948 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-12T15:31:27,949 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-12T15:31:27,951 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-12T15:31:27,954 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-12T15:31:27,954 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69893692, jitterRate=0.04149717092514038}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-12T15:31:27,955 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1731425487934Initializing all the Stores at 1731425487935 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425487935Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425487935Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425487935Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425487936 (+1 ms)Cleaning up temporary data from old regions at 1731425487948 (+12 ms)Region opened successfully at 1731425487955 (+7 ms) 2024-11-12T15:31:27,956 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-12T15:31:27,956 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-12T15:31:27,956 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-12T15:31:27,956 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-12T15:31:27,956 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-12T15:31:27,956 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-12T15:31:27,956 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-12T15:31:27,956 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-12T15:31:27,956 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-12T15:31:27,956 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-12T15:31:27,956 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-12T15:31:27,957 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-12T15:31:27,957 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731425487956Disabling compacts and flushes for region at 1731425487956Disabling writes for close at 1731425487956Writing region close event to WAL at 1731425487956Closed at 1731425487956 2024-11-12T15:31:27,958 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:27,959 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-12T15:31:27,959 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-12T15:31:27,961 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-12T15:31:27,962 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-12T15:31:27,977 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-12T15:31:27,978 DEBUG [RS:2;a6f201fa5419:39817 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@34eab060, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:27,978 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-12T15:31:27,978 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-12T15:31:27,978 DEBUG [RS:1;a6f201fa5419:36601 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@19067164, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:27,978 DEBUG [RS:0;a6f201fa5419:35231 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@34d2006b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a6f201fa5419/172.17.0.2:0 2024-11-12T15:31:27,989 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;a6f201fa5419:39817 2024-11-12T15:31:27,989 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-12T15:31:27,989 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-12T15:31:27,989 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-12T15:31:27,990 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(2659): reportForDuty to master=a6f201fa5419,41733,1731425487244 with port=39817, startcode=1731425487513 2024-11-12T15:31:27,990 DEBUG [RS:1;a6f201fa5419:36601 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;a6f201fa5419:36601 2024-11-12T15:31:27,990 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-12T15:31:27,990 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-12T15:31:27,990 DEBUG [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-12T15:31:27,990 DEBUG [RS:2;a6f201fa5419:39817 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-12T15:31:27,991 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(2659): reportForDuty to master=a6f201fa5419,41733,1731425487244 with port=36601, startcode=1731425487468 2024-11-12T15:31:27,991 DEBUG [RS:1;a6f201fa5419:36601 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-12T15:31:27,993 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35895, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-11-12T15:31:27,993 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41629, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-11-12T15:31:27,993 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a6f201fa5419:35231 2024-11-12T15:31:27,993 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-12T15:31:27,993 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-12T15:31:27,993 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41733 {}] master.ServerManager(363): Checking decommissioned status of RegionServer a6f201fa5419,39817,1731425487513 2024-11-12T15:31:27,993 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-12T15:31:27,993 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41733 {}] master.ServerManager(517): Registering regionserver=a6f201fa5419,39817,1731425487513 2024-11-12T15:31:27,994 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(2659): reportForDuty to master=a6f201fa5419,41733,1731425487244 with port=35231, startcode=1731425487430 2024-11-12T15:31:27,994 DEBUG [RS:0;a6f201fa5419:35231 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-12T15:31:27,996 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41733 {}] master.ServerManager(363): Checking decommissioned status of RegionServer a6f201fa5419,36601,1731425487468 2024-11-12T15:31:27,996 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587 2024-11-12T15:31:27,996 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41733 {}] master.ServerManager(517): Registering regionserver=a6f201fa5419,36601,1731425487468 2024-11-12T15:31:27,996 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35051 2024-11-12T15:31:27,996 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-12T15:31:27,996 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59657, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-11-12T15:31:27,998 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41733 {}] master.ServerManager(363): Checking decommissioned status of RegionServer a6f201fa5419,35231,1731425487430 2024-11-12T15:31:27,998 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41733 {}] master.ServerManager(517): Registering regionserver=a6f201fa5419,35231,1731425487430 2024-11-12T15:31:27,998 DEBUG [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587 2024-11-12T15:31:27,998 DEBUG [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35051 2024-11-12T15:31:27,998 DEBUG [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-12T15:31:28,000 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587 2024-11-12T15:31:28,000 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:35051 2024-11-12T15:31:28,000 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-12T15:31:28,008 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-12T15:31:28,045 DEBUG [RS:2;a6f201fa5419:39817 {}] zookeeper.ZKUtil(111): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a6f201fa5419,39817,1731425487513 2024-11-12T15:31:28,045 WARN [RS:2;a6f201fa5419:39817 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:28,045 INFO [RS:2;a6f201fa5419:39817 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-12T15:31:28,045 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a6f201fa5419,39817,1731425487513] 2024-11-12T15:31:28,045 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a6f201fa5419,36601,1731425487468] 2024-11-12T15:31:28,045 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a6f201fa5419,35231,1731425487430] 2024-11-12T15:31:28,045 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,39817,1731425487513 2024-11-12T15:31:28,045 DEBUG [RS:1;a6f201fa5419:36601 {}] zookeeper.ZKUtil(111): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a6f201fa5419,36601,1731425487468 2024-11-12T15:31:28,045 DEBUG [RS:0;a6f201fa5419:35231 {}] zookeeper.ZKUtil(111): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,045 WARN [RS:1;a6f201fa5419:36601 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:28,045 WARN [RS:0;a6f201fa5419:35231 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-12T15:31:28,045 INFO [RS:1;a6f201fa5419:36601 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-12T15:31:28,045 INFO [RS:0;a6f201fa5419:35231 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-12T15:31:28,045 DEBUG [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,36601,1731425487468 2024-11-12T15:31:28,045 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,051 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-12T15:31:28,051 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-12T15:31:28,051 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-12T15:31:28,054 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-12T15:31:28,054 INFO [RS:0;a6f201fa5419:35231 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-12T15:31:28,054 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,055 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-12T15:31:28,056 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-12T15:31:28,056 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,056 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,056 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,056 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,056 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,056 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,057 INFO [RS:1;a6f201fa5419:36601 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:28,057 DEBUG [RS:0;a6f201fa5419:35231 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:28,057 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,058 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-12T15:31:28,059 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-12T15:31:28,059 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-12T15:31:28,059 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,35231,1731425487430-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:28,060 INFO [RS:2;a6f201fa5419:39817 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,060 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,061 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:28,061 DEBUG [RS:1;a6f201fa5419:36601 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:28,064 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-12T15:31:28,065 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,065 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,065 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,065 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,065 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,065 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,36601,1731425487468-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:28,065 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-12T15:31:28,065 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,065 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a6f201fa5419:0, corePoolSize=2, maxPoolSize=2 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,066 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a6f201fa5419:0, corePoolSize=1, maxPoolSize=1 2024-11-12T15:31:28,067 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:28,067 DEBUG [RS:2;a6f201fa5419:39817 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0, corePoolSize=3, maxPoolSize=3 2024-11-12T15:31:28,071 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,071 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,071 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,071 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,071 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,071 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,39817,1731425487513-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:28,079 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-12T15:31:28,079 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-12T15:31:28,079 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,36601,1731425487468-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,079 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,35231,1731425487430-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,079 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,079 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,080 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.Replication(171): a6f201fa5419,35231,1731425487430 started 2024-11-12T15:31:28,080 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.Replication(171): a6f201fa5419,36601,1731425487468 started 2024-11-12T15:31:28,084 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-12T15:31:28,084 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,39817,1731425487513-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,084 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,084 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.Replication(171): a6f201fa5419,39817,1731425487513 started 2024-11-12T15:31:28,093 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,093 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,093 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1482): Serving as a6f201fa5419,35231,1731425487430, RpcServer on a6f201fa5419/172.17.0.2:35231, sessionid=0x1012fac451d0001 2024-11-12T15:31:28,093 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(1482): Serving as a6f201fa5419,36601,1731425487468, RpcServer on a6f201fa5419/172.17.0.2:36601, sessionid=0x1012fac451d0002 2024-11-12T15:31:28,093 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-12T15:31:28,093 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-12T15:31:28,093 DEBUG [RS:1;a6f201fa5419:36601 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a6f201fa5419,36601,1731425487468 2024-11-12T15:31:28,093 DEBUG [RS:0;a6f201fa5419:35231 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,093 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,36601,1731425487468' 2024-11-12T15:31:28,093 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,35231,1731425487430' 2024-11-12T15:31:28,093 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-12T15:31:28,093 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-12T15:31:28,094 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-12T15:31:28,094 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-12T15:31:28,094 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-12T15:31:28,094 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-12T15:31:28,094 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-12T15:31:28,094 DEBUG [RS:1;a6f201fa5419:36601 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a6f201fa5419,36601,1731425487468 2024-11-12T15:31:28,094 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-12T15:31:28,094 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,36601,1731425487468' 2024-11-12T15:31:28,094 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-12T15:31:28,094 DEBUG [RS:0;a6f201fa5419:35231 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,094 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,35231,1731425487430' 2024-11-12T15:31:28,094 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-12T15:31:28,095 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-12T15:31:28,095 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-12T15:31:28,095 DEBUG [RS:1;a6f201fa5419:36601 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-12T15:31:28,095 INFO [RS:1;a6f201fa5419:36601 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-12T15:31:28,095 INFO [RS:1;a6f201fa5419:36601 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-12T15:31:28,095 DEBUG [RS:0;a6f201fa5419:35231 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-12T15:31:28,095 INFO [RS:0;a6f201fa5419:35231 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-12T15:31:28,095 INFO [RS:0;a6f201fa5419:35231 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-12T15:31:28,097 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,098 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1482): Serving as a6f201fa5419,39817,1731425487513, RpcServer on a6f201fa5419/172.17.0.2:39817, sessionid=0x1012fac451d0003 2024-11-12T15:31:28,098 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-12T15:31:28,098 DEBUG [RS:2;a6f201fa5419:39817 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a6f201fa5419,39817,1731425487513 2024-11-12T15:31:28,098 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,39817,1731425487513' 2024-11-12T15:31:28,098 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-12T15:31:28,098 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-12T15:31:28,099 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-12T15:31:28,099 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-12T15:31:28,099 DEBUG [RS:2;a6f201fa5419:39817 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a6f201fa5419,39817,1731425487513 2024-11-12T15:31:28,099 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a6f201fa5419,39817,1731425487513' 2024-11-12T15:31:28,099 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-12T15:31:28,099 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-12T15:31:28,100 DEBUG [RS:2;a6f201fa5419:39817 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-12T15:31:28,100 INFO [RS:2;a6f201fa5419:39817 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-12T15:31:28,100 INFO [RS:2;a6f201fa5419:39817 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-12T15:31:28,112 WARN [a6f201fa5419:41733 {}] assignment.AssignmentManager(2443): No servers available; cannot place 1 unassigned regions. 2024-11-12T15:31:28,201 INFO [RS:1;a6f201fa5419:36601 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C36601%2C1731425487468, suffix=, logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,36601,1731425487468, archiveDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs, maxLogs=32 2024-11-12T15:31:28,201 INFO [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C35231%2C1731425487430, suffix=, logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,35231,1731425487430, archiveDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs, maxLogs=32 2024-11-12T15:31:28,204 INFO [RS:2;a6f201fa5419:39817 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C39817%2C1731425487513, suffix=, logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,39817,1731425487513, archiveDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs, maxLogs=32 2024-11-12T15:31:28,204 INFO [RS:0;a6f201fa5419:35231 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a6f201fa5419%2C35231%2C1731425487430.1731425488204 2024-11-12T15:31:28,205 INFO [RS:1;a6f201fa5419:36601 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a6f201fa5419%2C36601%2C1731425487468.1731425488205 2024-11-12T15:31:28,207 INFO [RS:2;a6f201fa5419:39817 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a6f201fa5419%2C39817%2C1731425487513.1731425488207 2024-11-12T15:31:28,217 INFO [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,35231,1731425487430/a6f201fa5419%2C35231%2C1731425487430.1731425488204 2024-11-12T15:31:28,217 INFO [RS:1;a6f201fa5419:36601 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,36601,1731425487468/a6f201fa5419%2C36601%2C1731425487468.1731425488205 2024-11-12T15:31:28,221 DEBUG [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46373:46373),(127.0.0.1/127.0.0.1:38265:38265),(127.0.0.1/127.0.0.1:38599:38599)] 2024-11-12T15:31:28,221 DEBUG [RS:1;a6f201fa5419:36601 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46373:46373),(127.0.0.1/127.0.0.1:38265:38265),(127.0.0.1/127.0.0.1:38599:38599)] 2024-11-12T15:31:28,221 INFO [RS:2;a6f201fa5419:39817 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,39817,1731425487513/a6f201fa5419%2C39817%2C1731425487513.1731425488207 2024-11-12T15:31:28,222 DEBUG [RS:2;a6f201fa5419:39817 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46373:46373),(127.0.0.1/127.0.0.1:38599:38599),(127.0.0.1/127.0.0.1:38265:38265)] 2024-11-12T15:31:28,363 DEBUG [a6f201fa5419:41733 {}] assignment.AssignmentManager(2464): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-12T15:31:28,363 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(204): Hosts are {a6f201fa5419=0} racks are {/default-rack=0} 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-12T15:31:28,367 INFO [a6f201fa5419:41733 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-12T15:31:28,367 INFO [a6f201fa5419:41733 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-12T15:31:28,367 INFO [a6f201fa5419:41733 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-12T15:31:28,367 DEBUG [a6f201fa5419:41733 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-12T15:31:28,368 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,371 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a6f201fa5419,35231,1731425487430, state=OPENING 2024-11-12T15:31:28,393 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-12T15:31:28,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:28,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:28,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:28,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:28,404 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-12T15:31:28,404 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,404 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,404 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,404 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=a6f201fa5419,35231,1731425487430}] 2024-11-12T15:31:28,404 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,560 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-12T15:31:28,563 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58857, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-12T15:31:28,569 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-12T15:31:28,570 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-12T15:31:28,573 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a6f201fa5419%2C35231%2C1731425487430.meta, suffix=.meta, logDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,35231,1731425487430, archiveDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs, maxLogs=32 2024-11-12T15:31:28,575 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a6f201fa5419%2C35231%2C1731425487430.meta.1731425488575.meta 2024-11-12T15:31:28,586 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/WALs/a6f201fa5419,35231,1731425487430/a6f201fa5419%2C35231%2C1731425487430.meta.1731425488575.meta 2024-11-12T15:31:28,589 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38599:38599),(127.0.0.1/127.0.0.1:38265:38265),(127.0.0.1/127.0.0.1:46373:46373)] 2024-11-12T15:31:28,592 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-12T15:31:28,593 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-12T15:31:28,593 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-12T15:31:28,593 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-12T15:31:28,593 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-12T15:31:28,593 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:28,593 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-12T15:31:28,593 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-12T15:31:28,595 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-12T15:31:28,596 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-12T15:31:28,596 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:28,597 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:28,597 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-12T15:31:28,598 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-12T15:31:28,598 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:28,598 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:28,598 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-12T15:31:28,599 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-12T15:31:28,599 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:28,599 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:28,599 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-12T15:31:28,600 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-12T15:31:28,600 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:28,601 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-12T15:31:28,601 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-12T15:31:28,602 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740 2024-11-12T15:31:28,603 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740 2024-11-12T15:31:28,605 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-12T15:31:28,605 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-12T15:31:28,605 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-12T15:31:28,607 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-12T15:31:28,608 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68681173, jitterRate=0.023429229855537415}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-12T15:31:28,608 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-12T15:31:28,609 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1731425488593Writing region info on filesystem at 1731425488593Initializing all the Stores at 1731425488594 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425488595 (+1 ms)Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425488595Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425488595Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731425488595Cleaning up temporary data from old regions at 1731425488605 (+10 ms)Running coprocessor post-open hooks at 1731425488608 (+3 ms)Region opened successfully at 1731425488609 (+1 ms) 2024-11-12T15:31:28,610 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1731425488559 2024-11-12T15:31:28,613 DEBUG [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-12T15:31:28,613 INFO [RS_OPEN_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-12T15:31:28,614 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,615 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a6f201fa5419,35231,1731425487430, state=OPEN 2024-11-12T15:31:28,624 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:28,624 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:28,624 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:28,624 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-12T15:31:28,625 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=a6f201fa5419,35231,1731425487430 2024-11-12T15:31:28,625 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,625 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,625 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,625 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-12T15:31:28,628 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-12T15:31:28,628 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=a6f201fa5419,35231,1731425487430 in 221 msec 2024-11-12T15:31:28,632 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-12T15:31:28,632 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 669 msec 2024-11-12T15:31:28,633 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-12T15:31:28,633 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-12T15:31:28,635 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-12T15:31:28,635 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=a6f201fa5419,35231,1731425487430, seqNum=-1] 2024-11-12T15:31:28,635 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-12T15:31:28,636 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56743, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-12T15:31:28,646 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 741 msec 2024-11-12T15:31:28,646 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1731425488646, completionTime=-1 2024-11-12T15:31:28,646 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-12T15:31:28,646 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] assignment.AssignmentManager(1756): Joining cluster... 2024-11-12T15:31:28,648 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] assignment.AssignmentManager(1768): Number of RegionServers=3 2024-11-12T15:31:28,648 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1731425548648 2024-11-12T15:31:28,648 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1731425608648 2024-11-12T15:31:28,648 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] assignment.AssignmentManager(1775): Joined the cluster in 2 msec 2024-11-12T15:31:28,649 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-11-12T15:31:28,649 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41733,1731425487244-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,649 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41733,1731425487244-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,649 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41733,1731425487244-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,649 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a6f201fa5419:41733, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,649 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,650 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,652 DEBUG [master/a6f201fa5419:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.068sec 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41733,1731425487244-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-12T15:31:28,655 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41733,1731425487244-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-12T15:31:28,658 DEBUG [master/a6f201fa5419:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-12T15:31:28,658 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-12T15:31:28,658 INFO [master/a6f201fa5419:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a6f201fa5419,41733,1731425487244-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-12T15:31:28,738 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1590b344, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-12T15:31:28,738 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request a6f201fa5419,41733,-1 for getting cluster id 2024-11-12T15:31:28,738 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-12T15:31:28,739 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '7c8fa3e3-b30e-40e6-880d-848e204d9f72' 2024-11-12T15:31:28,740 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-12T15:31:28,740 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "7c8fa3e3-b30e-40e6-880d-848e204d9f72" 2024-11-12T15:31:28,740 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4139576b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-12T15:31:28,740 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [a6f201fa5419,41733,-1] 2024-11-12T15:31:28,740 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-12T15:31:28,741 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:28,742 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46572, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-12T15:31:28,743 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7acc573, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-12T15:31:28,743 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-12T15:31:28,745 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=a6f201fa5419,35231,1731425487430, seqNum=-1] 2024-11-12T15:31:28,745 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-12T15:31:28,747 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41606, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-12T15:31:28,749 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=a6f201fa5419,41733,1731425487244 2024-11-12T15:31:28,750 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-12T15:31:28,751 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is a6f201fa5419,41733,1731425487244 2024-11-12T15:31:28,752 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@1273bfd8 2024-11-12T15:31:28,752 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-12T15:31:28,754 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46576, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-12T15:31:28,754 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-12T15:31:28,756 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-12T15:31:28,758 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-12T15:31:28,758 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:28,758 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-12T15:31:28,759 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:28,760 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-12T15:31:28,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741837_1013 (size=392) 2024-11-12T15:31:28,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741837_1013 (size=392) 2024-11-12T15:31:28,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741837_1013 (size=392) 2024-11-12T15:31:28,771 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 6870ab082070744ac59dc8d39a2aa52f, NAME => 'TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587 2024-11-12T15:31:28,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741838_1014 (size=51) 2024-11-12T15:31:28,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741838_1014 (size=51) 2024-11-12T15:31:28,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741838_1014 (size=51) 2024-11-12T15:31:28,781 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:28,781 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 6870ab082070744ac59dc8d39a2aa52f, disabling compactions & flushes 2024-11-12T15:31:28,781 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:28,781 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:28,781 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. after waiting 0 ms 2024-11-12T15:31:28,781 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:28,781 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:28,781 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 6870ab082070744ac59dc8d39a2aa52f: Waiting for close lock at 1731425488781Disabling compacts and flushes for region at 1731425488781Disabling writes for close at 1731425488781Writing region close event to WAL at 1731425488781Closed at 1731425488781 2024-11-12T15:31:28,783 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-12T15:31:28,784 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1731425488784"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1731425488784"}]},"ts":"1731425488784"} 2024-11-12T15:31:28,787 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-12T15:31:28,789 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-12T15:31:28,790 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731425488789"}]},"ts":"1731425488789"} 2024-11-12T15:31:28,793 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-12T15:31:28,793 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {a6f201fa5419=0} racks are {/default-rack=0} 2024-11-12T15:31:28,794 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-12T15:31:28,794 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-12T15:31:28,794 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-12T15:31:28,794 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-12T15:31:28,795 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-12T15:31:28,795 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-12T15:31:28,795 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-12T15:31:28,795 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-12T15:31:28,795 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-12T15:31:28,795 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-12T15:31:28,795 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6870ab082070744ac59dc8d39a2aa52f, ASSIGN}] 2024-11-12T15:31:28,797 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6870ab082070744ac59dc8d39a2aa52f, ASSIGN 2024-11-12T15:31:28,799 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6870ab082070744ac59dc8d39a2aa52f, ASSIGN; state=OFFLINE, location=a6f201fa5419,39817,1731425487513; forceNewPlan=false, retain=false 2024-11-12T15:31:28,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:28,949 INFO [a6f201fa5419:41733 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-12T15:31:28,950 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=6870ab082070744ac59dc8d39a2aa52f, regionState=OPENING, regionLocation=a6f201fa5419,39817,1731425487513 2024-11-12T15:31:28,955 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6870ab082070744ac59dc8d39a2aa52f, ASSIGN because future has completed 2024-11-12T15:31:28,956 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 6870ab082070744ac59dc8d39a2aa52f, server=a6f201fa5419,39817,1731425487513}] 2024-11-12T15:31:29,027 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-11-12T15:31:29,028 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-11-12T15:31:29,033 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-12T15:31:29,033 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-11-12T15:31:29,034 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-11-12T15:31:29,034 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-11-12T15:31:29,080 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:29,110 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-12T15:31:29,114 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51137, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-12T15:31:29,124 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,125 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 6870ab082070744ac59dc8d39a2aa52f, NAME => 'TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f.', STARTKEY => '', ENDKEY => ''} 2024-11-12T15:31:29,125 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,125 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-12T15:31:29,125 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,125 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,127 INFO [StoreOpener-6870ab082070744ac59dc8d39a2aa52f-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,129 INFO [StoreOpener-6870ab082070744ac59dc8d39a2aa52f-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6870ab082070744ac59dc8d39a2aa52f columnFamilyName cf 2024-11-12T15:31:29,129 DEBUG [StoreOpener-6870ab082070744ac59dc8d39a2aa52f-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-12T15:31:29,130 INFO [StoreOpener-6870ab082070744ac59dc8d39a2aa52f-1 {}] regionserver.HStore(327): Store=6870ab082070744ac59dc8d39a2aa52f/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-12T15:31:29,130 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,131 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,131 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,131 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,131 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,133 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,136 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-12T15:31:29,137 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 6870ab082070744ac59dc8d39a2aa52f; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68224570, jitterRate=0.016625314950942993}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-12T15:31:29,137 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,138 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 6870ab082070744ac59dc8d39a2aa52f: Running coprocessor pre-open hook at 1731425489125Writing region info on filesystem at 1731425489125Initializing all the Stores at 1731425489126 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731425489127 (+1 ms)Cleaning up temporary data from old regions at 1731425489132 (+5 ms)Running coprocessor post-open hooks at 1731425489137 (+5 ms)Region opened successfully at 1731425489138 (+1 ms) 2024-11-12T15:31:29,139 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f., pid=6, masterSystemTime=1731425489110 2024-11-12T15:31:29,143 DEBUG [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,143 INFO [RS_OPEN_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,144 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=6870ab082070744ac59dc8d39a2aa52f, regionState=OPEN, openSeqNum=2, regionLocation=a6f201fa5419,39817,1731425487513 2024-11-12T15:31:29,147 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 6870ab082070744ac59dc8d39a2aa52f, server=a6f201fa5419,39817,1731425487513 because future has completed 2024-11-12T15:31:29,153 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-12T15:31:29,155 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 6870ab082070744ac59dc8d39a2aa52f, server=a6f201fa5419,39817,1731425487513 in 193 msec 2024-11-12T15:31:29,158 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-12T15:31:29,158 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=6870ab082070744ac59dc8d39a2aa52f, ASSIGN in 358 msec 2024-11-12T15:31:29,159 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-12T15:31:29,159 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731425489159"}]},"ts":"1731425489159"} 2024-11-12T15:31:29,163 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-12T15:31:29,165 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-12T15:31:29,168 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 410 msec 2024-11-12T15:31:29,390 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-12T15:31:29,390 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-12T15:31:29,390 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-12T15:31:29,391 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-12T15:31:29,394 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-12T15:31:29,395 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-12T15:31:29,395 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-12T15:31:29,399 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f., hostname=a6f201fa5419,39817,1731425487513, seqNum=2] 2024-11-12T15:31:29,399 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-12T15:31:29,401 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38544, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-12T15:31:29,405 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-11-12T15:31:29,407 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-12T15:31:29,408 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:29,408 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-12T15:31:29,410 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-12T15:31:29,410 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-12T15:31:29,520 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:29,568 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39817 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-12T15:31:29,569 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,569 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 6870ab082070744ac59dc8d39a2aa52f 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-12T15:31:29,587 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/.tmp/cf/b5de29072c3044cea3b8f267334619be is 36, key is row/cf:cq/1731425489402/Put/seqid=0 2024-11-12T15:31:29,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741839_1015 (size=4787) 2024-11-12T15:31:29,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741839_1015 (size=4787) 2024-11-12T15:31:29,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741839_1015 (size=4787) 2024-11-12T15:31:29,596 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/.tmp/cf/b5de29072c3044cea3b8f267334619be 2024-11-12T15:31:29,605 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/.tmp/cf/b5de29072c3044cea3b8f267334619be as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/cf/b5de29072c3044cea3b8f267334619be 2024-11-12T15:31:29,614 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/cf/b5de29072c3044cea3b8f267334619be, entries=1, sequenceid=5, filesize=4.7 K 2024-11-12T15:31:29,615 INFO [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 6870ab082070744ac59dc8d39a2aa52f in 46ms, sequenceid=5, compaction requested=false 2024-11-12T15:31:29,615 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 6870ab082070744ac59dc8d39a2aa52f: 2024-11-12T15:31:29,615 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,616 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a6f201fa5419:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-12T15:31:29,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-12T15:31:29,622 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-12T15:31:29,622 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 208 msec 2024-11-12T15:31:29,625 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 218 msec 2024-11-12T15:31:29,730 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41733 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-12T15:31:29,731 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-12T15:31:29,760 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-12T15:31:29,760 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-12T15:31:29,761 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:29,761 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,761 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,761 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-12T15:31:29,761 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=695592510, stopped=false 2024-11-12T15:31:29,761 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=a6f201fa5419,41733,1731425487244 2024-11-12T15:31:29,761 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-12T15:31:29,835 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:29,835 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:29,835 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:29,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-12T15:31:29,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:29,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:29,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:29,836 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:29,836 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-12T15:31:29,836 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-12T15:31:29,837 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:29,838 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,838 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:29,838 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:29,838 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'a6f201fa5419,35231,1731425487430' ***** 2024-11-12T15:31:29,838 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-12T15:31:29,839 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'a6f201fa5419,36601,1731425487468' ***** 2024-11-12T15:31:29,839 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:29,839 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-12T15:31:29,839 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-12T15:31:29,839 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server 'a6f201fa5419,39817,1731425487513' ***** 2024-11-12T15:31:29,839 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-12T15:31:29,839 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-12T15:31:29,839 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-12T15:31:29,839 INFO [RS:0;a6f201fa5419:35231 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-12T15:31:29,839 INFO [RS:1;a6f201fa5419:36601 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-12T15:31:29,839 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-12T15:31:29,840 INFO [RS:0;a6f201fa5419:35231 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-12T15:31:29,840 INFO [RS:1;a6f201fa5419:36601 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-12T15:31:29,840 INFO [RS:2;a6f201fa5419:39817 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-12T15:31:29,840 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(959): stopping server a6f201fa5419,35231,1731425487430 2024-11-12T15:31:29,839 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-12T15:31:29,840 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(959): stopping server a6f201fa5419,36601,1731425487468 2024-11-12T15:31:29,840 INFO [RS:2;a6f201fa5419:39817 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-12T15:31:29,840 INFO [RS:0;a6f201fa5419:35231 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:29,840 INFO [RS:1;a6f201fa5419:36601 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:29,840 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-12T15:31:29,840 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(3091): Received CLOSE for 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,840 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-12T15:31:29,840 INFO [RS:0;a6f201fa5419:35231 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;a6f201fa5419:35231. 2024-11-12T15:31:29,840 INFO [RS:1;a6f201fa5419:36601 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;a6f201fa5419:36601. 2024-11-12T15:31:29,840 DEBUG [RS:0;a6f201fa5419:35231 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:29,840 DEBUG [RS:1;a6f201fa5419:36601 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:29,840 DEBUG [RS:1;a6f201fa5419:36601 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,840 DEBUG [RS:0;a6f201fa5419:35231 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,841 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(976): stopping server a6f201fa5419,36601,1731425487468; all regions closed. 2024-11-12T15:31:29,841 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-12T15:31:29,841 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(959): stopping server a6f201fa5419,39817,1731425487513 2024-11-12T15:31:29,841 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-12T15:31:29,841 INFO [RS:2;a6f201fa5419:39817 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:29,841 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-12T15:31:29,841 INFO [RS:2;a6f201fa5419:39817 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;a6f201fa5419:39817. 2024-11-12T15:31:29,841 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 6870ab082070744ac59dc8d39a2aa52f, disabling compactions & flushes 2024-11-12T15:31:29,841 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-12T15:31:29,841 INFO [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,841 DEBUG [RS:2;a6f201fa5419:39817 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-12T15:31:29,841 DEBUG [RS:2;a6f201fa5419:39817 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,841 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,842 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. after waiting 0 ms 2024-11-12T15:31:29,842 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-12T15:31:29,842 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,842 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1325): Online Regions={6870ab082070744ac59dc8d39a2aa52f=TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f.} 2024-11-12T15:31:29,842 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-12T15:31:29,842 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:29,842 DEBUG [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1351): Waiting on 6870ab082070744ac59dc8d39a2aa52f 2024-11-12T15:31:29,842 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-11-12T15:31:29,842 DEBUG [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-12T15:31:29,842 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:29,842 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-12T15:31:29,842 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-12T15:31:29,842 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:29,842 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-12T15:31:29,842 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:29,842 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-12T15:31:29,842 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-12T15:31:29,842 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:29,843 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-12T15:31:29,845 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741834_1010 (size=93) 2024-11-12T15:31:29,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741834_1010 (size=93) 2024-11-12T15:31:29,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741834_1010 (size=93) 2024-11-12T15:31:29,848 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/default/TestHBaseWalOnEC/6870ab082070744ac59dc8d39a2aa52f/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-12T15:31:29,849 INFO [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,849 DEBUG [RS:1;a6f201fa5419:36601 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs 2024-11-12T15:31:29,849 INFO [RS:1;a6f201fa5419:36601 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog a6f201fa5419%2C36601%2C1731425487468:(num 1731425488205) 2024-11-12T15:31:29,849 DEBUG [RS:1;a6f201fa5419:36601 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:29,849 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 6870ab082070744ac59dc8d39a2aa52f: Waiting for close lock at 1731425489841Running coprocessor pre-close hooks at 1731425489841Disabling compacts and flushes for region at 1731425489841Disabling writes for close at 1731425489842 (+1 ms)Writing region close event to WAL at 1731425489843 (+1 ms)Running coprocessor post-close hooks at 1731425489848 (+5 ms)Closed at 1731425489849 (+1 ms) 2024-11-12T15:31:29,849 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:29,849 INFO [RS:1;a6f201fa5419:36601 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:29,849 DEBUG [RS_CLOSE_REGION-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f. 2024-11-12T15:31:29,849 INFO [RS:1;a6f201fa5419:36601 {}] hbase.ChoreService(370): Chore service for: regionserver/a6f201fa5419:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:29,850 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-12T15:31:29,850 INFO [regionserver/a6f201fa5419:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:29,850 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-12T15:31:29,850 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-12T15:31:29,850 INFO [RS:1;a6f201fa5419:36601 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:29,850 INFO [RS:1;a6f201fa5419:36601 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36601 2024-11-12T15:31:29,860 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/info/9e59c45b55594a55bdf0b1b2ad990337 is 153, key is TestHBaseWalOnEC,,1731425488754.6870ab082070744ac59dc8d39a2aa52f./info:regioninfo/1731425489143/Put/seqid=0 2024-11-12T15:31:29,861 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a6f201fa5419,36601,1731425487468 2024-11-12T15:31:29,861 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-12T15:31:29,861 INFO [RS:1;a6f201fa5419:36601 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:29,861 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a6f201fa5419,36601,1731425487468] 2024-11-12T15:31:29,862 WARN [IPC Server handler 3 on default port 35051 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-12T15:31:29,862 WARN [IPC Server handler 3 on default port 35051 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-12T15:31:29,862 WARN [IPC Server handler 3 on default port 35051 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-12T15:31:29,866 INFO [regionserver/a6f201fa5419:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:29,868 INFO [regionserver/a6f201fa5419:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:29,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741840_1016 (size=6637) 2024-11-12T15:31:29,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741840_1016 (size=6637) 2024-11-12T15:31:29,871 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/info/9e59c45b55594a55bdf0b1b2ad990337 2024-11-12T15:31:29,873 INFO [regionserver/a6f201fa5419:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:29,882 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/a6f201fa5419,36601,1731425487468 already deleted, retry=false 2024-11-12T15:31:29,882 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; a6f201fa5419,36601,1731425487468 expired; onlineServers=2 2024-11-12T15:31:29,894 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/ns/64c2345e20654eb5b0a225cabab8fd84 is 43, key is default/ns:d/1731425488637/Put/seqid=0 2024-11-12T15:31:29,895 WARN [IPC Server handler 0 on default port 35051 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-12T15:31:29,895 WARN [IPC Server handler 0 on default port 35051 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-12T15:31:29,895 WARN [IPC Server handler 0 on default port 35051 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-12T15:31:29,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741841_1017 (size=5153) 2024-11-12T15:31:29,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741841_1017 (size=5153) 2024-11-12T15:31:29,900 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/ns/64c2345e20654eb5b0a225cabab8fd84 2024-11-12T15:31:29,921 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/table/9fac9b22e3ff40e48cd91d26a850814b is 52, key is TestHBaseWalOnEC/table:state/1731425489159/Put/seqid=0 2024-11-12T15:31:29,923 WARN [IPC Server handler 2 on default port 35051 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-12T15:31:29,923 WARN [IPC Server handler 2 on default port 35051 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-12T15:31:29,923 WARN [IPC Server handler 2 on default port 35051 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-12T15:31:29,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741842_1018 (size=5249) 2024-11-12T15:31:29,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741842_1018 (size=5249) 2024-11-12T15:31:29,928 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/table/9fac9b22e3ff40e48cd91d26a850814b 2024-11-12T15:31:29,935 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/info/9e59c45b55594a55bdf0b1b2ad990337 as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/info/9e59c45b55594a55bdf0b1b2ad990337 2024-11-12T15:31:29,943 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/info/9e59c45b55594a55bdf0b1b2ad990337, entries=10, sequenceid=11, filesize=6.5 K 2024-11-12T15:31:29,944 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/ns/64c2345e20654eb5b0a225cabab8fd84 as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/ns/64c2345e20654eb5b0a225cabab8fd84 2024-11-12T15:31:29,950 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/ns/64c2345e20654eb5b0a225cabab8fd84, entries=2, sequenceid=11, filesize=5.0 K 2024-11-12T15:31:29,952 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/.tmp/table/9fac9b22e3ff40e48cd91d26a850814b as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/table/9fac9b22e3ff40e48cd91d26a850814b 2024-11-12T15:31:29,958 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/table/9fac9b22e3ff40e48cd91d26a850814b, entries=2, sequenceid=11, filesize=5.1 K 2024-11-12T15:31:29,960 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 117ms, sequenceid=11, compaction requested=false 2024-11-12T15:31:29,965 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-12T15:31:29,966 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-12T15:31:29,966 INFO [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-12T15:31:29,966 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731425489842Running coprocessor pre-close hooks at 1731425489842Disabling compacts and flushes for region at 1731425489842Disabling writes for close at 1731425489842Obtaining lock to block concurrent updates at 1731425489843 (+1 ms)Preparing flush snapshotting stores in 1588230740 at 1731425489843Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1731425489843Flushing stores of hbase:meta,,1.1588230740 at 1731425489844 (+1 ms)Flushing 1588230740/info: creating writer at 1731425489844Flushing 1588230740/info: appending metadata at 1731425489860 (+16 ms)Flushing 1588230740/info: closing flushed file at 1731425489860Flushing 1588230740/ns: creating writer at 1731425489879 (+19 ms)Flushing 1588230740/ns: appending metadata at 1731425489893 (+14 ms)Flushing 1588230740/ns: closing flushed file at 1731425489893Flushing 1588230740/table: creating writer at 1731425489907 (+14 ms)Flushing 1588230740/table: appending metadata at 1731425489921 (+14 ms)Flushing 1588230740/table: closing flushed file at 1731425489921Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@71629247: reopening flushed file at 1731425489934 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@6d7b927d: reopening flushed file at 1731425489943 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1e7708db: reopening flushed file at 1731425489951 (+8 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 117ms, sequenceid=11, compaction requested=false at 1731425489960 (+9 ms)Writing region close event to WAL at 1731425489961 (+1 ms)Running coprocessor post-close hooks at 1731425489965 (+4 ms)Closed at 1731425489966 (+1 ms) 2024-11-12T15:31:29,966 DEBUG [RS_CLOSE_META-regionserver/a6f201fa5419:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-12T15:31:29,971 INFO [RS:1;a6f201fa5419:36601 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:29,971 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:29,971 INFO [RS:1;a6f201fa5419:36601 {}] regionserver.HRegionServer(1031): Exiting; stopping=a6f201fa5419,36601,1731425487468; zookeeper connection closed. 2024-11-12T15:31:29,971 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36601-0x1012fac451d0002, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:29,972 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7bc104 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7bc104 2024-11-12T15:31:30,042 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(976): stopping server a6f201fa5419,39817,1731425487513; all regions closed. 2024-11-12T15:31:30,042 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(976): stopping server a6f201fa5419,35231,1731425487430; all regions closed. 2024-11-12T15:31:30,043 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,043 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741835_1011 (size=1298) 2024-11-12T15:31:30,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741836_1012 (size=2751) 2024-11-12T15:31:30,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741836_1012 (size=2751) 2024-11-12T15:31:30,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741835_1011 (size=1298) 2024-11-12T15:31:30,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741835_1011 (size=1298) 2024-11-12T15:31:30,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741836_1012 (size=2751) 2024-11-12T15:31:30,049 DEBUG [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs 2024-11-12T15:31:30,049 INFO [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog a6f201fa5419%2C35231%2C1731425487430.meta:.meta(num 1731425488575) 2024-11-12T15:31:30,050 DEBUG [RS:2;a6f201fa5419:39817 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog a6f201fa5419%2C39817%2C1731425487513:(num 1731425488207) 2024-11-12T15:31:30,050 DEBUG [RS:2;a6f201fa5419:39817 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:30,050 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:30,050 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,050 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] hbase.ChoreService(370): Chore service for: regionserver/a6f201fa5419:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:30,050 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,050 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-12T15:31:30,050 INFO [regionserver/a6f201fa5419:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-12T15:31:30,050 INFO [RS:2;a6f201fa5419:39817 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:30,051 INFO [RS:2;a6f201fa5419:39817 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39817 2024-11-12T15:31:30,053 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741833_1009 (size=93) 2024-11-12T15:31:30,053 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741833_1009 (size=93) 2024-11-12T15:31:30,054 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741833_1009 (size=93) 2024-11-12T15:31:30,055 DEBUG [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/oldWALs 2024-11-12T15:31:30,055 INFO [RS:0;a6f201fa5419:35231 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog a6f201fa5419%2C35231%2C1731425487430:(num 1731425488204) 2024-11-12T15:31:30,055 DEBUG [RS:0;a6f201fa5419:35231 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-12T15:31:30,055 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.LeaseManager(133): Closed leases 2024-11-12T15:31:30,055 INFO [RS:0;a6f201fa5419:35231 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:30,056 INFO [RS:0;a6f201fa5419:35231 {}] hbase.ChoreService(370): Chore service for: regionserver/a6f201fa5419:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:30,056 INFO [RS:0;a6f201fa5419:35231 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:30,056 INFO [regionserver/a6f201fa5419:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:30,056 INFO [RS:0;a6f201fa5419:35231 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:35231 2024-11-12T15:31:30,077 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-12T15:31:30,077 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a6f201fa5419,39817,1731425487513 2024-11-12T15:31:30,077 INFO [RS:2;a6f201fa5419:39817 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:30,077 ERROR [Time-limited test-EventThread {}] zookeeper.ClientCnxn$EventThread(581): Error while calling watcher. java.util.concurrent.RejectedExecutionException: Task org.apache.hadoop.hbase.trace.TraceUtil$$Lambda$370/0x00007efe1c8f5a38@1d9a5e55 rejected from java.util.concurrent.ThreadPoolExecutor@5898c382[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 14] at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2065) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:833) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1365) ~[?:?] at java.util.concurrent.Executors$DelegatedExecutorService.execute(Executors.java:721) ~[?:?] at org.apache.hadoop.hbase.zookeeper.ZKWatcher.process(ZKWatcher.java:613) ~[hbase-zookeeper-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:579) ~[zookeeper-3.8.4.jar:3.8.4] at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:554) ~[zookeeper-3.8.4.jar:3.8.4] 2024-11-12T15:31:30,088 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a6f201fa5419,35231,1731425487430 2024-11-12T15:31:30,088 INFO [RS:0;a6f201fa5419:35231 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:30,098 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a6f201fa5419,39817,1731425487513] 2024-11-12T15:31:30,119 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/a6f201fa5419,39817,1731425487513 already deleted, retry=false 2024-11-12T15:31:30,119 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; a6f201fa5419,39817,1731425487513 expired; onlineServers=1 2024-11-12T15:31:30,119 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a6f201fa5419,35231,1731425487430] 2024-11-12T15:31:30,130 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/a6f201fa5419,35231,1731425487430 already deleted, retry=false 2024-11-12T15:31:30,130 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; a6f201fa5419,35231,1731425487430 expired; onlineServers=0 2024-11-12T15:31:30,130 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master 'a6f201fa5419,41733,1731425487244' ***** 2024-11-12T15:31:30,130 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-12T15:31:30,130 INFO [M:0;a6f201fa5419:41733 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-12T15:31:30,130 INFO [M:0;a6f201fa5419:41733 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-12T15:31:30,130 DEBUG [M:0;a6f201fa5419:41733 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-12T15:31:30,130 DEBUG [M:0;a6f201fa5419:41733 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-12T15:31:30,130 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-12T15:31:30,130 DEBUG [master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.large.0-1731425487911 {}] cleaner.HFileCleaner(306): Exit Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.large.0-1731425487911,5,FailOnTimeoutGroup] 2024-11-12T15:31:30,130 DEBUG [master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.small.0-1731425487911 {}] cleaner.HFileCleaner(306): Exit Thread[master/a6f201fa5419:0:becomeActiveMaster-HFileCleaner.small.0-1731425487911,5,FailOnTimeoutGroup] 2024-11-12T15:31:30,130 INFO [M:0;a6f201fa5419:41733 {}] hbase.ChoreService(370): Chore service for: master/a6f201fa5419:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-12T15:31:30,131 INFO [M:0;a6f201fa5419:41733 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-12T15:31:30,131 DEBUG [M:0;a6f201fa5419:41733 {}] master.HMaster(1795): Stopping service threads 2024-11-12T15:31:30,131 INFO [M:0;a6f201fa5419:41733 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-12T15:31:30,131 INFO [M:0;a6f201fa5419:41733 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-12T15:31:30,131 INFO [M:0;a6f201fa5419:41733 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-12T15:31:30,131 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-12T15:31:30,140 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-12T15:31:30,140 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-12T15:31:30,140 DEBUG [M:0;a6f201fa5419:41733 {}] zookeeper.ZKUtil(347): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-12T15:31:30,140 WARN [M:0;a6f201fa5419:41733 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-12T15:31:30,141 INFO [M:0;a6f201fa5419:41733 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/.lastflushedseqids 2024-11-12T15:31:30,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741843_1019 (size=127) 2024-11-12T15:31:30,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741843_1019 (size=127) 2024-11-12T15:31:30,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741843_1019 (size=127) 2024-11-12T15:31:30,198 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:30,198 INFO [RS:2;a6f201fa5419:39817 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:30,198 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39817-0x1012fac451d0003, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:30,199 INFO [RS:2;a6f201fa5419:39817 {}] regionserver.HRegionServer(1031): Exiting; stopping=a6f201fa5419,39817,1731425487513; zookeeper connection closed. 2024-11-12T15:31:30,199 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7ee7e3b9 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7ee7e3b9 2024-11-12T15:31:30,209 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:30,209 INFO [RS:0;a6f201fa5419:35231 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:30,209 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35231-0x1012fac451d0001, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:30,209 INFO [RS:0;a6f201fa5419:35231 {}] regionserver.HRegionServer(1031): Exiting; stopping=a6f201fa5419,35231,1731425487430; zookeeper connection closed. 2024-11-12T15:31:30,210 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7c871b3d {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7c871b3d 2024-11-12T15:31:30,210 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-12T15:31:30,551 INFO [M:0;a6f201fa5419:41733 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-12T15:31:30,552 INFO [M:0;a6f201fa5419:41733 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-12T15:31:30,552 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-12T15:31:30,552 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:30,552 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:30,552 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-12T15:31:30,552 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:30,553 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.83 KB heapSize=34.12 KB 2024-11-12T15:31:30,577 DEBUG [M:0;a6f201fa5419:41733 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/143f4c32988742a0abde17452e8087e8 is 82, key is hbase:meta,,1/info:regioninfo/1731425488614/Put/seqid=0 2024-11-12T15:31:30,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741844_1020 (size=5672) 2024-11-12T15:31:30,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741844_1020 (size=5672) 2024-11-12T15:31:30,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741844_1020 (size=5672) 2024-11-12T15:31:30,585 INFO [M:0;a6f201fa5419:41733 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/143f4c32988742a0abde17452e8087e8 2024-11-12T15:31:30,606 DEBUG [M:0;a6f201fa5419:41733 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/224bb0085d454baa9ddc20ac84f56ae5 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1731425489167/Put/seqid=0 2024-11-12T15:31:30,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741845_1021 (size=6439) 2024-11-12T15:31:30,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741845_1021 (size=6439) 2024-11-12T15:31:30,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741845_1021 (size=6439) 2024-11-12T15:31:30,613 INFO [M:0;a6f201fa5419:41733 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/224bb0085d454baa9ddc20ac84f56ae5 2024-11-12T15:31:30,634 DEBUG [M:0;a6f201fa5419:41733 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/21f577028c2e422e8decd344e136aa92 is 69, key is a6f201fa5419,35231,1731425487430/rs:state/1731425487998/Put/seqid=0 2024-11-12T15:31:30,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741846_1022 (size=5294) 2024-11-12T15:31:30,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741846_1022 (size=5294) 2024-11-12T15:31:30,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741846_1022 (size=5294) 2024-11-12T15:31:30,641 INFO [M:0;a6f201fa5419:41733 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/21f577028c2e422e8decd344e136aa92 2024-11-12T15:31:30,648 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/143f4c32988742a0abde17452e8087e8 as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/143f4c32988742a0abde17452e8087e8 2024-11-12T15:31:30,655 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/143f4c32988742a0abde17452e8087e8, entries=8, sequenceid=72, filesize=5.5 K 2024-11-12T15:31:30,656 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/224bb0085d454baa9ddc20ac84f56ae5 as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/224bb0085d454baa9ddc20ac84f56ae5 2024-11-12T15:31:30,664 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/224bb0085d454baa9ddc20ac84f56ae5, entries=8, sequenceid=72, filesize=6.3 K 2024-11-12T15:31:30,665 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/21f577028c2e422e8decd344e136aa92 as hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/21f577028c2e422e8decd344e136aa92 2024-11-12T15:31:30,672 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:35051/user/jenkins/test-data/34ad3da4-b6bf-c78d-68f4-7f63b2ae2587/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/21f577028c2e422e8decd344e136aa92, entries=3, sequenceid=72, filesize=5.2 K 2024-11-12T15:31:30,674 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 122ms, sequenceid=72, compaction requested=false 2024-11-12T15:31:30,676 INFO [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-12T15:31:30,676 DEBUG [M:0;a6f201fa5419:41733 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731425490552Disabling compacts and flushes for region at 1731425490552Disabling writes for close at 1731425490552Obtaining lock to block concurrent updates at 1731425490553 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1731425490553Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27471, getHeapSize=34872, getOffHeapSize=0, getCellsCount=85 at 1731425490553Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1731425490554 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1731425490554Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1731425490577 (+23 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1731425490577Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1731425490592 (+15 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1731425490606 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1731425490606Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1731425490619 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1731425490633 (+14 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1731425490633Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@ed1ef13: reopening flushed file at 1731425490647 (+14 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@43cace05: reopening flushed file at 1731425490655 (+8 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@49284fa1: reopening flushed file at 1731425490664 (+9 ms)Finished flush of dataSize ~26.83 KB/27471, heapSize ~33.82 KB/34632, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 122ms, sequenceid=72, compaction requested=false at 1731425490674 (+10 ms)Writing region close event to WAL at 1731425490675 (+1 ms)Closed at 1731425490676 (+1 ms) 2024-11-12T15:31:30,676 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,676 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,676 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,677 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,677 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-12T15:31:30,679 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40349 is added to blk_1073741830_1006 (size=32674) 2024-11-12T15:31:30,679 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42159 is added to blk_1073741830_1006 (size=32674) 2024-11-12T15:31:30,679 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34531 is added to blk_1073741830_1006 (size=32674) 2024-11-12T15:31:30,680 INFO [M:0;a6f201fa5419:41733 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-12T15:31:30,680 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-12T15:31:30,680 INFO [M:0;a6f201fa5419:41733 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:41733 2024-11-12T15:31:30,680 INFO [M:0;a6f201fa5419:41733 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-12T15:31:30,830 INFO [M:0;a6f201fa5419:41733 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-12T15:31:30,830 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:30,830 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41733-0x1012fac451d0000, quorum=127.0.0.1:59933, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-12T15:31:30,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@17c6a72a{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:30,833 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@51f8341d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:30,833 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:30,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6ead761f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:30,833 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6b33dae7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:30,834 WARN [BP-1078066073-172.17.0.2-1731425484641 heartbeating to localhost/127.0.0.1:35051 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-12T15:31:30,834 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-12T15:31:30,834 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-12T15:31:30,834 WARN [BP-1078066073-172.17.0.2-1731425484641 heartbeating to localhost/127.0.0.1:35051 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1078066073-172.17.0.2-1731425484641 (Datanode Uuid 962cc0fc-f52e-4129-b24e-a4666e0ccdc2) service to localhost/127.0.0.1:35051 2024-11-12T15:31:30,835 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data5/current/BP-1078066073-172.17.0.2-1731425484641 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:30,835 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data6/current/BP-1078066073-172.17.0.2-1731425484641 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:30,835 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-12T15:31:30,837 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3c51f4f3{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:30,837 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4cd93a34{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:30,837 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:30,838 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@59bfc6ef{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:30,838 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7bcbe8bf{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:30,839 WARN [BP-1078066073-172.17.0.2-1731425484641 heartbeating to localhost/127.0.0.1:35051 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-12T15:31:30,839 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-12T15:31:30,839 WARN [BP-1078066073-172.17.0.2-1731425484641 heartbeating to localhost/127.0.0.1:35051 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1078066073-172.17.0.2-1731425484641 (Datanode Uuid 5acbd09d-044b-46bb-b65c-8899da80485b) service to localhost/127.0.0.1:35051 2024-11-12T15:31:30,839 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-12T15:31:30,839 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data3/current/BP-1078066073-172.17.0.2-1731425484641 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:30,839 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data4/current/BP-1078066073-172.17.0.2-1731425484641 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:30,839 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-12T15:31:30,841 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@10c16ff9{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-12T15:31:30,841 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2cb7157f{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:30,841 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:30,842 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3d8004b1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:30,842 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@75cf72b6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:30,844 WARN [BP-1078066073-172.17.0.2-1731425484641 heartbeating to localhost/127.0.0.1:35051 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-12T15:31:30,844 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-12T15:31:30,844 WARN [BP-1078066073-172.17.0.2-1731425484641 heartbeating to localhost/127.0.0.1:35051 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1078066073-172.17.0.2-1731425484641 (Datanode Uuid d11c1117-47ce-495d-9708-db9e0910f77a) service to localhost/127.0.0.1:35051 2024-11-12T15:31:30,844 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-12T15:31:30,845 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data1/current/BP-1078066073-172.17.0.2-1731425484641 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:30,845 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/cluster_5ee12f30-d4f5-ccfc-5b06-1e43182f73cf/data/data2/current/BP-1078066073-172.17.0.2-1731425484641 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-12T15:31:30,845 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-12T15:31:30,851 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@248097e{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-12T15:31:30,852 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@69db7ce7{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-12T15:31:30,852 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-12T15:31:30,852 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4ad7e349{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-12T15:31:30,852 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@25cf301b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3@2/hbase-server/target/test-data/d961fa3e-02da-4481-7005-2326ce89a9b6/hadoop.log.dir/,STOPPED} 2024-11-12T15:31:30,861 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-12T15:31:30,897 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-12T15:31:30,903 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=148 (was 88) - Thread LEAK? -, OpenFileDescriptor=518 (was 445) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=199 (was 190) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=8152 (was 8332)