2024-11-08 08:51:05,848 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-08 08:51:05,860 main DEBUG Took 0.010176 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-11-08 08:51:05,860 main DEBUG PluginManager 'Core' found 129 plugins 2024-11-08 08:51:05,860 main DEBUG PluginManager 'Level' found 0 plugins 2024-11-08 08:51:05,861 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-11-08 08:51:05,863 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,879 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-11-08 08:51:05,890 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,892 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,893 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,894 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,894 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,895 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,896 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,896 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,897 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,897 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,898 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,899 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,900 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,900 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,901 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,901 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,902 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,902 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,903 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,903 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,904 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,904 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,905 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,905 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-11-08 08:51:05,906 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,906 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-11-08 08:51:05,908 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-11-08 08:51:05,910 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-11-08 08:51:05,912 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-11-08 08:51:05,913 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-11-08 08:51:05,914 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-11-08 08:51:05,915 main DEBUG PluginManager 'Converter' found 47 plugins 2024-11-08 08:51:05,926 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-11-08 08:51:05,929 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-11-08 08:51:05,931 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-11-08 08:51:05,932 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-11-08 08:51:05,932 main DEBUG createAppenders(={Console}) 2024-11-08 08:51:05,933 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-11-08 08:51:05,934 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-11-08 08:51:05,934 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-11-08 08:51:05,935 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-11-08 08:51:05,935 main DEBUG OutputStream closed 2024-11-08 08:51:05,936 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-11-08 08:51:05,936 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-11-08 08:51:05,936 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-11-08 08:51:06,018 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-11-08 08:51:06,020 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-11-08 08:51:06,021 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-11-08 08:51:06,022 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-11-08 08:51:06,022 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-11-08 08:51:06,023 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-11-08 08:51:06,023 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-11-08 08:51:06,023 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-11-08 08:51:06,024 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-11-08 08:51:06,024 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-11-08 08:51:06,024 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-11-08 08:51:06,025 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-11-08 08:51:06,025 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-11-08 08:51:06,025 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-11-08 08:51:06,026 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-11-08 08:51:06,026 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-11-08 08:51:06,026 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-11-08 08:51:06,027 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-11-08 08:51:06,029 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-08 08:51:06,029 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-logging/target/hbase-logging-3.0.0-beta-2-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-11-08 08:51:06,029 main DEBUG Shutdown hook enabled. Registering a new one. 2024-11-08 08:51:06,030 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-11-08T08:51:06,043 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-11-08 08:51:06,045 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-11-08 08:51:06,046 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-11-08T08:51:06,317 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989 2024-11-08T08:51:06,347 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155, deleteOnExit=true 2024-11-08T08:51:06,349 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/test.cache.data in system properties and HBase conf 2024-11-08T08:51:06,349 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.tmp.dir in system properties and HBase conf 2024-11-08T08:51:06,350 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir in system properties and HBase conf 2024-11-08T08:51:06,351 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-08T08:51:06,351 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-08T08:51:06,351 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-08T08:51:06,430 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-11-08T08:51:06,513 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-08T08:51:06,517 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-08T08:51:06,518 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-08T08:51:06,518 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-08T08:51:06,519 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-08T08:51:06,519 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-08T08:51:06,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-08T08:51:06,520 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-08T08:51:06,521 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-08T08:51:06,521 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-08T08:51:06,522 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/nfs.dump.dir in system properties and HBase conf 2024-11-08T08:51:06,522 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/java.io.tmpdir in system properties and HBase conf 2024-11-08T08:51:06,522 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-08T08:51:06,523 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-08T08:51:06,523 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-08T08:51:07,632 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-11-08T08:51:07,699 INFO [Time-limited test {}] log.Log(170): Logging initialized @2483ms to org.eclipse.jetty.util.log.Slf4jLog 2024-11-08T08:51:07,766 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:07,825 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:07,845 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:07,845 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:07,846 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:07,857 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:07,861 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@760c69c0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:07,862 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@ce709a8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:08,030 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@62d6efd9{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/java.io.tmpdir/jetty-localhost-36415-hadoop-hdfs-3_4_1-tests_jar-_-any-15351453196341071149/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-08T08:51:08,045 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@353d35a1{HTTP/1.1, (http/1.1)}{localhost:36415} 2024-11-08T08:51:08,045 INFO [Time-limited test {}] server.Server(415): Started @2830ms 2024-11-08T08:51:08,827 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:08,834 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:08,835 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:08,835 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:08,835 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:08,836 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3a5de9e4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:08,837 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@69893329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:08,933 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1b97a472{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/java.io.tmpdir/jetty-localhost-43743-hadoop-hdfs-3_4_1-tests_jar-_-any-16462237776161236251/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:08,934 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3722a29b{HTTP/1.1, (http/1.1)}{localhost:43743} 2024-11-08T08:51:08,934 INFO [Time-limited test {}] server.Server(415): Started @3719ms 2024-11-08T08:51:08,982 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-08T08:51:09,083 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:09,089 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:09,090 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:09,091 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:09,091 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:09,092 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@510fec09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:09,093 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40eb7053{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:09,220 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@353955e9{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/java.io.tmpdir/jetty-localhost-41435-hadoop-hdfs-3_4_1-tests_jar-_-any-18026431035951378919/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:09,221 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@11738cd8{HTTP/1.1, (http/1.1)}{localhost:41435} 2024-11-08T08:51:09,221 INFO [Time-limited test {}] server.Server(415): Started @4006ms 2024-11-08T08:51:09,224 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-08T08:51:09,263 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:09,270 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:09,271 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:09,272 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:09,272 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-11-08T08:51:09,273 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@16cd567f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:09,274 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5822645a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:09,386 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3114ae69{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/java.io.tmpdir/jetty-localhost-38287-hadoop-hdfs-3_4_1-tests_jar-_-any-13674805309475842115/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:09,386 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3c70a874{HTTP/1.1, (http/1.1)}{localhost:38287} 2024-11-08T08:51:09,387 INFO [Time-limited test {}] server.Server(415): Started @4171ms 2024-11-08T08:51:09,389 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-08T08:51:10,665 WARN [Thread-120 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data1/current/BP-1794552092-172.17.0.3-1731055867092/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:10,665 WARN [Thread-121 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data2/current/BP-1794552092-172.17.0.3-1731055867092/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:10,665 WARN [Thread-124 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data3/current/BP-1794552092-172.17.0.3-1731055867092/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:10,666 WARN [Thread-125 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data4/current/BP-1794552092-172.17.0.3-1731055867092/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:10,697 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-08T08:51:10,697 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-08T08:51:10,745 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x18c19ac146f3a1c4 with lease ID 0x17f0e4e04a887630: Processing first storage report for DS-42b2faa6-d7f4-4bc0-8e8f-61fde4bb9e11 from datanode DatanodeRegistration(127.0.0.1:39689, datanodeUuid=060e9f99-6979-4f24-a4ff-dab4db6244fb, infoPort=41629, infoSecurePort=0, ipcPort=41349, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092) 2024-11-08T08:51:10,746 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x18c19ac146f3a1c4 with lease ID 0x17f0e4e04a887630: from storage DS-42b2faa6-d7f4-4bc0-8e8f-61fde4bb9e11 node DatanodeRegistration(127.0.0.1:39689, datanodeUuid=060e9f99-6979-4f24-a4ff-dab4db6244fb, infoPort=41629, infoSecurePort=0, ipcPort=41349, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-08T08:51:10,746 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xcf68aa8e09be2c6a with lease ID 0x17f0e4e04a887631: Processing first storage report for DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd from datanode DatanodeRegistration(127.0.0.1:37237, datanodeUuid=4192a270-ddb6-4969-9470-f4d8ec183401, infoPort=43503, infoSecurePort=0, ipcPort=43027, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092) 2024-11-08T08:51:10,746 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xcf68aa8e09be2c6a with lease ID 0x17f0e4e04a887631: from storage DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd node DatanodeRegistration(127.0.0.1:37237, datanodeUuid=4192a270-ddb6-4969-9470-f4d8ec183401, infoPort=43503, infoSecurePort=0, ipcPort=43027, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:10,747 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x18c19ac146f3a1c4 with lease ID 0x17f0e4e04a887630: Processing first storage report for DS-f3b75a7c-ffc5-4c18-b304-2b195de27997 from datanode DatanodeRegistration(127.0.0.1:39689, datanodeUuid=060e9f99-6979-4f24-a4ff-dab4db6244fb, infoPort=41629, infoSecurePort=0, ipcPort=41349, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092) 2024-11-08T08:51:10,747 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x18c19ac146f3a1c4 with lease ID 0x17f0e4e04a887630: from storage DS-f3b75a7c-ffc5-4c18-b304-2b195de27997 node DatanodeRegistration(127.0.0.1:39689, datanodeUuid=060e9f99-6979-4f24-a4ff-dab4db6244fb, infoPort=41629, infoSecurePort=0, ipcPort=41349, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:10,747 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xcf68aa8e09be2c6a with lease ID 0x17f0e4e04a887631: Processing first storage report for DS-5fabcd0b-0400-4c1d-ad00-06572f786d3e from datanode DatanodeRegistration(127.0.0.1:37237, datanodeUuid=4192a270-ddb6-4969-9470-f4d8ec183401, infoPort=43503, infoSecurePort=0, ipcPort=43027, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092) 2024-11-08T08:51:10,747 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xcf68aa8e09be2c6a with lease ID 0x17f0e4e04a887631: from storage DS-5fabcd0b-0400-4c1d-ad00-06572f786d3e node DatanodeRegistration(127.0.0.1:37237, datanodeUuid=4192a270-ddb6-4969-9470-f4d8ec183401, infoPort=43503, infoSecurePort=0, ipcPort=43027, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:10,782 WARN [Thread-142 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data5/current/BP-1794552092-172.17.0.3-1731055867092/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:10,782 WARN [Thread-143 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data6/current/BP-1794552092-172.17.0.3-1731055867092/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:10,799 WARN [Thread-103 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-08T08:51:10,803 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x72c1228931c18a7e with lease ID 0x17f0e4e04a887632: Processing first storage report for DS-9ae82b2b-674f-4ef9-9797-e484eaa017b4 from datanode DatanodeRegistration(127.0.0.1:41019, datanodeUuid=9da036be-4ee2-4e34-b32c-cac8bf124248, infoPort=43887, infoSecurePort=0, ipcPort=46677, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092) 2024-11-08T08:51:10,803 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x72c1228931c18a7e with lease ID 0x17f0e4e04a887632: from storage DS-9ae82b2b-674f-4ef9-9797-e484eaa017b4 node DatanodeRegistration(127.0.0.1:41019, datanodeUuid=9da036be-4ee2-4e34-b32c-cac8bf124248, infoPort=43887, infoSecurePort=0, ipcPort=46677, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:10,804 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x72c1228931c18a7e with lease ID 0x17f0e4e04a887632: Processing first storage report for DS-45880e0e-a9f5-425d-b8fc-d58cb00e942b from datanode DatanodeRegistration(127.0.0.1:41019, datanodeUuid=9da036be-4ee2-4e34-b32c-cac8bf124248, infoPort=43887, infoSecurePort=0, ipcPort=46677, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092) 2024-11-08T08:51:10,804 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x72c1228931c18a7e with lease ID 0x17f0e4e04a887632: from storage DS-45880e0e-a9f5-425d-b8fc-d58cb00e942b node DatanodeRegistration(127.0.0.1:41019, datanodeUuid=9da036be-4ee2-4e34-b32c-cac8bf124248, infoPort=43887, infoSecurePort=0, ipcPort=46677, storageInfo=lv=-57;cid=testClusterID;nsid=1905036219;c=1731055867092), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:10,823 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989 2024-11-08T08:51:10,889 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-11-08T08:51:10,998 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=158, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=98, ProcessCount=11, AvailableMemoryMB=9504 2024-11-08T08:51:11,001 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-08T08:51:11,011 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-11-08T08:51:11,086 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/zookeeper_0, clientPort=50272, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-08T08:51:11,095 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=50272 2024-11-08T08:51:11,117 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:11,121 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:11,217 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:11,217 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:11,255 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:33562 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:39689:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:33562 dst: /127.0.0.1:39689 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:11,276 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775792_1002 (size=7) 2024-11-08T08:51:11,675 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:11,688 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36 with version=8 2024-11-08T08:51:11,688 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/hbase-staging 2024-11-08T08:51:11,773 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-11-08T08:51:12,019 INFO [Time-limited test {}] client.ConnectionUtils(128): master/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:12,028 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:12,029 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:12,033 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:12,033 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:12,034 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:12,163 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-08T08:51:12,218 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-11-08T08:51:12,227 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-11-08T08:51:12,230 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:12,254 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 92730 (auto-detected) 2024-11-08T08:51:12,255 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:03 (auto-detected) 2024-11-08T08:51:12,270 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:42533 2024-11-08T08:51:12,289 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:42533 connecting to ZooKeeper ensemble=127.0.0.1:50272 2024-11-08T08:51:12,414 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:425330x0, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:12,417 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:42533-0x10119a460090000 connected 2024-11-08T08:51:12,959 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:12,966 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:12,978 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:12,983 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36, hbase.cluster.distributed=false 2024-11-08T08:51:13,008 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:13,012 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42533 2024-11-08T08:51:13,013 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42533 2024-11-08T08:51:13,013 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42533 2024-11-08T08:51:13,014 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42533 2024-11-08T08:51:13,014 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42533 2024-11-08T08:51:13,109 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:13,110 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,110 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,111 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:13,111 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,111 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:13,113 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-08T08:51:13,116 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:13,117 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:35983 2024-11-08T08:51:13,118 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:35983 connecting to ZooKeeper ensemble=127.0.0.1:50272 2024-11-08T08:51:13,119 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,123 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,140 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:359830x0, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:13,141 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:359830x0, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:13,142 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35983-0x10119a460090001 connected 2024-11-08T08:51:13,145 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-08T08:51:13,152 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-08T08:51:13,154 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-08T08:51:13,160 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:13,161 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35983 2024-11-08T08:51:13,161 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35983 2024-11-08T08:51:13,162 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35983 2024-11-08T08:51:13,162 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35983 2024-11-08T08:51:13,162 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35983 2024-11-08T08:51:13,179 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:13,179 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,179 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,180 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:13,180 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,180 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:13,180 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-08T08:51:13,181 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:13,182 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:44823 2024-11-08T08:51:13,184 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:44823 connecting to ZooKeeper ensemble=127.0.0.1:50272 2024-11-08T08:51:13,186 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,191 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,208 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:448230x0, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:13,209 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:44823-0x10119a460090002 connected 2024-11-08T08:51:13,209 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:13,210 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-08T08:51:13,211 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-08T08:51:13,213 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-08T08:51:13,215 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:13,216 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=44823 2024-11-08T08:51:13,216 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=44823 2024-11-08T08:51:13,217 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=44823 2024-11-08T08:51:13,218 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=44823 2024-11-08T08:51:13,218 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=44823 2024-11-08T08:51:13,234 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:13,234 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,234 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,234 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:13,234 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:13,235 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:13,235 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-08T08:51:13,235 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:13,236 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:39923 2024-11-08T08:51:13,238 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39923 connecting to ZooKeeper ensemble=127.0.0.1:50272 2024-11-08T08:51:13,239 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,242 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,256 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:399230x0, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:13,257 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39923-0x10119a460090003 connected 2024-11-08T08:51:13,257 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:13,258 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-08T08:51:13,259 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-08T08:51:13,260 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-08T08:51:13,263 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:13,264 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39923 2024-11-08T08:51:13,264 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39923 2024-11-08T08:51:13,265 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39923 2024-11-08T08:51:13,266 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39923 2024-11-08T08:51:13,266 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39923 2024-11-08T08:51:13,283 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;6ac96558fc7f:42533 2024-11-08T08:51:13,284 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:13,331 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,331 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,331 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,331 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,334 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:13,366 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:13,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:13,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:13,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,366 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,367 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-08T08:51:13,369 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/6ac96558fc7f,42533,1731055871860 from backup master directory 2024-11-08T08:51:13,383 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:13,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:13,383 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:13,383 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:13,385 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-11-08T08:51:13,387 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-11-08T08:51:13,446 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/hbase.id] with ID: 2d82fd20-f265-48cf-9cbc-69b67c190911 2024-11-08T08:51:13,446 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/.tmp/hbase.id 2024-11-08T08:51:13,454 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,454 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,457 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41596 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41596 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:13,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775776_1004 (size=42) 2024-11-08T08:51:13,465 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:13,466 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/.tmp/hbase.id]:[hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/hbase.id] 2024-11-08T08:51:13,521 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:13,528 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-08T08:51:13,547 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 18ms. 2024-11-08T08:51:13,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,561 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:13,575 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,575 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,578 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41614 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41614 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:13,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775760_1006 (size=196) 2024-11-08T08:51:13,584 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:13,600 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-08T08:51:13,601 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-08T08:51:13,606 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-08T08:51:13,632 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,632 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,635 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41634 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41634 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:13,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775744_1008 (size=1189) 2024-11-08T08:51:13,641 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:13,658 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store 2024-11-08T08:51:13,673 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,673 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:13,677 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:39958 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:37237:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39958 dst: /127.0.0.1:37237 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:13,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775728_1010 (size=34) 2024-11-08T08:51:13,682 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:13,686 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-11-08T08:51:13,690 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:13,691 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-08T08:51:13,691 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:13,691 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:13,693 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-08T08:51:13,693 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:13,693 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:13,694 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731055873691Disabling compacts and flushes for region at 1731055873691Disabling writes for close at 1731055873693 (+2 ms)Writing region close event to WAL at 1731055873693Closed at 1731055873693 2024-11-08T08:51:13,696 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/.initializing 2024-11-08T08:51:13,697 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/WALs/6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:13,705 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-08T08:51:13,718 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C42533%2C1731055871860, suffix=, logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/WALs/6ac96558fc7f,42533,1731055871860, archiveDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/oldWALs, maxLogs=10 2024-11-08T08:51:13,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775789_1002 (size=7) 2024-11-08T08:51:13,743 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775788_1002 (size=7) 2024-11-08T08:51:13,747 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/WALs/6ac96558fc7f,42533,1731055871860/6ac96558fc7f%2C42533%2C1731055871860.1731055873723, exclude list is [], retry=0 2024-11-08T08:51:13,750 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:13,750 WARN [IPC Server handler 3 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:13,751 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:13,765 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:414) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:473) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:468) ~[hbase-asyncfs-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:13,766 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37237,DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd,DISK] 2024-11-08T08:51:13,766 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41019,DS-9ae82b2b-674f-4ef9-9797-e484eaa017b4,DISK] 2024-11-08T08:51:13,769 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-11-08T08:51:13,806 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/WALs/6ac96558fc7f,42533,1731055871860/6ac96558fc7f%2C42533%2C1731055871860.1731055873723 2024-11-08T08:51:13,807 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43503:43503),(127.0.0.1/127.0.0.1:43887:43887)] 2024-11-08T08:51:13,808 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-08T08:51:13,808 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:13,812 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,813 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,849 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,872 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-08T08:51:13,875 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:13,877 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:13,877 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,881 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-08T08:51:13,881 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:13,882 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:13,882 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,885 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-08T08:51:13,885 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:13,886 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:13,886 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,889 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-08T08:51:13,889 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:13,890 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:13,891 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,894 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,895 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,900 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,900 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,903 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-08T08:51:13,907 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:13,913 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-08T08:51:13,914 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=60954254, jitterRate=-0.09171083569526672}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-08T08:51:13,922 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1731055873825Initializing all the Stores at 1731055873828 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055873828Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055873829 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055873829Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055873829Cleaning up temporary data from old regions at 1731055873900 (+71 ms)Region opened successfully at 1731055873922 (+22 ms) 2024-11-08T08:51:13,923 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-08T08:51:13,956 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@555799b1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:13,983 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-08T08:51:13,993 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-08T08:51:13,993 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-08T08:51:13,995 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-08T08:51:13,998 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 2 msec 2024-11-08T08:51:14,004 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 5 msec 2024-11-08T08:51:14,004 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-08T08:51:14,028 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-08T08:51:14,036 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-08T08:51:14,082 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-08T08:51:14,085 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-08T08:51:14,087 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-08T08:51:14,098 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-08T08:51:14,101 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-08T08:51:14,105 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-08T08:51:14,113 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-08T08:51:14,115 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-08T08:51:14,123 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-08T08:51:14,144 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-08T08:51:14,155 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-08T08:51:14,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:14,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:14,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:14,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,166 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,166 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:14,166 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,170 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=6ac96558fc7f,42533,1731055871860, sessionid=0x10119a460090000, setting cluster-up flag (Was=false) 2024-11-08T08:51:14,203 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,203 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,235 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-08T08:51:14,237 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:14,260 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,260 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,260 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,260 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:14,292 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-08T08:51:14,294 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:14,302 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-08T08:51:14,370 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(746): ClusterId : 2d82fd20-f265-48cf-9cbc-69b67c190911 2024-11-08T08:51:14,370 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(746): ClusterId : 2d82fd20-f265-48cf-9cbc-69b67c190911 2024-11-08T08:51:14,371 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(746): ClusterId : 2d82fd20-f265-48cf-9cbc-69b67c190911 2024-11-08T08:51:14,372 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-08T08:51:14,372 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-08T08:51:14,372 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-08T08:51:14,374 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:14,383 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-08T08:51:14,389 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-08T08:51:14,395 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-08T08:51:14,395 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-08T08:51:14,395 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-08T08:51:14,395 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-08T08:51:14,395 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-08T08:51:14,395 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-08T08:51:14,395 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 6ac96558fc7f,42533,1731055871860 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-08T08:51:14,410 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-08T08:51:14,410 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-08T08:51:14,410 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-08T08:51:14,410 DEBUG [RS:1;6ac96558fc7f:44823 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1ba93c19, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:14,410 DEBUG [RS:2;6ac96558fc7f:39923 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@46168e86, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:14,410 DEBUG [RS:0;6ac96558fc7f:35983 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5439c5d1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:14,411 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:14,411 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:14,412 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:14,412 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:14,412 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/6ac96558fc7f:0, corePoolSize=10, maxPoolSize=10 2024-11-08T08:51:14,412 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,412 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:14,413 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,423 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1731055904423 2024-11-08T08:51:14,426 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-08T08:51:14,427 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-08T08:51:14,431 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-08T08:51:14,432 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-08T08:51:14,432 DEBUG [RS:2;6ac96558fc7f:39923 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;6ac96558fc7f:39923 2024-11-08T08:51:14,433 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-08T08:51:14,433 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-08T08:51:14,434 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;6ac96558fc7f:35983 2024-11-08T08:51:14,434 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;6ac96558fc7f:44823 2024-11-08T08:51:14,434 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:14,434 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-08T08:51:14,437 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-08T08:51:14,437 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-08T08:51:14,437 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-08T08:51:14,437 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-08T08:51:14,437 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-08T08:51:14,437 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-08T08:51:14,437 DEBUG [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-08T08:51:14,437 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-08T08:51:14,437 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-08T08:51:14,435 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,440 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(2659): reportForDuty to master=6ac96558fc7f,42533,1731055871860 with port=35983, startcode=1731055873077 2024-11-08T08:51:14,440 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(2659): reportForDuty to master=6ac96558fc7f,42533,1731055871860 with port=39923, startcode=1731055873233 2024-11-08T08:51:14,440 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(2659): reportForDuty to master=6ac96558fc7f,42533,1731055871860 with port=44823, startcode=1731055873178 2024-11-08T08:51:14,442 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:14,442 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-08T08:51:14,453 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-08T08:51:14,453 DEBUG [RS:0;6ac96558fc7f:35983 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-08T08:51:14,453 DEBUG [RS:2;6ac96558fc7f:39923 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-08T08:51:14,453 DEBUG [RS:1;6ac96558fc7f:44823 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-08T08:51:14,454 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-08T08:51:14,454 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-08T08:51:14,456 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-08T08:51:14,457 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-08T08:51:14,460 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.large.0-1731055874458,5,FailOnTimeoutGroup] 2024-11-08T08:51:14,461 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.small.0-1731055874460,5,FailOnTimeoutGroup] 2024-11-08T08:51:14,461 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,461 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-08T08:51:14,462 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=3) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-08T08:51:14,462 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,462 WARN [IPC Server handler 3 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:14,462 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:14,463 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,463 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,467 DEBUG [PEWorker-1 {}] util.FSTableDescriptors(635): Failed write hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321; retrying up to 10 times org.apache.hadoop.ipc.RemoteException: File /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$addBlock$11(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1143) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.allocateNewBlock(DFSStripedOutputStream.java:508) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.writeChunk(DFSStripedOutputStream.java:561) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(FSOutputSummer.java:220) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:165) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:146) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.closeImpl(DFSStripedOutputStream.java:1234) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:861) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:77) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:106) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.util.FSTableDescriptors.writeTableDescriptor(FSTableDescriptors.java:631) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.util.FSTableDescriptors.tryUpdateAndGetMetaTableDescriptor(FSTableDescriptors.java:159) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.writeFsLayout(InitMetaProcedure.java:87) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.executeFromState(InitMetaProcedure.java:103) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.executeFromState(InitMetaProcedure.java:55) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:188) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:941) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1825) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.executeProcedure(ProcedureExecutor.java:1503) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.runProcedure(ProcedureExecutor.java:2154) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2181) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] 2024-11-08T08:51:14,483 WARN [IPC Server handler 0 on default port 36363 {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=3) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-08T08:51:14,483 WARN [IPC Server handler 0 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,483 WARN [IPC Server handler 0 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:14,484 WARN [IPC Server handler 0 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:14,487 DEBUG [PEWorker-1 {}] util.FSTableDescriptors(635): Failed write hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000002.1321; retrying up to 10 times org.apache.hadoop.ipc.RemoteException: File /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000002.1321 could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$addBlock$11(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1143) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.allocateNewBlock(DFSStripedOutputStream.java:508) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.writeChunk(DFSStripedOutputStream.java:561) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(FSOutputSummer.java:220) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:165) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:146) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.closeImpl(DFSStripedOutputStream.java:1234) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:861) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:77) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:106) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.util.FSTableDescriptors.writeTableDescriptor(FSTableDescriptors.java:631) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.util.FSTableDescriptors.tryUpdateAndGetMetaTableDescriptor(FSTableDescriptors.java:159) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.writeFsLayout(InitMetaProcedure.java:87) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.executeFromState(InitMetaProcedure.java:103) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.executeFromState(InitMetaProcedure.java:55) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:188) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:941) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1825) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.executeProcedure(ProcedureExecutor.java:1503) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.runProcedure(ProcedureExecutor.java:2154) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2181) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] 2024-11-08T08:51:14,502 WARN [IPC Server handler 1 on default port 36363 {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=3) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-08T08:51:14,502 WARN [IPC Server handler 1 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,502 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:58541, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-11-08T08:51:14,502 WARN [IPC Server handler 1 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:14,502 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:34317, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-11-08T08:51:14,502 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37081, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-11-08T08:51:14,502 WARN [IPC Server handler 1 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:14,503 DEBUG [PEWorker-1 {}] util.FSTableDescriptors(635): Failed write hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000003.1321; retrying up to 10 times org.apache.hadoop.ipc.RemoteException: File /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000003.1321 could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$addBlock$11(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1143) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.allocateNewBlock(DFSStripedOutputStream.java:508) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.writeChunk(DFSStripedOutputStream.java:561) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(FSOutputSummer.java:220) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:165) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:146) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.closeImpl(DFSStripedOutputStream.java:1234) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:861) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:77) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:106) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.util.FSTableDescriptors.writeTableDescriptor(FSTableDescriptors.java:631) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.util.FSTableDescriptors.tryUpdateAndGetMetaTableDescriptor(FSTableDescriptors.java:159) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.writeFsLayout(InitMetaProcedure.java:87) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.executeFromState(InitMetaProcedure.java:103) ~[classes/:?] at org.apache.hadoop.hbase.master.procedure.InitMetaProcedure.executeFromState(InitMetaProcedure.java:55) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:188) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:941) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1825) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.executeProcedure(ProcedureExecutor.java:1503) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.runProcedure(ProcedureExecutor.java:2154) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2181) ~[hbase-procedure-3.0.0-beta-2-SNAPSHOT.jar:3.0.0-beta-2-SNAPSHOT] 2024-11-08T08:51:14,509 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42533 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:14,512 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42533 {}] master.ServerManager(517): Registering regionserver=6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:14,515 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:14,515 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:14,524 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42533 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:14,524 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42533 {}] master.ServerManager(517): Registering regionserver=6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:14,525 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41648 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41648 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:14,529 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42533 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:14,529 DEBUG [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36 2024-11-08T08:51:14,529 DEBUG [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:36363 2024-11-08T08:51:14,529 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36 2024-11-08T08:51:14,529 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42533 {}] master.ServerManager(517): Registering regionserver=6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:14,529 DEBUG [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-08T08:51:14,529 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:36363 2024-11-08T08:51:14,529 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-08T08:51:14,534 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36 2024-11-08T08:51:14,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775712_1013 (size=1321) 2024-11-08T08:51:14,534 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:36363 2024-11-08T08:51:14,534 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-08T08:51:14,536 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:14,538 DEBUG [PEWorker-1 {}] util.FSTableDescriptors(591): Deleted hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-08T08:51:14,539 DEBUG [PEWorker-1 {}] util.FSTableDescriptors(591): Deleted hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000002.1321 2024-11-08T08:51:14,540 DEBUG [PEWorker-1 {}] util.FSTableDescriptors(591): Deleted hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000003.1321 2024-11-08T08:51:14,541 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/.tabledesc/.tableinfo.0000000004.1321 2024-11-08T08:51:14,541 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36 2024-11-08T08:51:14,549 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:14,549 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:14,553 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41656 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41656 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:14,557 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775696_1015 (size=32) 2024-11-08T08:51:14,558 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:14,559 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:14,561 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-08T08:51:14,564 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-08T08:51:14,564 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:14,565 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:14,566 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-08T08:51:14,569 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-08T08:51:14,569 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:14,570 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:14,571 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-08T08:51:14,572 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-08T08:51:14,574 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-08T08:51:14,574 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:14,575 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:14,575 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-08T08:51:14,578 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-08T08:51:14,578 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:14,579 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:14,579 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-08T08:51:14,581 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740 2024-11-08T08:51:14,582 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740 2024-11-08T08:51:14,585 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-08T08:51:14,585 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-08T08:51:14,586 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-08T08:51:14,588 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-08T08:51:14,593 DEBUG [RS:2;6ac96558fc7f:39923 {}] zookeeper.ZKUtil(111): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:14,593 DEBUG [RS:1;6ac96558fc7f:44823 {}] zookeeper.ZKUtil(111): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:14,593 WARN [RS:1;6ac96558fc7f:44823 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:14,593 DEBUG [RS:0;6ac96558fc7f:35983 {}] zookeeper.ZKUtil(111): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:14,593 WARN [RS:2;6ac96558fc7f:39923 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:14,594 WARN [RS:0;6ac96558fc7f:35983 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:14,594 INFO [RS:2;6ac96558fc7f:39923 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-08T08:51:14,594 INFO [RS:1;6ac96558fc7f:44823 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-08T08:51:14,594 INFO [RS:0;6ac96558fc7f:35983 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-08T08:51:14,594 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:14,594 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:14,594 DEBUG [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:14,595 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [6ac96558fc7f,39923,1731055873233] 2024-11-08T08:51:14,596 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [6ac96558fc7f,44823,1731055873178] 2024-11-08T08:51:14,596 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [6ac96558fc7f,35983,1731055873077] 2024-11-08T08:51:14,610 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-08T08:51:14,611 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67701839, jitterRate=0.008836016058921814}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-08T08:51:14,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1731055874559Initializing all the Stores at 1731055874561 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055874561Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055874561Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055874561Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055874561Cleaning up temporary data from old regions at 1731055874585 (+24 ms)Region opened successfully at 1731055874616 (+31 ms) 2024-11-08T08:51:14,616 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-08T08:51:14,616 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-08T08:51:14,617 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-08T08:51:14,617 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-08T08:51:14,617 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-08T08:51:14,621 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-08T08:51:14,621 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731055874616Disabling compacts and flushes for region at 1731055874616Disabling writes for close at 1731055874617 (+1 ms)Writing region close event to WAL at 1731055874620 (+3 ms)Closed at 1731055874621 (+1 ms) 2024-11-08T08:51:14,624 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:14,625 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-08T08:51:14,626 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-08T08:51:14,626 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-08T08:51:14,626 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-08T08:51:14,632 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-08T08:51:14,644 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-08T08:51:14,649 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-08T08:51:14,649 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-08T08:51:14,652 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-08T08:51:14,656 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-08T08:51:14,661 INFO [RS:0;6ac96558fc7f:35983 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-08T08:51:14,661 INFO [RS:2;6ac96558fc7f:39923 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-08T08:51:14,661 INFO [RS:1;6ac96558fc7f:44823 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-08T08:51:14,662 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,662 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,662 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,665 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-08T08:51:14,665 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-08T08:51:14,665 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-08T08:51:14,671 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-08T08:51:14,671 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-08T08:51:14,671 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-08T08:51:14,673 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,673 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,673 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,673 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,673 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,674 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,675 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,675 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,675 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,675 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:14,675 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:14,675 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:14,675 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:14,675 DEBUG [RS:0;6ac96558fc7f:35983 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:14,675 DEBUG [RS:1;6ac96558fc7f:44823 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:14,675 DEBUG [RS:2;6ac96558fc7f:39923 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:14,680 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,35983,1731055873077-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,681 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,39923,1731055873233-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:14,684 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,684 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,684 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,685 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,685 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,685 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,44823,1731055873178-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:14,708 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-08T08:51:14,708 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-08T08:51:14,709 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-08T08:51:14,711 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,39923,1731055873233-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,711 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,44823,1731055873178-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,711 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,35983,1731055873077-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,711 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,711 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,711 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,711 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.Replication(171): 6ac96558fc7f,39923,1731055873233 started 2024-11-08T08:51:14,711 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.Replication(171): 6ac96558fc7f,44823,1731055873178 started 2024-11-08T08:51:14,711 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.Replication(171): 6ac96558fc7f,35983,1731055873077 started 2024-11-08T08:51:14,737 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,737 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,737 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:14,737 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(1482): Serving as 6ac96558fc7f,39923,1731055873233, RpcServer on 6ac96558fc7f/172.17.0.3:39923, sessionid=0x10119a460090003 2024-11-08T08:51:14,737 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1482): Serving as 6ac96558fc7f,44823,1731055873178, RpcServer on 6ac96558fc7f/172.17.0.3:44823, sessionid=0x10119a460090002 2024-11-08T08:51:14,737 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1482): Serving as 6ac96558fc7f,35983,1731055873077, RpcServer on 6ac96558fc7f/172.17.0.3:35983, sessionid=0x10119a460090001 2024-11-08T08:51:14,738 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-08T08:51:14,738 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-08T08:51:14,738 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-08T08:51:14,738 DEBUG [RS:0;6ac96558fc7f:35983 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:14,738 DEBUG [RS:1;6ac96558fc7f:44823 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:14,738 DEBUG [RS:2;6ac96558fc7f:39923 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:14,738 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,35983,1731055873077' 2024-11-08T08:51:14,738 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,39923,1731055873233' 2024-11-08T08:51:14,739 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-08T08:51:14,739 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-08T08:51:14,740 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-08T08:51:14,740 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-08T08:51:14,740 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=2) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-08T08:51:14,740 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,740 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-08T08:51:14,740 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-08T08:51:14,740 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=1) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-08T08:51:14,741 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-08T08:51:14,741 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-08T08:51:14,741 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,741 DEBUG [RS:0;6ac96558fc7f:35983 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:14,741 DEBUG [RS:2;6ac96558fc7f:39923 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:14,741 WARN [RedundancyMonitor {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK, ARCHIVE], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:14,741 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,35983,1731055873077' 2024-11-08T08:51:14,741 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,39923,1731055873233' 2024-11-08T08:51:14,741 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-08T08:51:14,741 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-08T08:51:14,741 WARN [RedundancyMonitor {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=false) All required storage types are unavailable: unavailableStorages=[DISK, ARCHIVE], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:14,742 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-08T08:51:14,742 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-08T08:51:14,742 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,44823,1731055873178' 2024-11-08T08:51:14,742 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-08T08:51:14,742 DEBUG [RS:0;6ac96558fc7f:35983 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-08T08:51:14,742 DEBUG [RS:2;6ac96558fc7f:39923 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-08T08:51:14,742 INFO [RS:0;6ac96558fc7f:35983 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-08T08:51:14,742 INFO [RS:2;6ac96558fc7f:39923 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-08T08:51:14,742 INFO [RS:2;6ac96558fc7f:39923 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-08T08:51:14,742 INFO [RS:0;6ac96558fc7f:35983 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-08T08:51:14,743 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-08T08:51:14,743 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-08T08:51:14,743 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-08T08:51:14,743 DEBUG [RS:1;6ac96558fc7f:44823 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:14,743 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,44823,1731055873178' 2024-11-08T08:51:14,743 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-08T08:51:14,744 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-08T08:51:14,744 DEBUG [RS:1;6ac96558fc7f:44823 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-08T08:51:14,744 INFO [RS:1;6ac96558fc7f:44823 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-08T08:51:14,744 INFO [RS:1;6ac96558fc7f:44823 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-08T08:51:14,801 WARN [6ac96558fc7f:42533 {}] assignment.AssignmentManager(2443): No servers available; cannot place 1 unassigned regions. 2024-11-08T08:51:14,847 INFO [RS:2;6ac96558fc7f:39923 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-08T08:51:14,847 INFO [RS:0;6ac96558fc7f:35983 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-08T08:51:14,847 INFO [RS:1;6ac96558fc7f:44823 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-11-08T08:51:14,850 INFO [RS:0;6ac96558fc7f:35983 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C35983%2C1731055873077, suffix=, logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,35983,1731055873077, archiveDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs, maxLogs=32 2024-11-08T08:51:14,850 INFO [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C44823%2C1731055873178, suffix=, logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178, archiveDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs, maxLogs=32 2024-11-08T08:51:14,850 INFO [RS:2;6ac96558fc7f:39923 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C39923%2C1731055873233, suffix=, logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,39923,1731055873233, archiveDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs, maxLogs=32 2024-11-08T08:51:14,867 DEBUG [RS:1;6ac96558fc7f:44823 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178/6ac96558fc7f%2C44823%2C1731055873178.1731055874854, exclude list is [], retry=0 2024-11-08T08:51:14,868 DEBUG [RS:2;6ac96558fc7f:39923 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,39923,1731055873233/6ac96558fc7f%2C39923%2C1731055873233.1731055874854, exclude list is [], retry=0 2024-11-08T08:51:14,868 DEBUG [RS:0;6ac96558fc7f:35983 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,35983,1731055873077/6ac96558fc7f%2C35983%2C1731055873077.1731055874854, exclude list is [], retry=0 2024-11-08T08:51:14,871 WARN [IPC Server handler 1 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,871 WARN [IPC Server handler 1 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:14,871 WARN [IPC Server handler 1 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:14,872 WARN [IPC Server handler 4 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:14,872 WARN [IPC Server handler 4 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:14,872 WARN [IPC Server handler 4 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:14,874 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41019,DS-9ae82b2b-674f-4ef9-9797-e484eaa017b4,DISK] 2024-11-08T08:51:14,875 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39689,DS-42b2faa6-d7f4-4bc0-8e8f-61fde4bb9e11,DISK] 2024-11-08T08:51:14,875 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37237,DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd,DISK] 2024-11-08T08:51:14,875 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37237,DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd,DISK] 2024-11-08T08:51:14,876 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39689,DS-42b2faa6-d7f4-4bc0-8e8f-61fde4bb9e11,DISK] 2024-11-08T08:51:14,876 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37237,DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd,DISK] 2024-11-08T08:51:14,877 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39689,DS-42b2faa6-d7f4-4bc0-8e8f-61fde4bb9e11,DISK] 2024-11-08T08:51:14,901 INFO [RS:0;6ac96558fc7f:35983 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,35983,1731055873077/6ac96558fc7f%2C35983%2C1731055873077.1731055874854 2024-11-08T08:51:14,901 INFO [RS:2;6ac96558fc7f:39923 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,39923,1731055873233/6ac96558fc7f%2C39923%2C1731055873233.1731055874854 2024-11-08T08:51:14,901 DEBUG [RS:2;6ac96558fc7f:39923 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43503:43503),(127.0.0.1/127.0.0.1:41629:41629)] 2024-11-08T08:51:14,901 DEBUG [RS:0;6ac96558fc7f:35983 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43503:43503),(127.0.0.1/127.0.0.1:41629:41629)] 2024-11-08T08:51:14,901 INFO [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178/6ac96558fc7f%2C44823%2C1731055873178.1731055874854 2024-11-08T08:51:14,902 DEBUG [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43887:43887),(127.0.0.1/127.0.0.1:43503:43503),(127.0.0.1/127.0.0.1:41629:41629)] 2024-11-08T08:51:15,054 DEBUG [6ac96558fc7f:42533 {}] assignment.AssignmentManager(2464): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-08T08:51:15,060 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(204): Hosts are {6ac96558fc7f=0} racks are {/default-rack=0} 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-08T08:51:15,067 INFO [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-08T08:51:15,067 INFO [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-08T08:51:15,067 INFO [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-08T08:51:15,067 DEBUG [6ac96558fc7f:42533 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-08T08:51:15,073 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:15,080 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 6ac96558fc7f,44823,1731055873178, state=OPENING 2024-11-08T08:51:15,134 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-08T08:51:15,145 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:15,145 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:15,145 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:15,145 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:15,146 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,146 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,146 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,146 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,148 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-08T08:51:15,151 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=6ac96558fc7f,44823,1731055873178}] 2024-11-08T08:51:15,325 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-08T08:51:15,327 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:34305, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-08T08:51:15,341 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-08T08:51:15,342 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-11-08T08:51:15,342 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-11-08T08:51:15,347 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C44823%2C1731055873178.meta, suffix=.meta, logDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178, archiveDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs, maxLogs=32 2024-11-08T08:51:15,362 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(525): When create output stream for /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178/6ac96558fc7f%2C44823%2C1731055873178.meta.1731055875349.meta, exclude list is [], retry=0 2024-11-08T08:51:15,365 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:15,365 WARN [IPC Server handler 3 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:15,365 WARN [IPC Server handler 3 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:15,368 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39689,DS-42b2faa6-d7f4-4bc0-8e8f-61fde4bb9e11,DISK] 2024-11-08T08:51:15,368 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:37237,DS-b58d1eed-b31e-4bde-b0c3-bb1c38f771fd,DISK] 2024-11-08T08:51:15,371 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,44823,1731055873178/6ac96558fc7f%2C44823%2C1731055873178.meta.1731055875349.meta 2024-11-08T08:51:15,371 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:43503:43503),(127.0.0.1/127.0.0.1:41629:41629)] 2024-11-08T08:51:15,371 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-08T08:51:15,373 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-08T08:51:15,376 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-08T08:51:15,380 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-08T08:51:15,384 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-08T08:51:15,385 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:15,385 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-08T08:51:15,385 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-08T08:51:15,388 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-08T08:51:15,390 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-08T08:51:15,390 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:15,391 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:15,391 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-08T08:51:15,393 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-08T08:51:15,393 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:15,394 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:15,394 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-08T08:51:15,396 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-08T08:51:15,396 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:15,396 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:15,397 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-08T08:51:15,398 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-08T08:51:15,398 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:15,399 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:15,399 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-08T08:51:15,400 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740 2024-11-08T08:51:15,403 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740 2024-11-08T08:51:15,406 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-08T08:51:15,406 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-08T08:51:15,407 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-08T08:51:15,410 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-08T08:51:15,412 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61372634, jitterRate=-0.08547648787498474}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-08T08:51:15,413 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-08T08:51:15,414 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1731055875385Writing region info on filesystem at 1731055875385Initializing all the Stores at 1731055875388 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055875388Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055875388Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055875388Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055875388Cleaning up temporary data from old regions at 1731055875406 (+18 ms)Running coprocessor post-open hooks at 1731055875413 (+7 ms)Region opened successfully at 1731055875414 (+1 ms) 2024-11-08T08:51:15,421 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1731055875317 2024-11-08T08:51:15,433 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-08T08:51:15,433 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-08T08:51:15,435 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:15,437 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 6ac96558fc7f,44823,1731055873178, state=OPEN 2024-11-08T08:51:15,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:15,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:15,450 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:15,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:15,450 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,450 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,450 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,450 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:15,451 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:15,458 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-08T08:51:15,458 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=6ac96558fc7f,44823,1731055873178 in 301 msec 2024-11-08T08:51:15,465 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-08T08:51:15,465 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 828 msec 2024-11-08T08:51:15,466 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:15,467 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-08T08:51:15,485 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-08T08:51:15,486 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=6ac96558fc7f,44823,1731055873178, seqNum=-1] 2024-11-08T08:51:15,528 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-08T08:51:15,531 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:34563, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-08T08:51:15,560 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.2240 sec 2024-11-08T08:51:15,560 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1731055875560, completionTime=-1 2024-11-08T08:51:15,569 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-08T08:51:15,569 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] assignment.AssignmentManager(1756): Joining cluster... 2024-11-08T08:51:15,602 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] assignment.AssignmentManager(1768): Number of RegionServers=3 2024-11-08T08:51:15,602 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1731055935602 2024-11-08T08:51:15,602 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1731055995602 2024-11-08T08:51:15,602 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] assignment.AssignmentManager(1775): Joined the cluster in 33 msec 2024-11-08T08:51:15,604 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-11-08T08:51:15,612 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42533,1731055871860-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,612 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42533,1731055871860-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,612 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42533,1731055871860-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,614 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-6ac96558fc7f:42533, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,615 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,615 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,622 DEBUG [master/6ac96558fc7f:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-08T08:51:15,649 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.265sec 2024-11-08T08:51:15,651 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-08T08:51:15,652 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-08T08:51:15,654 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-08T08:51:15,654 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-08T08:51:15,654 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-08T08:51:15,655 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42533,1731055871860-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:15,656 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42533,1731055871860-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-08T08:51:15,661 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-08T08:51:15,662 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-08T08:51:15,662 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42533,1731055871860-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:15,681 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@c0bd1c0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-08T08:51:15,684 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-11-08T08:51:15,684 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-11-08T08:51:15,688 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 6ac96558fc7f,42533,-1 for getting cluster id 2024-11-08T08:51:15,691 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-08T08:51:15,698 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '2d82fd20-f265-48cf-9cbc-69b67c190911' 2024-11-08T08:51:15,701 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-08T08:51:15,701 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "2d82fd20-f265-48cf-9cbc-69b67c190911" 2024-11-08T08:51:15,702 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2a25b292, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-08T08:51:15,702 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [6ac96558fc7f,42533,-1] 2024-11-08T08:51:15,704 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-08T08:51:15,706 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:15,707 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42440, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-08T08:51:15,710 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1aa9a4ec, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-08T08:51:15,711 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-08T08:51:15,718 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=6ac96558fc7f,44823,1731055873178, seqNum=-1] 2024-11-08T08:51:15,718 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-08T08:51:15,721 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60184, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-08T08:51:15,742 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:15,746 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-08T08:51:15,750 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:15,754 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@477921b 2024-11-08T08:51:15,755 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-08T08:51:15,758 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42446, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-08T08:51:15,766 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-08T08:51:15,776 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-08T08:51:15,779 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-08T08:51:15,781 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-08T08:51:15,781 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:15,784 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-08T08:51:15,786 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:15,793 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:15,793 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:15,800 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41678 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41678 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:15,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775680_1021 (size=392) 2024-11-08T08:51:15,807 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:15,810 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 0aed20f1745462a12c55ec714d3d4a79, NAME => 'TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36 2024-11-08T08:51:15,817 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:15,817 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:15,823 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41692 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41692 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:15,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775664_1023 (size=51) 2024-11-08T08:51:15,831 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:15,832 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:15,832 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 0aed20f1745462a12c55ec714d3d4a79, disabling compactions & flushes 2024-11-08T08:51:15,832 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:15,832 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:15,832 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. after waiting 0 ms 2024-11-08T08:51:15,832 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:15,832 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:15,832 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 0aed20f1745462a12c55ec714d3d4a79: Waiting for close lock at 1731055875832Disabling compacts and flushes for region at 1731055875832Disabling writes for close at 1731055875832Writing region close event to WAL at 1731055875832Closed at 1731055875832 2024-11-08T08:51:15,834 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-08T08:51:15,839 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1731055875835"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1731055875835"}]},"ts":"1731055875835"} 2024-11-08T08:51:15,845 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-08T08:51:15,847 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-08T08:51:15,850 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731055875847"}]},"ts":"1731055875847"} 2024-11-08T08:51:15,856 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-08T08:51:15,856 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {6ac96558fc7f=0} racks are {/default-rack=0} 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-08T08:51:15,858 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-08T08:51:15,858 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-08T08:51:15,858 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-08T08:51:15,858 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-08T08:51:15,860 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0aed20f1745462a12c55ec714d3d4a79, ASSIGN}] 2024-11-08T08:51:15,863 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0aed20f1745462a12c55ec714d3d4a79, ASSIGN 2024-11-08T08:51:15,865 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0aed20f1745462a12c55ec714d3d4a79, ASSIGN; state=OFFLINE, location=6ac96558fc7f,35983,1731055873077; forceNewPlan=false, retain=false 2024-11-08T08:51:15,895 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:16,018 INFO [6ac96558fc7f:42533 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-08T08:51:16,019 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=0aed20f1745462a12c55ec714d3d4a79, regionState=OPENING, regionLocation=6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:16,023 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0aed20f1745462a12c55ec714d3d4a79, ASSIGN because future has completed 2024-11-08T08:51:16,024 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 0aed20f1745462a12c55ec714d3d4a79, server=6ac96558fc7f,35983,1731055873077}] 2024-11-08T08:51:16,106 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:16,179 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-08T08:51:16,181 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37989, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-08T08:51:16,187 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:16,187 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 0aed20f1745462a12c55ec714d3d4a79, NAME => 'TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79.', STARTKEY => '', ENDKEY => ''} 2024-11-08T08:51:16,188 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,188 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:16,188 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,188 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,191 INFO [StoreOpener-0aed20f1745462a12c55ec714d3d4a79-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,193 INFO [StoreOpener-0aed20f1745462a12c55ec714d3d4a79-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 0aed20f1745462a12c55ec714d3d4a79 columnFamilyName cf 2024-11-08T08:51:16,194 DEBUG [StoreOpener-0aed20f1745462a12c55ec714d3d4a79-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:16,195 INFO [StoreOpener-0aed20f1745462a12c55ec714d3d4a79-1 {}] regionserver.HStore(327): Store=0aed20f1745462a12c55ec714d3d4a79/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:16,195 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,197 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,197 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,198 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,198 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,201 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,207 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-08T08:51:16,208 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 0aed20f1745462a12c55ec714d3d4a79; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65980911, jitterRate=-0.016807809472084045}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-08T08:51:16,208 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:16,209 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 0aed20f1745462a12c55ec714d3d4a79: Running coprocessor pre-open hook at 1731055876188Writing region info on filesystem at 1731055876188Initializing all the Stores at 1731055876190 (+2 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055876191 (+1 ms)Cleaning up temporary data from old regions at 1731055876198 (+7 ms)Running coprocessor post-open hooks at 1731055876208 (+10 ms)Region opened successfully at 1731055876209 (+1 ms) 2024-11-08T08:51:16,211 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79., pid=6, masterSystemTime=1731055876179 2024-11-08T08:51:16,215 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:16,215 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:16,216 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=0aed20f1745462a12c55ec714d3d4a79, regionState=OPEN, openSeqNum=2, regionLocation=6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:16,220 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 0aed20f1745462a12c55ec714d3d4a79, server=6ac96558fc7f,35983,1731055873077 because future has completed 2024-11-08T08:51:16,227 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-08T08:51:16,229 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 0aed20f1745462a12c55ec714d3d4a79, server=6ac96558fc7f,35983,1731055873077 in 198 msec 2024-11-08T08:51:16,232 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-08T08:51:16,232 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=0aed20f1745462a12c55ec714d3d4a79, ASSIGN in 368 msec 2024-11-08T08:51:16,233 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-08T08:51:16,233 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731055876233"}]},"ts":"1731055876233"} 2024-11-08T08:51:16,237 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-08T08:51:16,238 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-08T08:51:16,242 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 468 msec 2024-11-08T08:51:16,414 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:16,415 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-08T08:51:16,415 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-08T08:51:16,416 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-08T08:51:16,421 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-08T08:51:16,421 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-08T08:51:16,422 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-08T08:51:16,430 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79., hostname=6ac96558fc7f,35983,1731055873077, seqNum=2] 2024-11-08T08:51:16,431 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-08T08:51:16,433 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:50378, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-08T08:51:16,442 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestHBaseWalOnEC 2024-11-08T08:51:16,446 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-08T08:51:16,448 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-08T08:51:16,449 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:16,451 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-08T08:51:16,452 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-08T08:51:16,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:16,615 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=35983 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-08T08:51:16,615 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:16,619 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 0aed20f1745462a12c55ec714d3d4a79 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-08T08:51:16,669 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/.tmp/cf/0625b8f9c5204aab8baef5cf0cfcea1d is 36, key is row/cf:cq/1731055876434/Put/seqid=0 2024-11-08T08:51:16,674 WARN [IPC Server handler 0 on default port 36363 {}] blockmanagement.BlockPlacementPolicyRackFaultTolerant(145): Only able to place 2 of total expected 3 (maxNodesPerRack=3, numOfReplicas=3) nodes evenly across racks, falling back to evenly place on the remaining racks. This may not guarantee rack-level fault tolerance. Please check if the racks are configured properly. 2024-11-08T08:51:16,675 WARN [IPC Server handler 0 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:16,675 WARN [IPC Server handler 0 on default port 36363 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:16,675 WARN [IPC Server handler 0 on default port 36363 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:16,676 WARN [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore(850): Failed flushing store file for 0aed20f1745462a12c55ec714d3d4a79/cf, retrying num=0 org.apache.hadoop.ipc.RemoteException: File /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/.tmp/cf/0625b8f9c5204aab8baef5cf0cfcea1d could only be written to 2 of the 3 required nodes for RS-3-2-1024k. There are 3 datanode(s) running and 3 node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2480) at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:293) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3075) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:932) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:603) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$addBlock$11(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:500) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy48.addBlock(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1143) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.allocateNewBlock(DFSStripedOutputStream.java:508) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSStripedOutputStream.writeChunk(DFSStripedOutputStream.java:561) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunks(FSOutputSummer.java:220) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:165) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:146) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:137) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:112) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at java.io.DataOutputStream.write(DataOutputStream.java:112) ~[?:?] at java.io.ByteArrayOutputStream.writeTo(ByteArrayOutputStream.java:161) ~[?:?] at org.apache.hadoop.hbase.io.hfile.FixedFileTrailer.serialize(FixedFileTrailer.java:197) ~[classes/:?] at org.apache.hadoop.hbase.io.hfile.HFileWriterImpl.finishClose(HFileWriterImpl.java:855) ~[classes/:?] at org.apache.hadoop.hbase.io.hfile.HFileWriterImpl.close(HFileWriterImpl.java:680) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.StoreFileWriter$SingleStoreFileWriter.close(StoreFileWriter.java:787) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.StoreFileWriter.close(StoreFileWriter.java:294) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.StoreFlusher.finalizeWriter(StoreFlusher.java:70) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.DefaultStoreFlusher.flushSnapshot(DefaultStoreFlusher.java:74) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore.flushCache(HStore.java:832) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl.flushCache(HStore.java:1975) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushCacheAndCommit(HRegion.java:3029) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2737) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2709) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.flushcache(HRegion.java:2579) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.flushcache(HRegion.java:2502) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.flush(HRegion.java:2472) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.FlushRegionCallable.doCall(FlushRegionCallable.java:56) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.BaseRSProcedureCallable.call(BaseRSProcedureCallable.java:35) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.BaseRSProcedureCallable.call(BaseRSProcedureCallable.java:23) ~[classes/:3.0.0-beta-2-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.handler.RSProcedureHandler.process(RSProcedureHandler.java:56) ~[classes/:?] at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:104) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:16,726 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775725_1010 (size=34) 2024-11-08T08:51:16,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775709_1013 (size=1321) 2024-11-08T08:51:16,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775772_1004 (size=42) 2024-11-08T08:51:16,729 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775773_1004 (size=42) 2024-11-08T08:51:16,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775708_1013 (size=1321) 2024-11-08T08:51:16,730 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775757_1006 (size=196) 2024-11-08T08:51:16,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775756_1006 (size=196) 2024-11-08T08:51:16,765 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:17,074 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:17,585 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:17,692 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/.tmp/cf/0bfa92a3fbdc43c18ec99800d53752bd is 36, key is row/cf:cq/1731055876434/Put/seqid=0 2024-11-08T08:51:17,696 WARN [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:17,696 WARN [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:17,701 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_20974069_22 at /127.0.0.1:40050 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:37237:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:40050 dst: /127.0.0.1:37237 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:17,708 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775648_1025 (size=4787) 2024-11-08T08:51:17,709 WARN [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:17,709 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/.tmp/cf/0bfa92a3fbdc43c18ec99800d53752bd 2024-11-08T08:51:17,751 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/.tmp/cf/0bfa92a3fbdc43c18ec99800d53752bd as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/cf/0bfa92a3fbdc43c18ec99800d53752bd 2024-11-08T08:51:17,762 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/cf/0bfa92a3fbdc43c18ec99800d53752bd, entries=1, sequenceid=5, filesize=4.7 K 2024-11-08T08:51:17,770 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 0aed20f1745462a12c55ec714d3d4a79 in 1149ms, sequenceid=5, compaction requested=false 2024-11-08T08:51:17,771 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-11-08T08:51:17,773 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 0aed20f1745462a12c55ec714d3d4a79: 2024-11-08T08:51:17,773 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:17,774 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-08T08:51:17,776 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-08T08:51:17,782 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-08T08:51:17,782 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 1.3260 sec 2024-11-08T08:51:17,787 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 1.3410 sec 2024-11-08T08:51:18,595 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42533 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:18,595 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-08T08:51:18,608 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-08T08:51:18,609 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-08T08:51:18,609 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:18,613 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:18,613 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:18,613 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-08T08:51:18,614 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-08T08:51:18,614 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1441288993, stopped=false 2024-11-08T08:51:18,614 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=6ac96558fc7f,42533,1731055871860 2024-11-08T08:51:19,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:19,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:19,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:19,206 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:19,206 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:19,206 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:19,206 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-08T08:51:19,206 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:19,206 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:19,206 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-08T08:51:19,207 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:19,207 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,208 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:19,208 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:19,208 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '6ac96558fc7f,35983,1731055873077' ***** 2024-11-08T08:51:19,208 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-08T08:51:19,209 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:19,209 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '6ac96558fc7f,44823,1731055873178' ***** 2024-11-08T08:51:19,209 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:19,210 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-08T08:51:19,210 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-08T08:51:19,210 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '6ac96558fc7f,39923,1731055873233' ***** 2024-11-08T08:51:19,210 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-08T08:51:19,210 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-08T08:51:19,211 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-08T08:51:19,211 INFO [RS:0;6ac96558fc7f:35983 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-08T08:51:19,211 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-08T08:51:19,211 INFO [RS:2;6ac96558fc7f:39923 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-08T08:51:19,212 INFO [RS:0;6ac96558fc7f:35983 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-08T08:51:19,212 INFO [RS:2;6ac96558fc7f:39923 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-08T08:51:19,212 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-08T08:51:19,212 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(959): stopping server 6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:19,212 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-08T08:51:19,212 INFO [RS:1;6ac96558fc7f:44823 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-08T08:51:19,212 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(3091): Received CLOSE for 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:19,212 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:19,212 INFO [RS:1;6ac96558fc7f:44823 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-08T08:51:19,212 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(959): stopping server 6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:19,212 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:19,212 INFO [RS:2;6ac96558fc7f:39923 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;6ac96558fc7f:39923. 2024-11-08T08:51:19,213 INFO [RS:1;6ac96558fc7f:44823 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;6ac96558fc7f:44823. 2024-11-08T08:51:19,213 DEBUG [RS:2;6ac96558fc7f:39923 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:19,213 DEBUG [RS:2;6ac96558fc7f:39923 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,213 DEBUG [RS:1;6ac96558fc7f:44823 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:19,213 DEBUG [RS:1;6ac96558fc7f:44823 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,213 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(959): stopping server 6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:19,213 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(976): stopping server 6ac96558fc7f,39923,1731055873233; all regions closed. 2024-11-08T08:51:19,213 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:19,213 INFO [RS:0;6ac96558fc7f:35983 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;6ac96558fc7f:35983. 2024-11-08T08:51:19,213 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-08T08:51:19,213 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-08T08:51:19,213 DEBUG [RS:0;6ac96558fc7f:35983 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:19,213 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-08T08:51:19,214 DEBUG [RS:0;6ac96558fc7f:35983 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,214 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 0aed20f1745462a12c55ec714d3d4a79, disabling compactions & flushes 2024-11-08T08:51:19,214 INFO [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:19,214 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:19,214 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1325): Online Regions={0aed20f1745462a12c55ec714d3d4a79=TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79.} 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. after waiting 0 ms 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:19,214 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-08T08:51:19,214 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-08T08:51:19,214 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-08T08:51:19,214 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-08T08:51:19,215 DEBUG [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-08T08:51:19,215 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-08T08:51:19,215 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-08T08:51:19,216 DEBUG [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1351): Waiting on 0aed20f1745462a12c55ec714d3d4a79 2024-11-08T08:51:19,220 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_1073741827_1017 (size=93) 2024-11-08T08:51:19,221 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_1073741827_1017 (size=93) 2024-11-08T08:51:19,227 DEBUG [RS:2;6ac96558fc7f:39923 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs 2024-11-08T08:51:19,227 INFO [RS:2;6ac96558fc7f:39923 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 6ac96558fc7f%2C39923%2C1731055873233:(num 1731055874854) 2024-11-08T08:51:19,227 DEBUG [RS:2;6ac96558fc7f:39923 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.ChoreService(370): Chore service for: regionserver/6ac96558fc7f:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-08T08:51:19,228 INFO [regionserver/6ac96558fc7f:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-08T08:51:19,228 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:19,229 INFO [RS:2;6ac96558fc7f:39923 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:39923 2024-11-08T08:51:19,232 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/default/TestHBaseWalOnEC/0aed20f1745462a12c55ec714d3d4a79/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-08T08:51:19,235 INFO [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:19,235 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 0aed20f1745462a12c55ec714d3d4a79: Waiting for close lock at 1731055879213Running coprocessor pre-close hooks at 1731055879214 (+1 ms)Disabling compacts and flushes for region at 1731055879214Disabling writes for close at 1731055879214Writing region close event to WAL at 1731055879217 (+3 ms)Running coprocessor post-close hooks at 1731055879233 (+16 ms)Closed at 1731055879235 (+2 ms) 2024-11-08T08:51:19,236 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79. 2024-11-08T08:51:19,251 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/info/60d764e75391438099c707759b07bb81 is 153, key is TestHBaseWalOnEC,,1731055875760.0aed20f1745462a12c55ec714d3d4a79./info:regioninfo/1731055876216/Put/seqid=0 2024-11-08T08:51:19,254 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:19,254 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:19,261 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1126177094_22 at /127.0.0.1:41754 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41754 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:19,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775632_1027 (size=6637) 2024-11-08T08:51:19,267 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:19,267 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/info/60d764e75391438099c707759b07bb81 2024-11-08T08:51:19,288 INFO [regionserver/6ac96558fc7f:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:19,288 INFO [regionserver/6ac96558fc7f:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:19,296 INFO [regionserver/6ac96558fc7f:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:19,299 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/ns/d6115071a0564719b76142795dfe2bc5 is 43, key is default/ns:d/1731055875536/Put/seqid=0 2024-11-08T08:51:19,302 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:19,302 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:19,306 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1126177094_22 at /127.0.0.1:49652 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:39689:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49652 dst: /127.0.0.1:39689 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:19,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775616_1029 (size=5153) 2024-11-08T08:51:19,311 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:19,312 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/ns/d6115071a0564719b76142795dfe2bc5 2024-11-08T08:51:19,341 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/table/722e0f78d76043abaaf581c8bf4e39aa is 52, key is TestHBaseWalOnEC/table:state/1731055876233/Put/seqid=0 2024-11-08T08:51:19,344 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:19,344 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:19,347 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1126177094_22 at /127.0.0.1:41794 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41794 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:19,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775600_1031 (size=5249) 2024-11-08T08:51:19,352 WARN [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:19,352 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/table/722e0f78d76043abaaf581c8bf4e39aa 2024-11-08T08:51:19,363 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/info/60d764e75391438099c707759b07bb81 as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/info/60d764e75391438099c707759b07bb81 2024-11-08T08:51:19,373 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/info/60d764e75391438099c707759b07bb81, entries=10, sequenceid=11, filesize=6.5 K 2024-11-08T08:51:19,375 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/ns/d6115071a0564719b76142795dfe2bc5 as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/ns/d6115071a0564719b76142795dfe2bc5 2024-11-08T08:51:19,385 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/ns/d6115071a0564719b76142795dfe2bc5, entries=2, sequenceid=11, filesize=5.0 K 2024-11-08T08:51:19,387 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/.tmp/table/722e0f78d76043abaaf581c8bf4e39aa as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/table/722e0f78d76043abaaf581c8bf4e39aa 2024-11-08T08:51:19,398 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/table/722e0f78d76043abaaf581c8bf4e39aa, entries=2, sequenceid=11, filesize=5.1 K 2024-11-08T08:51:19,400 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 185ms, sequenceid=11, compaction requested=false 2024-11-08T08:51:19,400 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-11-08T08:51:19,412 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-08T08:51:19,413 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-08T08:51:19,413 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-08T08:51:19,413 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731055879214Running coprocessor pre-close hooks at 1731055879214Disabling compacts and flushes for region at 1731055879214Disabling writes for close at 1731055879215 (+1 ms)Obtaining lock to block concurrent updates at 1731055879215Preparing flush snapshotting stores in 1588230740 at 1731055879215Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1731055879216 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1731055879217 (+1 ms)Flushing 1588230740/info: creating writer at 1731055879217Flushing 1588230740/info: appending metadata at 1731055879246 (+29 ms)Flushing 1588230740/info: closing flushed file at 1731055879246Flushing 1588230740/ns: creating writer at 1731055879280 (+34 ms)Flushing 1588230740/ns: appending metadata at 1731055879298 (+18 ms)Flushing 1588230740/ns: closing flushed file at 1731055879299 (+1 ms)Flushing 1588230740/table: creating writer at 1731055879322 (+23 ms)Flushing 1588230740/table: appending metadata at 1731055879340 (+18 ms)Flushing 1588230740/table: closing flushed file at 1731055879340Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@348e0b1a: reopening flushed file at 1731055879361 (+21 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@716748e3: reopening flushed file at 1731055879374 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@58706e90: reopening flushed file at 1731055879385 (+11 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 185ms, sequenceid=11, compaction requested=false at 1731055879400 (+15 ms)Writing region close event to WAL at 1731055879402 (+2 ms)Running coprocessor post-close hooks at 1731055879413 (+11 ms)Closed at 1731055879413 2024-11-08T08:51:19,414 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-08T08:51:19,415 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(976): stopping server 6ac96558fc7f,44823,1731055873178; all regions closed. 2024-11-08T08:51:19,416 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(976): stopping server 6ac96558fc7f,35983,1731055873077; all regions closed. 2024-11-08T08:51:19,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_1073741829_1019 (size=2751) 2024-11-08T08:51:19,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_1073741829_1019 (size=2751) 2024-11-08T08:51:19,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_1073741828_1018 (size=1298) 2024-11-08T08:51:19,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_1073741828_1018 (size=1298) 2024-11-08T08:51:19,420 WARN [Close-WAL-Writer-0 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(650): complete file /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/WALs/6ac96558fc7f,35983,1731055873077/6ac96558fc7f%2C35983%2C1731055873077.1731055874854 not finished, retry = 0 2024-11-08T08:51:19,423 DEBUG [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs 2024-11-08T08:51:19,423 INFO [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 6ac96558fc7f%2C44823%2C1731055873178.meta:.meta(num 1731055875349) 2024-11-08T08:51:19,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_1073741826_1016 (size=93) 2024-11-08T08:51:19,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_1073741826_1016 (size=93) 2024-11-08T08:51:19,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_1073741826_1016 (size=93) 2024-11-08T08:51:19,434 DEBUG [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs 2024-11-08T08:51:19,434 INFO [RS:1;6ac96558fc7f:44823 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 6ac96558fc7f%2C44823%2C1731055873178:(num 1731055874854) 2024-11-08T08:51:19,434 DEBUG [RS:1;6ac96558fc7f:44823 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,434 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:19,434 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:19,434 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.ChoreService(370): Chore service for: regionserver/6ac96558fc7f:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:19,435 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:19,435 INFO [regionserver/6ac96558fc7f:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:19,435 INFO [RS:1;6ac96558fc7f:44823 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:44823 2024-11-08T08:51:19,527 DEBUG [RS:0;6ac96558fc7f:35983 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/oldWALs 2024-11-08T08:51:19,527 INFO [RS:0;6ac96558fc7f:35983 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 6ac96558fc7f%2C35983%2C1731055873077:(num 1731055874854) 2024-11-08T08:51:19,527 DEBUG [RS:0;6ac96558fc7f:35983 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:19,527 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:19,527 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:19,528 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.ChoreService(370): Chore service for: regionserver/6ac96558fc7f:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:19,528 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-08T08:51:19,528 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-08T08:51:19,528 INFO [regionserver/6ac96558fc7f:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:19,528 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-08T08:51:19,529 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:19,529 INFO [RS:0;6ac96558fc7f:35983 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:35983 2024-11-08T08:51:19,540 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/6ac96558fc7f,39923,1731055873233 2024-11-08T08:51:19,540 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-08T08:51:19,540 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:20,013 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775676_1021 (size=392) 2024-11-08T08:51:20,014 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775644_1025 (size=4787) 2024-11-08T08:51:20,014 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775645_1025 (size=4787) 2024-11-08T08:51:20,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775677_1021 (size=392) 2024-11-08T08:51:20,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775740_1008 (size=1189) 2024-11-08T08:51:20,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775741_1008 (size=1189) 2024-11-08T08:51:20,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775724_1010 (size=34) 2024-11-08T08:51:20,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_-9223372036854775693_1015 (size=32) 2024-11-08T08:51:20,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39689 is added to blk_-9223372036854775692_1015 (size=32) 2024-11-08T08:51:20,109 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/6ac96558fc7f,44823,1731055873178 2024-11-08T08:51:20,109 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:20,109 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/6ac96558fc7f,35983,1731055873077 2024-11-08T08:51:20,110 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:20,276 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [6ac96558fc7f,39923,1731055873233] 2024-11-08T08:51:20,371 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/6ac96558fc7f,39923,1731055873233 already deleted, retry=false 2024-11-08T08:51:20,371 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 6ac96558fc7f,39923,1731055873233 expired; onlineServers=2 2024-11-08T08:51:20,371 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [6ac96558fc7f,44823,1731055873178] 2024-11-08T08:51:20,376 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:20,376 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39923-0x10119a460090003, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:20,376 INFO [RS:2;6ac96558fc7f:39923 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:20,376 INFO [RS:2;6ac96558fc7f:39923 {}] regionserver.HRegionServer(1031): Exiting; stopping=6ac96558fc7f,39923,1731055873233; zookeeper connection closed. 2024-11-08T08:51:20,377 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@2dbf0007 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@2dbf0007 2024-11-08T08:51:20,382 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/6ac96558fc7f,44823,1731055873178 already deleted, retry=false 2024-11-08T08:51:20,382 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 6ac96558fc7f,44823,1731055873178 expired; onlineServers=1 2024-11-08T08:51:20,382 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [6ac96558fc7f,35983,1731055873077] 2024-11-08T08:51:20,392 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/6ac96558fc7f,35983,1731055873077 already deleted, retry=false 2024-11-08T08:51:20,392 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 6ac96558fc7f,35983,1731055873077 expired; onlineServers=0 2024-11-08T08:51:20,393 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '6ac96558fc7f,42533,1731055871860' ***** 2024-11-08T08:51:20,393 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-08T08:51:20,393 INFO [M:0;6ac96558fc7f:42533 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:20,393 INFO [M:0;6ac96558fc7f:42533 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:20,393 DEBUG [M:0;6ac96558fc7f:42533 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-08T08:51:20,393 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-08T08:51:20,393 DEBUG [M:0;6ac96558fc7f:42533 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-08T08:51:20,393 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.small.0-1731055874460 {}] cleaner.HFileCleaner(306): Exit Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.small.0-1731055874460,5,FailOnTimeoutGroup] 2024-11-08T08:51:20,393 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.large.0-1731055874458 {}] cleaner.HFileCleaner(306): Exit Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.large.0-1731055874458,5,FailOnTimeoutGroup] 2024-11-08T08:51:20,394 INFO [M:0;6ac96558fc7f:42533 {}] hbase.ChoreService(370): Chore service for: master/6ac96558fc7f:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:20,394 INFO [M:0;6ac96558fc7f:42533 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:20,394 DEBUG [M:0;6ac96558fc7f:42533 {}] master.HMaster(1795): Stopping service threads 2024-11-08T08:51:20,394 INFO [M:0;6ac96558fc7f:42533 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-08T08:51:20,394 INFO [M:0;6ac96558fc7f:42533 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-08T08:51:20,395 INFO [M:0;6ac96558fc7f:42533 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-08T08:51:20,395 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-08T08:51:20,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:20,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:20,403 DEBUG [M:0;6ac96558fc7f:42533 {}] zookeeper.ZKUtil(347): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-08T08:51:20,403 WARN [M:0;6ac96558fc7f:42533 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-08T08:51:20,404 INFO [M:0;6ac96558fc7f:42533 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/.lastflushedseqids 2024-11-08T08:51:20,413 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:20,413 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:20,415 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41846 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41846 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:20,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775584_1033 (size=127) 2024-11-08T08:51:20,419 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:20,420 INFO [M:0;6ac96558fc7f:42533 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-08T08:51:20,420 INFO [M:0;6ac96558fc7f:42533 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-08T08:51:20,420 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-08T08:51:20,420 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:20,420 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:20,420 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-08T08:51:20,420 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:20,420 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.82 KB heapSize=34.11 KB 2024-11-08T08:51:20,438 DEBUG [M:0;6ac96558fc7f:42533 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/05152fcc909d4d9facbf8487f1ca487d is 82, key is hbase:meta,,1/info:regioninfo/1731055875434/Put/seqid=0 2024-11-08T08:51:20,441 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:20,441 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:20,444 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41864 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41864 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:20,447 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:20,447 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:20,447 INFO [RS:1;6ac96558fc7f:44823 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:20,447 INFO [RS:0;6ac96558fc7f:35983 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:20,447 DEBUG [pool-71-thread-1-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44823-0x10119a460090002, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:20,447 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35983-0x10119a460090001, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:20,447 INFO [RS:1;6ac96558fc7f:44823 {}] regionserver.HRegionServer(1031): Exiting; stopping=6ac96558fc7f,44823,1731055873178; zookeeper connection closed. 2024-11-08T08:51:20,447 INFO [RS:0;6ac96558fc7f:35983 {}] regionserver.HRegionServer(1031): Exiting; stopping=6ac96558fc7f,35983,1731055873077; zookeeper connection closed. 2024-11-08T08:51:20,448 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7a4ea9b2 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7a4ea9b2 2024-11-08T08:51:20,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775568_1035 (size=5672) 2024-11-08T08:51:20,451 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@256265cc {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@256265cc 2024-11-08T08:51:20,451 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-08T08:51:20,851 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:20,852 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/05152fcc909d4d9facbf8487f1ca487d 2024-11-08T08:51:20,881 DEBUG [M:0;6ac96558fc7f:42533 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/9bd396ba36a14267afe80ee17064f628 is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1731055876241/Put/seqid=0 2024-11-08T08:51:20,883 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:20,883 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:20,886 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41876 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41876 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:20,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775552_1037 (size=6438) 2024-11-08T08:51:20,894 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-hbase.properties,hadoop-metrics2.properties 2024-11-08T08:51:21,291 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:21,292 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/9bd396ba36a14267afe80ee17064f628 2024-11-08T08:51:21,324 DEBUG [M:0;6ac96558fc7f:42533 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/f5d053d0021d46aca411f1ce12a69aa3 is 69, key is 6ac96558fc7f,35983,1731055873077/rs:state/1731055874530/Put/seqid=0 2024-11-08T08:51:21,326 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:21,326 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-11-08T08:51:21,329 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1962198541_22 at /127.0.0.1:41898 [Receiving block BP-1794552092-172.17.0.3-1731055867092:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:41019:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41898 dst: /127.0.0.1:41019 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-11-08T08:51:21,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_-9223372036854775536_1039 (size=5294) 2024-11-08T08:51:21,336 WARN [M:0;6ac96558fc7f:42533 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-11-08T08:51:21,336 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/f5d053d0021d46aca411f1ce12a69aa3 2024-11-08T08:51:21,344 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/05152fcc909d4d9facbf8487f1ca487d as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/05152fcc909d4d9facbf8487f1ca487d 2024-11-08T08:51:21,352 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/05152fcc909d4d9facbf8487f1ca487d, entries=8, sequenceid=72, filesize=5.5 K 2024-11-08T08:51:21,353 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/9bd396ba36a14267afe80ee17064f628 as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/9bd396ba36a14267afe80ee17064f628 2024-11-08T08:51:21,362 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/9bd396ba36a14267afe80ee17064f628, entries=8, sequenceid=72, filesize=6.3 K 2024-11-08T08:51:21,364 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/f5d053d0021d46aca411f1ce12a69aa3 as hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/f5d053d0021d46aca411f1ce12a69aa3 2024-11-08T08:51:21,373 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/f5d053d0021d46aca411f1ce12a69aa3, entries=3, sequenceid=72, filesize=5.2 K 2024-11-08T08:51:21,375 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 954ms, sequenceid=72, compaction requested=false 2024-11-08T08:51:21,376 INFO [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:21,377 DEBUG [M:0;6ac96558fc7f:42533 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731055880420Disabling compacts and flushes for region at 1731055880420Disabling writes for close at 1731055880420Obtaining lock to block concurrent updates at 1731055880420Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1731055880420Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27462, getHeapSize=34864, getOffHeapSize=0, getCellsCount=85 at 1731055880421 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1731055880422 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1731055880422Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1731055880438 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1731055880438Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1731055880864 (+426 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1731055880880 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1731055880880Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1731055881306 (+426 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1731055881324 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1731055881324Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3fbea8c: reopening flushed file at 1731055881343 (+19 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@7124f23b: reopening flushed file at 1731055881352 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@cb82513: reopening flushed file at 1731055881362 (+10 ms)Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 954ms, sequenceid=72, compaction requested=false at 1731055881375 (+13 ms)Writing region close event to WAL at 1731055881376 (+1 ms)Closed at 1731055881376 2024-11-08T08:51:21,380 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41019 is added to blk_1073741825_1011 (size=32665) 2024-11-08T08:51:21,380 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37237 is added to blk_1073741825_1011 (size=32665) 2024-11-08T08:51:21,381 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:21,381 INFO [M:0;6ac96558fc7f:42533 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-08T08:51:21,381 INFO [M:0;6ac96558fc7f:42533 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:42533 2024-11-08T08:51:21,381 INFO [M:0;6ac96558fc7f:42533 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:21,550 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:21,550 INFO [M:0;6ac96558fc7f:42533 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:21,550 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42533-0x10119a460090000, quorum=127.0.0.1:50272, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:21,589 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3114ae69{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:21,595 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3c70a874{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:21,595 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:21,595 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5822645a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:21,595 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@16cd567f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:21,599 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-08T08:51:21,599 WARN [BP-1794552092-172.17.0.3-1731055867092 heartbeating to localhost/127.0.0.1:36363 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-08T08:51:21,599 WARN [BP-1794552092-172.17.0.3-1731055867092 heartbeating to localhost/127.0.0.1:36363 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1794552092-172.17.0.3-1731055867092 (Datanode Uuid 9da036be-4ee2-4e34-b32c-cac8bf124248) service to localhost/127.0.0.1:36363 2024-11-08T08:51:21,599 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-08T08:51:21,600 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data5/current/BP-1794552092-172.17.0.3-1731055867092 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:21,600 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data6/current/BP-1794552092-172.17.0.3-1731055867092 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:21,600 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-08T08:51:21,603 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@353955e9{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:21,603 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@11738cd8{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:21,603 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:21,604 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40eb7053{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:21,604 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@510fec09{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:21,605 WARN [BP-1794552092-172.17.0.3-1731055867092 heartbeating to localhost/127.0.0.1:36363 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-08T08:51:21,605 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-08T08:51:21,605 WARN [BP-1794552092-172.17.0.3-1731055867092 heartbeating to localhost/127.0.0.1:36363 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1794552092-172.17.0.3-1731055867092 (Datanode Uuid 4192a270-ddb6-4969-9470-f4d8ec183401) service to localhost/127.0.0.1:36363 2024-11-08T08:51:21,605 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-08T08:51:21,606 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data3/current/BP-1794552092-172.17.0.3-1731055867092 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:21,606 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data4/current/BP-1794552092-172.17.0.3-1731055867092 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:21,606 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-08T08:51:21,609 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1b97a472{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:21,610 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3722a29b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:21,610 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:21,610 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@69893329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:21,610 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3a5de9e4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:21,611 WARN [BP-1794552092-172.17.0.3-1731055867092 heartbeating to localhost/127.0.0.1:36363 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-08T08:51:21,611 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-08T08:51:21,611 WARN [BP-1794552092-172.17.0.3-1731055867092 heartbeating to localhost/127.0.0.1:36363 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1794552092-172.17.0.3-1731055867092 (Datanode Uuid 060e9f99-6979-4f24-a4ff-dab4db6244fb) service to localhost/127.0.0.1:36363 2024-11-08T08:51:21,611 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-08T08:51:21,612 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-08T08:51:21,616 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data2/current/BP-1794552092-172.17.0.3-1731055867092 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:21,616 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/cluster_a487b40c-e2f9-73c7-a1f5-5206253d0155/data/data1/current/BP-1794552092-172.17.0.3-1731055867092 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:21,624 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@62d6efd9{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-08T08:51:21,624 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@353d35a1{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:21,625 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:21,625 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@ce709a8{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:21,625 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@760c69c0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:21,633 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-08T08:51:21,661 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-08T08:51:21,669 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=89 (was 158), OpenFileDescriptor=433 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=121 (was 98) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=9195 (was 9504) 2024-11-08T08:51:21,675 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=89, OpenFileDescriptor=433, MaxFileDescriptor=1048576, SystemLoadAverage=121, ProcessCount=11, AvailableMemoryMB=9195 2024-11-08T08:51:21,676 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-11-08T08:51:21,676 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.log.dir so I do NOT create it in target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400 2024-11-08T08:51:21,676 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/8f79ed63-853b-cd97-4692-3c78fd9ff989/hadoop.tmp.dir so I do NOT create it in target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400 2024-11-08T08:51:21,676 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7, deleteOnExit=true 2024-11-08T08:51:21,676 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/test.cache.data in system properties and HBase conf 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.tmp.dir in system properties and HBase conf 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir in system properties and HBase conf 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/mapreduce.cluster.local.dir in system properties and HBase conf 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-11-08T08:51:21,677 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-11-08T08:51:21,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-08T08:51:21,678 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-11-08T08:51:21,679 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/nfs.dump.dir in system properties and HBase conf 2024-11-08T08:51:21,679 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/java.io.tmpdir in system properties and HBase conf 2024-11-08T08:51:21,679 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/dfs.journalnode.edits.dir in system properties and HBase conf 2024-11-08T08:51:21,679 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-11-08T08:51:21,679 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-11-08T08:51:22,184 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:22,190 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:22,197 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:22,197 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:22,197 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:22,198 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:22,199 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@642ad787{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:22,199 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@778ab8c6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:22,293 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@26e4a54d{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/java.io.tmpdir/jetty-localhost-36933-hadoop-hdfs-3_4_1-tests_jar-_-any-13143419870960541076/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-08T08:51:22,294 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@249bb0e5{HTTP/1.1, (http/1.1)}{localhost:36933} 2024-11-08T08:51:22,294 INFO [Time-limited test {}] server.Server(415): Started @17078ms 2024-11-08T08:51:22,642 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:22,645 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:22,647 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:22,647 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:22,647 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:22,648 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@22e42287{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:22,648 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@49cc577f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:22,743 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6e45ef60{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/java.io.tmpdir/jetty-localhost-42439-hadoop-hdfs-3_4_1-tests_jar-_-any-17114443368069322646/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:22,743 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6be33115{HTTP/1.1, (http/1.1)}{localhost:42439} 2024-11-08T08:51:22,743 INFO [Time-limited test {}] server.Server(415): Started @17528ms 2024-11-08T08:51:22,745 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-08T08:51:22,781 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:22,784 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:22,785 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:22,785 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:22,785 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:22,786 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5ffbd1c8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:22,786 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@19cdd4f9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:22,882 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@47310287{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/java.io.tmpdir/jetty-localhost-44657-hadoop-hdfs-3_4_1-tests_jar-_-any-5119550135831350304/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:22,882 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2b50a4e4{HTTP/1.1, (http/1.1)}{localhost:44657} 2024-11-08T08:51:22,883 INFO [Time-limited test {}] server.Server(415): Started @17667ms 2024-11-08T08:51:22,884 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-08T08:51:22,913 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-11-08T08:51:22,916 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-11-08T08:51:22,916 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-11-08T08:51:22,917 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-11-08T08:51:22,917 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-11-08T08:51:22,917 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6229ef54{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,AVAILABLE} 2024-11-08T08:51:22,918 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@15b92333{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-11-08T08:51:23,016 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7261385{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/java.io.tmpdir/jetty-localhost-33293-hadoop-hdfs-3_4_1-tests_jar-_-any-7249908954780295280/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:23,017 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@242a3163{HTTP/1.1, (http/1.1)}{localhost:33293} 2024-11-08T08:51:23,017 INFO [Time-limited test {}] server.Server(415): Started @17802ms 2024-11-08T08:51:23,023 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-11-08T08:51:24,151 WARN [Thread-594 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data1/current/BP-614830020-172.17.0.3-1731055881703/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:24,153 WARN [Thread-595 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data2/current/BP-614830020-172.17.0.3-1731055881703/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:24,174 WARN [Thread-534 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-08T08:51:24,177 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xaf5ac58787070295 with lease ID 0x45714478bf9c184a: Processing first storage report for DS-f63f22d4-58f6-4ff2-9901-7f12b9eb6c25 from datanode DatanodeRegistration(127.0.0.1:36041, datanodeUuid=9618e1bd-b775-4920-a84c-544796c2f2d0, infoPort=41195, infoSecurePort=0, ipcPort=33831, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703) 2024-11-08T08:51:24,177 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xaf5ac58787070295 with lease ID 0x45714478bf9c184a: from storage DS-f63f22d4-58f6-4ff2-9901-7f12b9eb6c25 node DatanodeRegistration(127.0.0.1:36041, datanodeUuid=9618e1bd-b775-4920-a84c-544796c2f2d0, infoPort=41195, infoSecurePort=0, ipcPort=33831, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:24,177 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xaf5ac58787070295 with lease ID 0x45714478bf9c184a: Processing first storage report for DS-2f0eb553-2dc1-4c41-ac95-e2358fe83ecd from datanode DatanodeRegistration(127.0.0.1:36041, datanodeUuid=9618e1bd-b775-4920-a84c-544796c2f2d0, infoPort=41195, infoSecurePort=0, ipcPort=33831, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703) 2024-11-08T08:51:24,177 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xaf5ac58787070295 with lease ID 0x45714478bf9c184a: from storage DS-2f0eb553-2dc1-4c41-ac95-e2358fe83ecd node DatanodeRegistration(127.0.0.1:36041, datanodeUuid=9618e1bd-b775-4920-a84c-544796c2f2d0, infoPort=41195, infoSecurePort=0, ipcPort=33831, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:24,486 WARN [Thread-605 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data3/current/BP-614830020-172.17.0.3-1731055881703/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:24,486 WARN [Thread-606 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data4/current/BP-614830020-172.17.0.3-1731055881703/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:24,506 WARN [Thread-557 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-08T08:51:24,509 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xbe589cac666144a2 with lease ID 0x45714478bf9c184b: Processing first storage report for DS-fc1d3247-b4af-48b5-aa83-2e1e6667a348 from datanode DatanodeRegistration(127.0.0.1:36421, datanodeUuid=1de92640-8f18-4d49-b1a2-248977cbb17b, infoPort=43555, infoSecurePort=0, ipcPort=37849, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703) 2024-11-08T08:51:24,509 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbe589cac666144a2 with lease ID 0x45714478bf9c184b: from storage DS-fc1d3247-b4af-48b5-aa83-2e1e6667a348 node DatanodeRegistration(127.0.0.1:36421, datanodeUuid=1de92640-8f18-4d49-b1a2-248977cbb17b, infoPort=43555, infoSecurePort=0, ipcPort=37849, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:24,509 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xbe589cac666144a2 with lease ID 0x45714478bf9c184b: Processing first storage report for DS-0d22acf3-4737-4437-b21c-88e39894039d from datanode DatanodeRegistration(127.0.0.1:36421, datanodeUuid=1de92640-8f18-4d49-b1a2-248977cbb17b, infoPort=43555, infoSecurePort=0, ipcPort=37849, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703) 2024-11-08T08:51:24,509 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbe589cac666144a2 with lease ID 0x45714478bf9c184b: from storage DS-0d22acf3-4737-4437-b21c-88e39894039d node DatanodeRegistration(127.0.0.1:36421, datanodeUuid=1de92640-8f18-4d49-b1a2-248977cbb17b, infoPort=43555, infoSecurePort=0, ipcPort=37849, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-11-08T08:51:24,529 WARN [Thread-616 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data5/current/BP-614830020-172.17.0.3-1731055881703/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:24,529 WARN [Thread-617 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data6/current/BP-614830020-172.17.0.3-1731055881703/current, will proceed with Du for space computation calculation, 2024-11-08T08:51:24,548 WARN [Thread-579 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-11-08T08:51:24,551 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3cfb7d07ba30b941 with lease ID 0x45714478bf9c184c: Processing first storage report for DS-5ee3f843-e74a-45db-9ffe-8b9d09ada6ca from datanode DatanodeRegistration(127.0.0.1:38231, datanodeUuid=b9fc0088-a82b-4736-9c38-4f3597a87d67, infoPort=35871, infoSecurePort=0, ipcPort=35309, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703) 2024-11-08T08:51:24,551 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3cfb7d07ba30b941 with lease ID 0x45714478bf9c184c: from storage DS-5ee3f843-e74a-45db-9ffe-8b9d09ada6ca node DatanodeRegistration(127.0.0.1:38231, datanodeUuid=b9fc0088-a82b-4736-9c38-4f3597a87d67, infoPort=35871, infoSecurePort=0, ipcPort=35309, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:24,551 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3cfb7d07ba30b941 with lease ID 0x45714478bf9c184c: Processing first storage report for DS-70772705-ac08-4dee-9552-e985030b6816 from datanode DatanodeRegistration(127.0.0.1:38231, datanodeUuid=b9fc0088-a82b-4736-9c38-4f3597a87d67, infoPort=35871, infoSecurePort=0, ipcPort=35309, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703) 2024-11-08T08:51:24,551 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3cfb7d07ba30b941 with lease ID 0x45714478bf9c184c: from storage DS-70772705-ac08-4dee-9552-e985030b6816 node DatanodeRegistration(127.0.0.1:38231, datanodeUuid=b9fc0088-a82b-4736-9c38-4f3597a87d67, infoPort=35871, infoSecurePort=0, ipcPort=35309, storageInfo=lv=-57;cid=testClusterID;nsid=545764125;c=1731055881703), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-11-08T08:51:24,571 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400 2024-11-08T08:51:24,575 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/zookeeper_0, clientPort=64730, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-11-08T08:51:24,576 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=64730 2024-11-08T08:51:24,576 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,578 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741825_1001 (size=7) 2024-11-08T08:51:24,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741825_1001 (size=7) 2024-11-08T08:51:24,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741825_1001 (size=7) 2024-11-08T08:51:24,596 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90 with version=8 2024-11-08T08:51:24,596 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:36363/user/jenkins/test-data/eda57a83-efe4-f75b-95ad-3448a52b0d36/hbase-staging 2024-11-08T08:51:24,598 INFO [Time-limited test {}] client.ConnectionUtils(128): master/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-11-08T08:51:24,598 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:24,599 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46451 2024-11-08T08:51:24,600 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:46451 connecting to ZooKeeper ensemble=127.0.0.1:64730 2024-11-08T08:51:24,655 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:464510x0, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:24,655 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:46451-0x10119a494c90000 connected 2024-11-08T08:51:24,744 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,746 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,749 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:24,750 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90, hbase.cluster.distributed=false 2024-11-08T08:51:24,752 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:24,753 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46451 2024-11-08T08:51:24,753 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46451 2024-11-08T08:51:24,753 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46451 2024-11-08T08:51:24,754 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46451 2024-11-08T08:51:24,754 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46451 2024-11-08T08:51:24,773 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-08T08:51:24,773 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:24,774 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:46683 2024-11-08T08:51:24,775 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:46683 connecting to ZooKeeper ensemble=127.0.0.1:64730 2024-11-08T08:51:24,776 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,778 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,792 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:466830x0, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:24,793 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:46683-0x10119a494c90001 connected 2024-11-08T08:51:24,793 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:24,793 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-08T08:51:24,794 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-08T08:51:24,795 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-08T08:51:24,796 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:24,797 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=46683 2024-11-08T08:51:24,797 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=46683 2024-11-08T08:51:24,798 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=46683 2024-11-08T08:51:24,798 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=46683 2024-11-08T08:51:24,798 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=46683 2024-11-08T08:51:24,817 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:24,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,817 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:24,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,817 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:24,818 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-08T08:51:24,818 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:24,818 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:39649 2024-11-08T08:51:24,820 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39649 connecting to ZooKeeper ensemble=127.0.0.1:64730 2024-11-08T08:51:24,820 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,822 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,834 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:396490x0, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:24,835 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39649-0x10119a494c90002 connected 2024-11-08T08:51:24,835 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:24,835 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-08T08:51:24,836 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-08T08:51:24,837 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-08T08:51:24,838 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:24,841 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39649 2024-11-08T08:51:24,841 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39649 2024-11-08T08:51:24,844 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39649 2024-11-08T08:51:24,844 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39649 2024-11-08T08:51:24,845 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39649 2024-11-08T08:51:24,861 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/6ac96558fc7f:0 server-side Connection retries=45 2024-11-08T08:51:24,861 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,861 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,862 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-11-08T08:51:24,862 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-11-08T08:51:24,862 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-11-08T08:51:24,862 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-11-08T08:51:24,862 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-11-08T08:51:24,862 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.3:42453 2024-11-08T08:51:24,864 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:42453 connecting to ZooKeeper ensemble=127.0.0.1:64730 2024-11-08T08:51:24,865 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,866 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,881 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:424530x0, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-11-08T08:51:24,882 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:424530x0, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:24,882 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:42453-0x10119a494c90003 connected 2024-11-08T08:51:24,882 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-11-08T08:51:24,882 DEBUG [Time-limited test {}] mob.MobFileCache(123): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-11-08T08:51:24,883 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-11-08T08:51:24,884 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-11-08T08:51:24,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42453 2024-11-08T08:51:24,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42453 2024-11-08T08:51:24,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42453 2024-11-08T08:51:24,886 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42453 2024-11-08T08:51:24,886 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42453 2024-11-08T08:51:24,896 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;6ac96558fc7f:46451 2024-11-08T08:51:24,897 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:24,908 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,908 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,908 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,908 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,908 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:24,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:24,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:24,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:24,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,918 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,919 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,919 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-11-08T08:51:24,920 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/6ac96558fc7f,46451,1731055884597 from backup master directory 2024-11-08T08:51:24,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:24,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,929 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:24,929 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:24,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-11-08T08:51:24,935 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/hbase.id] with ID: bd746c6f-4269-47fe-8a62-117605dafd0f 2024-11-08T08:51:24,935 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/.tmp/hbase.id 2024-11-08T08:51:24,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741826_1002 (size=42) 2024-11-08T08:51:24,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741826_1002 (size=42) 2024-11-08T08:51:24,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741826_1002 (size=42) 2024-11-08T08:51:24,949 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/.tmp/hbase.id]:[hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/hbase.id] 2024-11-08T08:51:24,965 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-11-08T08:51:24,965 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-11-08T08:51:24,967 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 2ms. 2024-11-08T08:51:24,976 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,976 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,976 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,976 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:24,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741827_1003 (size=196) 2024-11-08T08:51:24,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741827_1003 (size=196) 2024-11-08T08:51:24,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741827_1003 (size=196) 2024-11-08T08:51:24,988 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-08T08:51:24,989 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-11-08T08:51:24,989 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-08T08:51:25,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741828_1004 (size=1189) 2024-11-08T08:51:25,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741828_1004 (size=1189) 2024-11-08T08:51:25,001 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741828_1004 (size=1189) 2024-11-08T08:51:25,002 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store 2024-11-08T08:51:25,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741829_1005 (size=34) 2024-11-08T08:51:25,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741829_1005 (size=34) 2024-11-08T08:51:25,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741829_1005 (size=34) 2024-11-08T08:51:25,013 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:25,013 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-08T08:51:25,013 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:25,013 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:25,013 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-08T08:51:25,014 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:25,014 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:25,014 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731055885013Disabling compacts and flushes for region at 1731055885013Disabling writes for close at 1731055885013Writing region close event to WAL at 1731055885014 (+1 ms)Closed at 1731055885014 2024-11-08T08:51:25,015 WARN [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/.initializing 2024-11-08T08:51:25,015 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/WALs/6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:25,019 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C46451%2C1731055884597, suffix=, logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/WALs/6ac96558fc7f,46451,1731055884597, archiveDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/oldWALs, maxLogs=10 2024-11-08T08:51:25,019 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 6ac96558fc7f%2C46451%2C1731055884597.1731055885019 2024-11-08T08:51:25,031 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/WALs/6ac96558fc7f,46451,1731055884597/6ac96558fc7f%2C46451%2C1731055884597.1731055885019 2024-11-08T08:51:25,034 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43555:43555),(127.0.0.1/127.0.0.1:35871:35871),(127.0.0.1/127.0.0.1:41195:41195)] 2024-11-08T08:51:25,036 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-11-08T08:51:25,037 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:25,037 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,037 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,039 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,042 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-11-08T08:51:25,042 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,043 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,043 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,046 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-11-08T08:51:25,046 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,046 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:25,047 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,050 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-11-08T08:51:25,050 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,051 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:25,051 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,053 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-11-08T08:51:25,053 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,054 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:25,054 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,055 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,056 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,057 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,057 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,058 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-08T08:51:25,060 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-11-08T08:51:25,063 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-08T08:51:25,064 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69146550, jitterRate=0.030363887548446655}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-08T08:51:25,066 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1731055885037Initializing all the Stores at 1731055885039 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885039Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055885039Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055885039Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055885039Cleaning up temporary data from old regions at 1731055885057 (+18 ms)Region opened successfully at 1731055885065 (+8 ms) 2024-11-08T08:51:25,066 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-11-08T08:51:25,071 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5e1a9adf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:25,072 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-11-08T08:51:25,072 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-11-08T08:51:25,073 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-11-08T08:51:25,073 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-11-08T08:51:25,073 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-11-08T08:51:25,074 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-11-08T08:51:25,074 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-11-08T08:51:25,077 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-11-08T08:51:25,078 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-11-08T08:51:25,087 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-11-08T08:51:25,087 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-11-08T08:51:25,088 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-11-08T08:51:25,097 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-11-08T08:51:25,098 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-11-08T08:51:25,099 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-11-08T08:51:25,108 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-11-08T08:51:25,109 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-11-08T08:51:25,118 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-11-08T08:51:25,121 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-11-08T08:51:25,133 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,144 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,145 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,146 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=6ac96558fc7f,46451,1731055884597, sessionid=0x10119a494c90000, setting cluster-up flag (Was=false) 2024-11-08T08:51:25,165 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,165 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,165 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,165 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,197 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-11-08T08:51:25,198 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:25,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,218 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,249 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-11-08T08:51:25,250 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:25,252 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-11-08T08:51:25,254 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:25,255 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-11-08T08:51:25,255 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-11-08T08:51:25,255 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 6ac96558fc7f,46451,1731055884597 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-11-08T08:51:25,257 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:25,257 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:25,257 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:25,257 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=5, maxPoolSize=5 2024-11-08T08:51:25,257 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/6ac96558fc7f:0, corePoolSize=10, maxPoolSize=10 2024-11-08T08:51:25,257 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,258 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:25,258 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,258 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1731055915258 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-11-08T08:51:25,259 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-11-08T08:51:25,260 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-11-08T08:51:25,260 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-11-08T08:51:25,260 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-11-08T08:51:25,260 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:25,260 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-11-08T08:51:25,261 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.large.0-1731055885260,5,FailOnTimeoutGroup] 2024-11-08T08:51:25,261 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.small.0-1731055885261,5,FailOnTimeoutGroup] 2024-11-08T08:51:25,261 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,261 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-11-08T08:51:25,261 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,261 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,262 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,262 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-11-08T08:51:25,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741831_1007 (size=1321) 2024-11-08T08:51:25,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741831_1007 (size=1321) 2024-11-08T08:51:25,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741831_1007 (size=1321) 2024-11-08T08:51:25,272 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-11-08T08:51:25,272 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90 2024-11-08T08:51:25,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741832_1008 (size=32) 2024-11-08T08:51:25,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741832_1008 (size=32) 2024-11-08T08:51:25,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741832_1008 (size=32) 2024-11-08T08:51:25,282 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:25,283 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-08T08:51:25,285 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-08T08:51:25,285 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,286 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,286 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-08T08:51:25,287 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-08T08:51:25,288 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,288 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,289 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-08T08:51:25,289 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(746): ClusterId : bd746c6f-4269-47fe-8a62-117605dafd0f 2024-11-08T08:51:25,289 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(746): ClusterId : bd746c6f-4269-47fe-8a62-117605dafd0f 2024-11-08T08:51:25,289 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(746): ClusterId : bd746c6f-4269-47fe-8a62-117605dafd0f 2024-11-08T08:51:25,289 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-08T08:51:25,289 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-08T08:51:25,289 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-11-08T08:51:25,290 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-08T08:51:25,290 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,291 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,291 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-08T08:51:25,293 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-08T08:51:25,293 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,293 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,294 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-08T08:51:25,294 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740 2024-11-08T08:51:25,295 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740 2024-11-08T08:51:25,296 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-08T08:51:25,296 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-08T08:51:25,297 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-08T08:51:25,298 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-08T08:51:25,302 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-08T08:51:25,303 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59207567, jitterRate=-0.1177385002374649}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-08T08:51:25,303 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-08T08:51:25,303 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-08T08:51:25,303 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-08T08:51:25,303 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-08T08:51:25,305 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1731055885282Initializing all the Stores at 1731055885283 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885283Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885283Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055885283Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885283Cleaning up temporary data from old regions at 1731055885296 (+13 ms)Region opened successfully at 1731055885304 (+8 ms) 2024-11-08T08:51:25,305 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-08T08:51:25,305 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-08T08:51:25,305 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-08T08:51:25,305 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-08T08:51:25,305 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-08T08:51:25,305 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-08T08:51:25,306 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731055885305Disabling compacts and flushes for region at 1731055885305Disabling writes for close at 1731055885305Writing region close event to WAL at 1731055885305Closed at 1731055885305 2024-11-08T08:51:25,307 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:25,307 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-11-08T08:51:25,308 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-11-08T08:51:25,310 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-08T08:51:25,312 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-11-08T08:51:25,313 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-08T08:51:25,313 DEBUG [RS:1;6ac96558fc7f:39649 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2b915cfe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:25,316 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-11-08T08:51:25,316 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-11-08T08:51:25,317 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-08T08:51:25,317 DEBUG [RS:0;6ac96558fc7f:46683 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@44aa6803, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:25,324 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;6ac96558fc7f:39649 2024-11-08T08:51:25,325 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-08T08:51:25,325 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-08T08:51:25,325 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-08T08:51:25,329 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(2659): reportForDuty to master=6ac96558fc7f,46451,1731055884597 with port=39649, startcode=1731055884817 2024-11-08T08:51:25,329 DEBUG [RS:1;6ac96558fc7f:39649 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-08T08:51:25,329 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-11-08T08:51:25,330 DEBUG [RS:2;6ac96558fc7f:42453 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7291cd39, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=6ac96558fc7f/172.17.0.3:0 2024-11-08T08:51:25,330 DEBUG [RS:0;6ac96558fc7f:46683 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;6ac96558fc7f:46683 2024-11-08T08:51:25,331 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-08T08:51:25,331 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-08T08:51:25,331 DEBUG [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-08T08:51:25,331 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(2659): reportForDuty to master=6ac96558fc7f,46451,1731055884597 with port=46683, startcode=1731055884773 2024-11-08T08:51:25,332 DEBUG [RS:0;6ac96558fc7f:46683 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-08T08:51:25,332 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54463, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-11-08T08:51:25,332 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,332 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] master.ServerManager(517): Registering regionserver=6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,333 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45041, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-11-08T08:51:25,334 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:25,334 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] master.ServerManager(517): Registering regionserver=6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:25,334 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90 2024-11-08T08:51:25,334 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41273 2024-11-08T08:51:25,334 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-08T08:51:25,336 DEBUG [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90 2024-11-08T08:51:25,336 DEBUG [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41273 2024-11-08T08:51:25,336 DEBUG [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-08T08:51:25,343 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;6ac96558fc7f:42453 2024-11-08T08:51:25,343 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-11-08T08:51:25,343 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-11-08T08:51:25,343 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(832): About to register with Master. 2024-11-08T08:51:25,344 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-08T08:51:25,382 DEBUG [RS:1;6ac96558fc7f:39649 {}] zookeeper.ZKUtil(111): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,382 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(2659): reportForDuty to master=6ac96558fc7f,46451,1731055884597 with port=42453, startcode=1731055884861 2024-11-08T08:51:25,382 WARN [RS:1;6ac96558fc7f:39649 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:25,382 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [6ac96558fc7f,46683,1731055884773] 2024-11-08T08:51:25,382 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [6ac96558fc7f,39649,1731055884817] 2024-11-08T08:51:25,382 INFO [RS:1;6ac96558fc7f:39649 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-08T08:51:25,383 DEBUG [RS:2;6ac96558fc7f:42453 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-11-08T08:51:25,383 DEBUG [RS:0;6ac96558fc7f:46683 {}] zookeeper.ZKUtil(111): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:25,383 WARN [RS:0;6ac96558fc7f:46683 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:25,383 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,383 INFO [RS:0;6ac96558fc7f:46683 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-08T08:51:25,383 DEBUG [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:25,386 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:44339, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-11-08T08:51:25,387 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:25,387 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] master.ServerManager(517): Registering regionserver=6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:25,390 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-08T08:51:25,390 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-08T08:51:25,390 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90 2024-11-08T08:51:25,390 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:41273 2024-11-08T08:51:25,390 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-11-08T08:51:25,393 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-08T08:51:25,393 INFO [RS:0;6ac96558fc7f:46683 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-08T08:51:25,393 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,394 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-08T08:51:25,395 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-08T08:51:25,395 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,395 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,395 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,395 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,395 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,395 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-08T08:51:25,395 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,395 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,396 INFO [RS:1;6ac96558fc7f:39649 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,396 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:25,396 DEBUG [RS:0;6ac96558fc7f:46683 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:25,398 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-08T08:51:25,399 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-08T08:51:25,399 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:25,399 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:25,400 DEBUG [RS:1;6ac96558fc7f:39649 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:25,403 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46683,1731055884773-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,404 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,39649,1731055884817-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:25,413 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-08T08:51:25,413 DEBUG [RS:2;6ac96558fc7f:42453 {}] zookeeper.ZKUtil(111): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:25,414 WARN [RS:2;6ac96558fc7f:42453 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-11-08T08:51:25,414 INFO [RS:2;6ac96558fc7f:42453 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-08T08:51:25,414 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:25,414 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [6ac96558fc7f,42453,1731055884861] 2024-11-08T08:51:25,418 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-11-08T08:51:25,421 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-08T08:51:25,421 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46683,1731055884773-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,422 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,422 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.Replication(171): 6ac96558fc7f,46683,1731055884773 started 2024-11-08T08:51:25,422 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-11-08T08:51:25,423 INFO [RS:2;6ac96558fc7f:42453 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-11-08T08:51:25,423 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,423 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-11-08T08:51:25,424 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-11-08T08:51:25,424 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,424 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,424 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/6ac96558fc7f:0, corePoolSize=2, maxPoolSize=2 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,39649,1731055884817-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/6ac96558fc7f:0, corePoolSize=1, maxPoolSize=1 2024-11-08T08:51:25,425 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:25,425 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,426 DEBUG [RS:2;6ac96558fc7f:42453 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0, corePoolSize=3, maxPoolSize=3 2024-11-08T08:51:25,426 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.Replication(171): 6ac96558fc7f,39649,1731055884817 started 2024-11-08T08:51:25,426 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,426 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,426 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,426 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,426 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,426 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42453,1731055884861-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:25,441 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-11-08T08:51:25,442 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,42453,1731055884861-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,442 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,442 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.Replication(171): 6ac96558fc7f,42453,1731055884861 started 2024-11-08T08:51:25,443 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,443 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(1482): Serving as 6ac96558fc7f,46683,1731055884773, RpcServer on 6ac96558fc7f/172.17.0.3:46683, sessionid=0x10119a494c90001 2024-11-08T08:51:25,443 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-08T08:51:25,443 DEBUG [RS:0;6ac96558fc7f:46683 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:25,443 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,46683,1731055884773' 2024-11-08T08:51:25,444 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-08T08:51:25,444 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-08T08:51:25,445 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-08T08:51:25,445 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-08T08:51:25,445 DEBUG [RS:0;6ac96558fc7f:46683 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:25,445 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,46683,1731055884773' 2024-11-08T08:51:25,445 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-08T08:51:25,445 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,445 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1482): Serving as 6ac96558fc7f,39649,1731055884817, RpcServer on 6ac96558fc7f/172.17.0.3:39649, sessionid=0x10119a494c90002 2024-11-08T08:51:25,445 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-08T08:51:25,445 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-08T08:51:25,445 DEBUG [RS:1;6ac96558fc7f:39649 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,445 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,39649,1731055884817' 2024-11-08T08:51:25,445 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-08T08:51:25,446 DEBUG [RS:0;6ac96558fc7f:46683 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-08T08:51:25,446 INFO [RS:0;6ac96558fc7f:46683 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-08T08:51:25,446 INFO [RS:0;6ac96558fc7f:46683 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-08T08:51:25,446 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-08T08:51:25,446 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-08T08:51:25,446 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-08T08:51:25,446 DEBUG [RS:1;6ac96558fc7f:39649 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,446 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,39649,1731055884817' 2024-11-08T08:51:25,446 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-08T08:51:25,447 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-08T08:51:25,447 DEBUG [RS:1;6ac96558fc7f:39649 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-08T08:51:25,447 INFO [RS:1;6ac96558fc7f:39649 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-08T08:51:25,447 INFO [RS:1;6ac96558fc7f:39649 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-08T08:51:25,456 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:25,456 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1482): Serving as 6ac96558fc7f,42453,1731055884861, RpcServer on 6ac96558fc7f/172.17.0.3:42453, sessionid=0x10119a494c90003 2024-11-08T08:51:25,456 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-11-08T08:51:25,456 DEBUG [RS:2;6ac96558fc7f:42453 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:25,456 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,42453,1731055884861' 2024-11-08T08:51:25,456 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-11-08T08:51:25,457 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-11-08T08:51:25,458 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-11-08T08:51:25,458 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-11-08T08:51:25,458 DEBUG [RS:2;6ac96558fc7f:42453 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:25,458 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '6ac96558fc7f,42453,1731055884861' 2024-11-08T08:51:25,458 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-11-08T08:51:25,459 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-11-08T08:51:25,459 DEBUG [RS:2;6ac96558fc7f:42453 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-11-08T08:51:25,459 INFO [RS:2;6ac96558fc7f:42453 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-11-08T08:51:25,459 INFO [RS:2;6ac96558fc7f:42453 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-11-08T08:51:25,462 WARN [6ac96558fc7f:46451 {}] assignment.AssignmentManager(2443): No servers available; cannot place 1 unassigned regions. 2024-11-08T08:51:25,548 INFO [RS:0;6ac96558fc7f:46683 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C46683%2C1731055884773, suffix=, logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,46683,1731055884773, archiveDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs, maxLogs=32 2024-11-08T08:51:25,549 INFO [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C39649%2C1731055884817, suffix=, logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,39649,1731055884817, archiveDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs, maxLogs=32 2024-11-08T08:51:25,551 INFO [RS:1;6ac96558fc7f:39649 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 6ac96558fc7f%2C39649%2C1731055884817.1731055885551 2024-11-08T08:51:25,551 INFO [RS:0;6ac96558fc7f:46683 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 6ac96558fc7f%2C46683%2C1731055884773.1731055885551 2024-11-08T08:51:25,562 INFO [RS:2;6ac96558fc7f:42453 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C42453%2C1731055884861, suffix=, logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,42453,1731055884861, archiveDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs, maxLogs=32 2024-11-08T08:51:25,563 INFO [RS:0;6ac96558fc7f:46683 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,46683,1731055884773/6ac96558fc7f%2C46683%2C1731055884773.1731055885551 2024-11-08T08:51:25,563 INFO [RS:2;6ac96558fc7f:42453 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 6ac96558fc7f%2C42453%2C1731055884861.1731055885563 2024-11-08T08:51:25,563 INFO [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,39649,1731055884817/6ac96558fc7f%2C39649%2C1731055884817.1731055885551 2024-11-08T08:51:25,565 DEBUG [RS:0;6ac96558fc7f:46683 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43555:43555),(127.0.0.1/127.0.0.1:35871:35871),(127.0.0.1/127.0.0.1:41195:41195)] 2024-11-08T08:51:25,565 DEBUG [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43555:43555),(127.0.0.1/127.0.0.1:35871:35871),(127.0.0.1/127.0.0.1:41195:41195)] 2024-11-08T08:51:25,576 INFO [RS:2;6ac96558fc7f:42453 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,42453,1731055884861/6ac96558fc7f%2C42453%2C1731055884861.1731055885563 2024-11-08T08:51:25,577 DEBUG [RS:2;6ac96558fc7f:42453 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41195:41195),(127.0.0.1/127.0.0.1:43555:43555),(127.0.0.1/127.0.0.1:35871:35871)] 2024-11-08T08:51:25,713 DEBUG [6ac96558fc7f:46451 {}] assignment.AssignmentManager(2464): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-11-08T08:51:25,713 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(204): Hosts are {6ac96558fc7f=0} racks are {/default-rack=0} 2024-11-08T08:51:25,719 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-08T08:51:25,719 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-08T08:51:25,719 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-08T08:51:25,719 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-08T08:51:25,719 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-08T08:51:25,719 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-08T08:51:25,719 INFO [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-08T08:51:25,720 INFO [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-08T08:51:25,720 INFO [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-08T08:51:25,720 DEBUG [6ac96558fc7f:46451 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-08T08:51:25,721 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:25,724 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 6ac96558fc7f,39649,1731055884817, state=OPENING 2024-11-08T08:51:25,786 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-11-08T08:51:25,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,797 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:25,797 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:25,797 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:25,797 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-11-08T08:51:25,798 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=6ac96558fc7f,39649,1731055884817}] 2024-11-08T08:51:25,798 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:25,798 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:25,954 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-08T08:51:25,955 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56515, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-08T08:51:25,960 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-11-08T08:51:25,960 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-11-08T08:51:25,963 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=6ac96558fc7f%2C39649%2C1731055884817.meta, suffix=.meta, logDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,39649,1731055884817, archiveDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs, maxLogs=32 2024-11-08T08:51:25,964 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 6ac96558fc7f%2C39649%2C1731055884817.meta.1731055885963.meta 2024-11-08T08:51:25,971 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/WALs/6ac96558fc7f,39649,1731055884817/6ac96558fc7f%2C39649%2C1731055884817.meta.1731055885963.meta 2024-11-08T08:51:25,972 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35871:35871),(127.0.0.1/127.0.0.1:41195:41195),(127.0.0.1/127.0.0.1:43555:43555)] 2024-11-08T08:51:25,975 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-11-08T08:51:25,975 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-11-08T08:51:25,975 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-11-08T08:51:25,975 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-11-08T08:51:25,976 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-11-08T08:51:25,976 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:25,976 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-11-08T08:51:25,976 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-11-08T08:51:25,978 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-11-08T08:51:25,979 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-11-08T08:51:25,979 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,980 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,980 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-11-08T08:51:25,981 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-11-08T08:51:25,981 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,982 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,982 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-11-08T08:51:25,983 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-11-08T08:51:25,983 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,984 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,984 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-11-08T08:51:25,985 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-11-08T08:51:25,985 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:25,985 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-11-08T08:51:25,985 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-11-08T08:51:25,986 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740 2024-11-08T08:51:25,988 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740 2024-11-08T08:51:25,990 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-11-08T08:51:25,990 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-11-08T08:51:25,991 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-11-08T08:51:25,993 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-11-08T08:51:25,994 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69429877, jitterRate=0.03458578884601593}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-11-08T08:51:25,994 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-11-08T08:51:25,996 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1731055885976Writing region info on filesystem at 1731055885976Initializing all the Stores at 1731055885977 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885977Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885978 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055885978Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1731055885978Cleaning up temporary data from old regions at 1731055885990 (+12 ms)Running coprocessor post-open hooks at 1731055885994 (+4 ms)Region opened successfully at 1731055885995 (+1 ms) 2024-11-08T08:51:25,998 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1731055885953 2024-11-08T08:51:26,001 DEBUG [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-11-08T08:51:26,002 INFO [RS_OPEN_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-11-08T08:51:26,002 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:26,004 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 6ac96558fc7f,39649,1731055884817, state=OPEN 2024-11-08T08:51:26,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:26,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:26,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:26,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-11-08T08:51:26,034 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:26,034 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:26,034 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:26,034 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:26,034 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-11-08T08:51:26,039 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-11-08T08:51:26,039 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=6ac96558fc7f,39649,1731055884817 in 237 msec 2024-11-08T08:51:26,043 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-11-08T08:51:26,043 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 731 msec 2024-11-08T08:51:26,044 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-11-08T08:51:26,044 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-11-08T08:51:26,046 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-08T08:51:26,046 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=6ac96558fc7f,39649,1731055884817, seqNum=-1] 2024-11-08T08:51:26,047 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-08T08:51:26,048 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56803, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-08T08:51:26,057 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 802 msec 2024-11-08T08:51:26,058 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1731055886058, completionTime=-1 2024-11-08T08:51:26,058 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-11-08T08:51:26,058 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] assignment.AssignmentManager(1756): Joining cluster... 2024-11-08T08:51:26,060 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] assignment.AssignmentManager(1768): Number of RegionServers=3 2024-11-08T08:51:26,060 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1731055946060 2024-11-08T08:51:26,060 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1731056006060 2024-11-08T08:51:26,060 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] assignment.AssignmentManager(1775): Joined the cluster in 2 msec 2024-11-08T08:51:26,061 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46451,1731055884597-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,061 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46451,1731055884597-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,061 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46451,1731055884597-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,061 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-6ac96558fc7f:46451, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,061 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,061 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,064 DEBUG [master/6ac96558fc7f:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 1.138sec 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46451,1731055884597-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-11-08T08:51:26,067 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46451,1731055884597-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-11-08T08:51:26,071 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-11-08T08:51:26,071 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-11-08T08:51:26,071 INFO [master/6ac96558fc7f:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=6ac96558fc7f,46451,1731055884597-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-11-08T08:51:26,089 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@324feec6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-08T08:51:26,090 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 6ac96558fc7f,46451,-1 for getting cluster id 2024-11-08T08:51:26,090 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-11-08T08:51:26,091 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = 'bd746c6f-4269-47fe-8a62-117605dafd0f' 2024-11-08T08:51:26,092 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-11-08T08:51:26,092 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "bd746c6f-4269-47fe-8a62-117605dafd0f" 2024-11-08T08:51:26,092 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@68b47d9b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-08T08:51:26,092 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [6ac96558fc7f,46451,-1] 2024-11-08T08:51:26,093 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-11-08T08:51:26,093 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:26,095 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:57666, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-11-08T08:51:26,096 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@12dcf010, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-11-08T08:51:26,096 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-11-08T08:51:26,097 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=6ac96558fc7f,39649,1731055884817, seqNum=-1] 2024-11-08T08:51:26,098 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-08T08:51:26,100 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43950, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-08T08:51:26,102 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:26,103 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-11-08T08:51:26,104 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is 6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:26,105 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@7e6dad3b 2024-11-08T08:51:26,105 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-11-08T08:51:26,107 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:57682, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-11-08T08:51:26,108 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.3 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-11-08T08:51:26,110 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-11-08T08:51:26,112 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-11-08T08:51:26,112 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:26,113 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-11-08T08:51:26,114 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:26,115 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-11-08T08:51:26,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741837_1013 (size=392) 2024-11-08T08:51:26,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741837_1013 (size=392) 2024-11-08T08:51:26,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741837_1013 (size=392) 2024-11-08T08:51:26,128 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 87db7e43283c3f608ba48288579bba7d, NAME => 'TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90 2024-11-08T08:51:26,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741838_1014 (size=51) 2024-11-08T08:51:26,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741838_1014 (size=51) 2024-11-08T08:51:26,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741838_1014 (size=51) 2024-11-08T08:51:26,139 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:26,140 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 87db7e43283c3f608ba48288579bba7d, disabling compactions & flushes 2024-11-08T08:51:26,140 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,140 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,140 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. after waiting 0 ms 2024-11-08T08:51:26,140 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,140 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,140 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 87db7e43283c3f608ba48288579bba7d: Waiting for close lock at 1731055886140Disabling compacts and flushes for region at 1731055886140Disabling writes for close at 1731055886140Writing region close event to WAL at 1731055886140Closed at 1731055886140 2024-11-08T08:51:26,143 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-11-08T08:51:26,143 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1731055886143"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1731055886143"}]},"ts":"1731055886143"} 2024-11-08T08:51:26,147 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-11-08T08:51:26,149 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-11-08T08:51:26,150 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731055886149"}]},"ts":"1731055886149"} 2024-11-08T08:51:26,153 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-11-08T08:51:26,153 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {6ac96558fc7f=0} racks are {/default-rack=0} 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-11-08T08:51:26,155 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-11-08T08:51:26,155 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-11-08T08:51:26,155 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-11-08T08:51:26,155 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-11-08T08:51:26,156 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=87db7e43283c3f608ba48288579bba7d, ASSIGN}] 2024-11-08T08:51:26,158 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=87db7e43283c3f608ba48288579bba7d, ASSIGN 2024-11-08T08:51:26,160 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=87db7e43283c3f608ba48288579bba7d, ASSIGN; state=OFFLINE, location=6ac96558fc7f,42453,1731055884861; forceNewPlan=false, retain=false 2024-11-08T08:51:26,225 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:26,311 INFO [6ac96558fc7f:46451 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-11-08T08:51:26,311 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=87db7e43283c3f608ba48288579bba7d, regionState=OPENING, regionLocation=6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:26,315 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=87db7e43283c3f608ba48288579bba7d, ASSIGN because future has completed 2024-11-08T08:51:26,316 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 87db7e43283c3f608ba48288579bba7d, server=6ac96558fc7f,42453,1731055884861}] 2024-11-08T08:51:26,434 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:26,470 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-11-08T08:51:26,472 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:36719, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-11-08T08:51:26,477 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,477 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 87db7e43283c3f608ba48288579bba7d, NAME => 'TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d.', STARTKEY => '', ENDKEY => ''} 2024-11-08T08:51:26,477 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,478 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-11-08T08:51:26,478 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,478 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,480 INFO [StoreOpener-87db7e43283c3f608ba48288579bba7d-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,481 INFO [StoreOpener-87db7e43283c3f608ba48288579bba7d-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 87db7e43283c3f608ba48288579bba7d columnFamilyName cf 2024-11-08T08:51:26,481 DEBUG [StoreOpener-87db7e43283c3f608ba48288579bba7d-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-11-08T08:51:26,482 INFO [StoreOpener-87db7e43283c3f608ba48288579bba7d-1 {}] regionserver.HStore(327): Store=87db7e43283c3f608ba48288579bba7d/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-11-08T08:51:26,482 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,483 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,483 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,484 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,484 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,486 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,488 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-11-08T08:51:26,489 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 87db7e43283c3f608ba48288579bba7d; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75234085, jitterRate=0.12107522785663605}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-11-08T08:51:26,489 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:26,490 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 87db7e43283c3f608ba48288579bba7d: Running coprocessor pre-open hook at 1731055886478Writing region info on filesystem at 1731055886478Initializing all the Stores at 1731055886479 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1731055886479Cleaning up temporary data from old regions at 1731055886484 (+5 ms)Running coprocessor post-open hooks at 1731055886489 (+5 ms)Region opened successfully at 1731055886490 (+1 ms) 2024-11-08T08:51:26,491 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d., pid=6, masterSystemTime=1731055886470 2024-11-08T08:51:26,495 DEBUG [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,495 INFO [RS_OPEN_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,497 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=87db7e43283c3f608ba48288579bba7d, regionState=OPEN, openSeqNum=2, regionLocation=6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:26,500 WARN [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=46451 {}] assignment.AssignmentManager(1535): Unable to acquire lock for regionNode state=OPEN, location=6ac96558fc7f,42453,1731055884861, table=TestHBaseWalOnEC, region=87db7e43283c3f608ba48288579bba7d. It is likely that another thread is currently holding the lock. To avoid deadlock, skip execution for now. 2024-11-08T08:51:26,500 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 87db7e43283c3f608ba48288579bba7d, server=6ac96558fc7f,42453,1731055884861 because future has completed 2024-11-08T08:51:26,506 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-11-08T08:51:26,506 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 87db7e43283c3f608ba48288579bba7d, server=6ac96558fc7f,42453,1731055884861 in 187 msec 2024-11-08T08:51:26,510 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-11-08T08:51:26,511 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=87db7e43283c3f608ba48288579bba7d, ASSIGN in 350 msec 2024-11-08T08:51:26,512 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-11-08T08:51:26,512 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1731055886512"}]},"ts":"1731055886512"} 2024-11-08T08:51:26,515 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-11-08T08:51:26,516 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-11-08T08:51:26,519 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 408 msec 2024-11-08T08:51:26,744 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-11-08T08:51:26,745 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-11-08T08:51:26,745 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-11-08T08:51:26,745 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-08T08:51:26,749 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-11-08T08:51:26,750 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-11-08T08:51:26,750 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-11-08T08:51:26,754 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d., hostname=6ac96558fc7f,42453,1731055884861, seqNum=2] 2024-11-08T08:51:26,754 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-11-08T08:51:26,756 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45838, version=3.0.0-beta-2-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-11-08T08:51:26,760 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.3 flush TestHBaseWalOnEC 2024-11-08T08:51:26,762 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-11-08T08:51:26,763 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:26,764 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-11-08T08:51:26,765 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-11-08T08:51:26,766 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-11-08T08:51:26,875 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:26,921 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=42453 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-11-08T08:51:26,921 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,921 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 87db7e43283c3f608ba48288579bba7d 1/1 column families, dataSize=32 B heapSize=360 B 2024-11-08T08:51:26,937 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/.tmp/cf/a7ec954ac5ce4f5daec99f4436ddebb7 is 36, key is row/cf:cq/1731055886757/Put/seqid=0 2024-11-08T08:51:26,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741839_1015 (size=4787) 2024-11-08T08:51:26,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741839_1015 (size=4787) 2024-11-08T08:51:26,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741839_1015 (size=4787) 2024-11-08T08:51:26,948 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/.tmp/cf/a7ec954ac5ce4f5daec99f4436ddebb7 2024-11-08T08:51:26,957 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/.tmp/cf/a7ec954ac5ce4f5daec99f4436ddebb7 as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/cf/a7ec954ac5ce4f5daec99f4436ddebb7 2024-11-08T08:51:26,965 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/cf/a7ec954ac5ce4f5daec99f4436ddebb7, entries=1, sequenceid=5, filesize=4.7 K 2024-11-08T08:51:26,967 INFO [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 87db7e43283c3f608ba48288579bba7d in 46ms, sequenceid=5, compaction requested=false 2024-11-08T08:51:26,967 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 87db7e43283c3f608ba48288579bba7d: 2024-11-08T08:51:26,967 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:26,967 DEBUG [RS_FLUSH_OPERATIONS-regionserver/6ac96558fc7f:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-11-08T08:51:26,968 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-11-08T08:51:26,973 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-11-08T08:51:26,973 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 204 msec 2024-11-08T08:51:26,977 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 215 msec 2024-11-08T08:51:27,085 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=46451 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-11-08T08:51:27,085 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-11-08T08:51:27,090 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-11-08T08:51:27,090 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-08T08:51:27,090 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:27,090 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,090 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,090 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-11-08T08:51:27,090 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-11-08T08:51:27,090 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=211998716, stopped=false 2024-11-08T08:51:27,090 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=6ac96558fc7f,46451,1731055884597 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-11-08T08:51:27,197 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:27,197 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:27,197 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-11-08T08:51:27,198 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:27,198 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:27,198 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:27,198 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:27,198 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,198 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '6ac96558fc7f,46683,1731055884773' ***** 2024-11-08T08:51:27,198 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-08T08:51:27,198 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '6ac96558fc7f,39649,1731055884817' ***** 2024-11-08T08:51:27,198 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-08T08:51:27,198 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '6ac96558fc7f,42453,1731055884861' ***** 2024-11-08T08:51:27,198 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-11-08T08:51:27,198 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-08T08:51:27,198 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-11-08T08:51:27,199 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-08T08:51:27,199 INFO [RS:0;6ac96558fc7f:46683 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-08T08:51:27,199 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-08T08:51:27,199 INFO [RS:0;6ac96558fc7f:46683 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-08T08:51:27,199 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HeapMemoryManager(220): Stopping 2024-11-08T08:51:27,199 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-08T08:51:27,199 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-11-08T08:51:27,199 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(959): stopping server 6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:27,200 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:27,200 INFO [RS:2;6ac96558fc7f:42453 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-08T08:51:27,200 INFO [RS:0;6ac96558fc7f:46683 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;6ac96558fc7f:46683. 2024-11-08T08:51:27,200 INFO [RS:2;6ac96558fc7f:42453 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-08T08:51:27,200 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(3091): Received CLOSE for 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:27,200 DEBUG [RS:0;6ac96558fc7f:46683 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:27,200 DEBUG [RS:0;6ac96558fc7f:46683 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-11-08T08:51:27,200 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(976): stopping server 6ac96558fc7f,46683,1731055884773; all regions closed. 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(959): stopping server 6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;6ac96558fc7f:39649. 2024-11-08T08:51:27,200 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(959): stopping server 6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:27,200 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:27,200 DEBUG [RS:1;6ac96558fc7f:39649 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:27,200 DEBUG [RS:1;6ac96558fc7f:39649 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,200 INFO [RS:2;6ac96558fc7f:42453 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;6ac96558fc7f:42453. 2024-11-08T08:51:27,200 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 87db7e43283c3f608ba48288579bba7d, disabling compactions & flushes 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-08T08:51:27,200 DEBUG [RS:2;6ac96558fc7f:42453 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-11-08T08:51:27,200 INFO [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:27,200 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-08T08:51:27,201 DEBUG [RS:2;6ac96558fc7f:42453 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,201 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:27,201 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-08T08:51:27,201 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. after waiting 0 ms 2024-11-08T08:51:27,201 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-08T08:51:27,201 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-11-08T08:51:27,201 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:27,201 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1325): Online Regions={87db7e43283c3f608ba48288579bba7d=TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d.} 2024-11-08T08:51:27,201 DEBUG [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1351): Waiting on 87db7e43283c3f608ba48288579bba7d 2024-11-08T08:51:27,201 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-11-08T08:51:27,201 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-11-08T08:51:27,201 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,201 DEBUG [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-11-08T08:51:27,201 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-11-08T08:51:27,201 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,201 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-11-08T08:51:27,201 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,201 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-11-08T08:51:27,201 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,201 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-11-08T08:51:27,201 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,202 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-11-08T08:51:27,203 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-11-08T08:51:27,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741834_1010 (size=93) 2024-11-08T08:51:27,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741834_1010 (size=93) 2024-11-08T08:51:27,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741834_1010 (size=93) 2024-11-08T08:51:27,205 INFO [regionserver/6ac96558fc7f:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:27,206 INFO [regionserver/6ac96558fc7f:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:27,207 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/default/TestHBaseWalOnEC/87db7e43283c3f608ba48288579bba7d/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-11-08T08:51:27,208 DEBUG [RS:0;6ac96558fc7f:46683 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs 2024-11-08T08:51:27,208 INFO [RS:0;6ac96558fc7f:46683 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 6ac96558fc7f%2C46683%2C1731055884773:(num 1731055885551) 2024-11-08T08:51:27,208 DEBUG [RS:0;6ac96558fc7f:46683 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,208 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:27,208 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:27,208 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.ChoreService(370): Chore service for: regionserver/6ac96558fc7f:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:27,209 INFO [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:27,209 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-08T08:51:27,209 INFO [regionserver/6ac96558fc7f:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:27,209 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-08T08:51:27,209 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 87db7e43283c3f608ba48288579bba7d: Waiting for close lock at 1731055887200Running coprocessor pre-close hooks at 1731055887200Disabling compacts and flushes for region at 1731055887200Disabling writes for close at 1731055887201 (+1 ms)Writing region close event to WAL at 1731055887202 (+1 ms)Running coprocessor post-close hooks at 1731055887208 (+6 ms)Closed at 1731055887208 2024-11-08T08:51:27,209 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-08T08:51:27,209 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:27,209 DEBUG [RS_CLOSE_REGION-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d. 2024-11-08T08:51:27,209 INFO [RS:0;6ac96558fc7f:46683 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46683 2024-11-08T08:51:27,217 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/6ac96558fc7f,46683,1731055884773 2024-11-08T08:51:27,217 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-08T08:51:27,218 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:27,222 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/info/d9e3acebb4604a0a86cde4c29d46162d is 153, key is TestHBaseWalOnEC,,1731055886107.87db7e43283c3f608ba48288579bba7d./info:regioninfo/1731055886496/Put/seqid=0 2024-11-08T08:51:27,223 WARN [IPC Server handler 3 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:27,223 WARN [IPC Server handler 3 on default port 41273 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:27,223 WARN [IPC Server handler 3 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:27,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741840_1016 (size=6637) 2024-11-08T08:51:27,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741840_1016 (size=6637) 2024-11-08T08:51:27,228 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [6ac96558fc7f,46683,1731055884773] 2024-11-08T08:51:27,228 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/info/d9e3acebb4604a0a86cde4c29d46162d 2024-11-08T08:51:27,229 INFO [regionserver/6ac96558fc7f:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:27,238 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/6ac96558fc7f,46683,1731055884773 already deleted, retry=false 2024-11-08T08:51:27,239 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 6ac96558fc7f,46683,1731055884773 expired; onlineServers=2 2024-11-08T08:51:27,251 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/ns/5ee6f93b0bd846b09ee4714e04afa8f0 is 43, key is default/ns:d/1731055886049/Put/seqid=0 2024-11-08T08:51:27,252 WARN [IPC Server handler 1 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:27,252 WARN [IPC Server handler 1 on default port 41273 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:27,252 WARN [IPC Server handler 1 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:27,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741841_1017 (size=5153) 2024-11-08T08:51:27,257 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741841_1017 (size=5153) 2024-11-08T08:51:27,257 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/ns/5ee6f93b0bd846b09ee4714e04afa8f0 2024-11-08T08:51:27,278 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/table/78c2e0a481604bbf965f35e58e39c0de is 52, key is TestHBaseWalOnEC/table:state/1731055886512/Put/seqid=0 2024-11-08T08:51:27,280 WARN [IPC Server handler 4 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:27,280 WARN [IPC Server handler 4 on default port 41273 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:27,280 WARN [IPC Server handler 4 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:27,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741842_1018 (size=5249) 2024-11-08T08:51:27,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741842_1018 (size=5249) 2024-11-08T08:51:27,285 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/table/78c2e0a481604bbf965f35e58e39c0de 2024-11-08T08:51:27,292 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/info/d9e3acebb4604a0a86cde4c29d46162d as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/info/d9e3acebb4604a0a86cde4c29d46162d 2024-11-08T08:51:27,301 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/info/d9e3acebb4604a0a86cde4c29d46162d, entries=10, sequenceid=11, filesize=6.5 K 2024-11-08T08:51:27,302 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/ns/5ee6f93b0bd846b09ee4714e04afa8f0 as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/ns/5ee6f93b0bd846b09ee4714e04afa8f0 2024-11-08T08:51:27,310 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/ns/5ee6f93b0bd846b09ee4714e04afa8f0, entries=2, sequenceid=11, filesize=5.0 K 2024-11-08T08:51:27,312 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/.tmp/table/78c2e0a481604bbf965f35e58e39c0de as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/table/78c2e0a481604bbf965f35e58e39c0de 2024-11-08T08:51:27,320 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/table/78c2e0a481604bbf965f35e58e39c0de, entries=2, sequenceid=11, filesize=5.1 K 2024-11-08T08:51:27,321 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 119ms, sequenceid=11, compaction requested=false 2024-11-08T08:51:27,327 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-11-08T08:51:27,327 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-11-08T08:51:27,327 INFO [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-11-08T08:51:27,328 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1731055887201Running coprocessor pre-close hooks at 1731055887201Disabling compacts and flushes for region at 1731055887201Disabling writes for close at 1731055887201Obtaining lock to block concurrent updates at 1731055887203 (+2 ms)Preparing flush snapshotting stores in 1588230740 at 1731055887203Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1731055887203Flushing stores of hbase:meta,,1.1588230740 at 1731055887204 (+1 ms)Flushing 1588230740/info: creating writer at 1731055887204Flushing 1588230740/info: appending metadata at 1731055887221 (+17 ms)Flushing 1588230740/info: closing flushed file at 1731055887221Flushing 1588230740/ns: creating writer at 1731055887235 (+14 ms)Flushing 1588230740/ns: appending metadata at 1731055887250 (+15 ms)Flushing 1588230740/ns: closing flushed file at 1731055887250Flushing 1588230740/table: creating writer at 1731055887264 (+14 ms)Flushing 1588230740/table: appending metadata at 1731055887278 (+14 ms)Flushing 1588230740/table: closing flushed file at 1731055887278Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3229b284: reopening flushed file at 1731055887291 (+13 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@8e5ddfd: reopening flushed file at 1731055887301 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@57ceb447: reopening flushed file at 1731055887310 (+9 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 119ms, sequenceid=11, compaction requested=false at 1731055887321 (+11 ms)Writing region close event to WAL at 1731055887322 (+1 ms)Running coprocessor post-close hooks at 1731055887327 (+5 ms)Closed at 1731055887327 2024-11-08T08:51:27,328 DEBUG [RS_CLOSE_META-regionserver/6ac96558fc7f:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-11-08T08:51:27,329 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,329 INFO [RS:0;6ac96558fc7f:46683 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:27,329 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:46683-0x10119a494c90001, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,329 INFO [RS:0;6ac96558fc7f:46683 {}] regionserver.HRegionServer(1031): Exiting; stopping=6ac96558fc7f,46683,1731055884773; zookeeper connection closed. 2024-11-08T08:51:27,329 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4ecf2a25 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4ecf2a25 2024-11-08T08:51:27,401 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(976): stopping server 6ac96558fc7f,42453,1731055884861; all regions closed. 2024-11-08T08:51:27,401 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(976): stopping server 6ac96558fc7f,39649,1731055884817; all regions closed. 2024-11-08T08:51:27,402 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,402 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,404 INFO [regionserver/6ac96558fc7f:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-11-08T08:51:27,404 INFO [regionserver/6ac96558fc7f:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-11-08T08:51:27,404 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741835_1011 (size=1298) 2024-11-08T08:51:27,405 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741836_1012 (size=2751) 2024-11-08T08:51:27,405 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741836_1012 (size=2751) 2024-11-08T08:51:27,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741836_1012 (size=2751) 2024-11-08T08:51:27,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741835_1011 (size=1298) 2024-11-08T08:51:27,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741835_1011 (size=1298) 2024-11-08T08:51:27,408 DEBUG [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs 2024-11-08T08:51:27,408 INFO [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 6ac96558fc7f%2C39649%2C1731055884817.meta:.meta(num 1731055885963) 2024-11-08T08:51:27,408 DEBUG [RS:2;6ac96558fc7f:42453 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs 2024-11-08T08:51:27,409 INFO [RS:2;6ac96558fc7f:42453 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 6ac96558fc7f%2C42453%2C1731055884861:(num 1731055885563) 2024-11-08T08:51:27,409 DEBUG [RS:2;6ac96558fc7f:42453 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,409 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:27,409 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:27,409 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,409 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,409 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,409 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.ChoreService(370): Chore service for: regionserver/6ac96558fc7f:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:27,409 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,409 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-11-08T08:51:27,409 INFO [regionserver/6ac96558fc7f:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:27,410 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-11-08T08:51:27,410 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-11-08T08:51:27,410 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,410 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:27,410 INFO [RS:2;6ac96558fc7f:42453 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:42453 2024-11-08T08:51:27,412 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741833_1009 (size=93) 2024-11-08T08:51:27,413 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741833_1009 (size=93) 2024-11-08T08:51:27,413 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741833_1009 (size=93) 2024-11-08T08:51:27,415 DEBUG [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/oldWALs 2024-11-08T08:51:27,415 INFO [RS:1;6ac96558fc7f:39649 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 6ac96558fc7f%2C39649%2C1731055884817:(num 1731055885551) 2024-11-08T08:51:27,415 DEBUG [RS:1;6ac96558fc7f:39649 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-11-08T08:51:27,415 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.LeaseManager(133): Closed leases 2024-11-08T08:51:27,415 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:27,415 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.ChoreService(370): Chore service for: regionserver/6ac96558fc7f:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:27,415 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:27,415 INFO [regionserver/6ac96558fc7f:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:27,416 INFO [RS:1;6ac96558fc7f:39649 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:39649 2024-11-08T08:51:27,423 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/6ac96558fc7f,42453,1731055884861 2024-11-08T08:51:27,423 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-11-08T08:51:27,423 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:27,434 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/6ac96558fc7f,39649,1731055884817 2024-11-08T08:51:27,434 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:27,445 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [6ac96558fc7f,42453,1731055884861] 2024-11-08T08:51:27,466 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/6ac96558fc7f,42453,1731055884861 already deleted, retry=false 2024-11-08T08:51:27,466 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 6ac96558fc7f,42453,1731055884861 expired; onlineServers=1 2024-11-08T08:51:27,466 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [6ac96558fc7f,39649,1731055884817] 2024-11-08T08:51:27,476 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/6ac96558fc7f,39649,1731055884817 already deleted, retry=false 2024-11-08T08:51:27,476 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 6ac96558fc7f,39649,1731055884817 expired; onlineServers=0 2024-11-08T08:51:27,477 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '6ac96558fc7f,46451,1731055884597' ***** 2024-11-08T08:51:27,477 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-11-08T08:51:27,477 INFO [M:0;6ac96558fc7f:46451 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-11-08T08:51:27,477 INFO [M:0;6ac96558fc7f:46451 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-11-08T08:51:27,478 DEBUG [M:0;6ac96558fc7f:46451 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-11-08T08:51:27,478 DEBUG [M:0;6ac96558fc7f:46451 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-11-08T08:51:27,478 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.large.0-1731055885260 {}] cleaner.HFileCleaner(306): Exit Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.large.0-1731055885260,5,FailOnTimeoutGroup] 2024-11-08T08:51:27,478 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-11-08T08:51:27,478 DEBUG [master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.small.0-1731055885261 {}] cleaner.HFileCleaner(306): Exit Thread[master/6ac96558fc7f:0:becomeActiveMaster-HFileCleaner.small.0-1731055885261,5,FailOnTimeoutGroup] 2024-11-08T08:51:27,479 INFO [M:0;6ac96558fc7f:46451 {}] hbase.ChoreService(370): Chore service for: master/6ac96558fc7f:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-11-08T08:51:27,479 INFO [M:0;6ac96558fc7f:46451 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-11-08T08:51:27,479 DEBUG [M:0;6ac96558fc7f:46451 {}] master.HMaster(1795): Stopping service threads 2024-11-08T08:51:27,479 INFO [M:0;6ac96558fc7f:46451 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-11-08T08:51:27,480 INFO [M:0;6ac96558fc7f:46451 {}] procedure2.ProcedureExecutor(723): Stopping 2024-11-08T08:51:27,480 INFO [M:0;6ac96558fc7f:46451 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-11-08T08:51:27,480 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-11-08T08:51:27,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-11-08T08:51:27,492 DEBUG [M:0;6ac96558fc7f:46451 {}] zookeeper.ZKUtil(347): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-11-08T08:51:27,492 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-11-08T08:51:27,492 WARN [M:0;6ac96558fc7f:46451 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-11-08T08:51:27,493 INFO [M:0;6ac96558fc7f:46451 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/.lastflushedseqids 2024-11-08T08:51:27,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741843_1019 (size=127) 2024-11-08T08:51:27,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741843_1019 (size=127) 2024-11-08T08:51:27,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741843_1019 (size=127) 2024-11-08T08:51:27,501 INFO [M:0;6ac96558fc7f:46451 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-11-08T08:51:27,501 INFO [M:0;6ac96558fc7f:46451 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-11-08T08:51:27,501 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-11-08T08:51:27,501 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:27,501 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:27,501 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-11-08T08:51:27,501 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:27,502 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.82 KB heapSize=34.11 KB 2024-11-08T08:51:27,523 DEBUG [M:0;6ac96558fc7f:46451 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/46ad707c2bc34dbf87298e2f301c9d02 is 82, key is hbase:meta,,1/info:regioninfo/1731055886002/Put/seqid=0 2024-11-08T08:51:27,525 WARN [IPC Server handler 3 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-11-08T08:51:27,525 WARN [IPC Server handler 3 on default port 41273 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=3, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-11-08T08:51:27,525 WARN [IPC Server handler 3 on default port 41273 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 3 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-11-08T08:51:27,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741844_1020 (size=5672) 2024-11-08T08:51:27,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741844_1020 (size=5672) 2024-11-08T08:51:27,530 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/46ad707c2bc34dbf87298e2f301c9d02 2024-11-08T08:51:27,545 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,545 INFO [RS:2;6ac96558fc7f:42453 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:27,545 INFO [RS:2;6ac96558fc7f:42453 {}] regionserver.HRegionServer(1031): Exiting; stopping=6ac96558fc7f,42453,1731055884861; zookeeper connection closed. 2024-11-08T08:51:27,545 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:42453-0x10119a494c90003, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,545 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@3726f6d9 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@3726f6d9 2024-11-08T08:51:27,553 DEBUG [M:0;6ac96558fc7f:46451 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/5b8aeedabfb949a687c2b248493490dc is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1731055886518/Put/seqid=0 2024-11-08T08:51:27,555 INFO [RS:1;6ac96558fc7f:39649 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:27,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,555 INFO [RS:1;6ac96558fc7f:39649 {}] regionserver.HRegionServer(1031): Exiting; stopping=6ac96558fc7f,39649,1731055884817; zookeeper connection closed. 2024-11-08T08:51:27,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39649-0x10119a494c90002, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,556 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6a4539c9 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6a4539c9 2024-11-08T08:51:27,556 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-11-08T08:51:27,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741845_1021 (size=6438) 2024-11-08T08:51:27,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741845_1021 (size=6438) 2024-11-08T08:51:27,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741845_1021 (size=6438) 2024-11-08T08:51:27,561 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.14 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/5b8aeedabfb949a687c2b248493490dc 2024-11-08T08:51:27,588 DEBUG [M:0;6ac96558fc7f:46451 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/93e8792022be44e7b22f3b1b44b81e16 is 69, key is 6ac96558fc7f,39649,1731055884817/rs:state/1731055885333/Put/seqid=0 2024-11-08T08:51:27,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741846_1022 (size=5294) 2024-11-08T08:51:27,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741846_1022 (size=5294) 2024-11-08T08:51:27,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741846_1022 (size=5294) 2024-11-08T08:51:27,597 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/93e8792022be44e7b22f3b1b44b81e16 2024-11-08T08:51:27,604 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/46ad707c2bc34dbf87298e2f301c9d02 as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/46ad707c2bc34dbf87298e2f301c9d02 2024-11-08T08:51:27,612 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/46ad707c2bc34dbf87298e2f301c9d02, entries=8, sequenceid=72, filesize=5.5 K 2024-11-08T08:51:27,614 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/5b8aeedabfb949a687c2b248493490dc as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/5b8aeedabfb949a687c2b248493490dc 2024-11-08T08:51:27,622 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/5b8aeedabfb949a687c2b248493490dc, entries=8, sequenceid=72, filesize=6.3 K 2024-11-08T08:51:27,624 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/93e8792022be44e7b22f3b1b44b81e16 as hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/93e8792022be44e7b22f3b1b44b81e16 2024-11-08T08:51:27,631 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:41273/user/jenkins/test-data/e3781d82-86be-749d-3004-1aabd830af90/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/93e8792022be44e7b22f3b1b44b81e16, entries=3, sequenceid=72, filesize=5.2 K 2024-11-08T08:51:27,633 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 132ms, sequenceid=72, compaction requested=false 2024-11-08T08:51:27,634 INFO [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-11-08T08:51:27,635 DEBUG [M:0;6ac96558fc7f:46451 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1731055887501Disabling compacts and flushes for region at 1731055887501Disabling writes for close at 1731055887501Obtaining lock to block concurrent updates at 1731055887502 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1731055887502Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27462, getHeapSize=34864, getOffHeapSize=0, getCellsCount=85 at 1731055887502Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1731055887503 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1731055887503Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1731055887523 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1731055887523Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1731055887536 (+13 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1731055887552 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1731055887552Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1731055887568 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1731055887588 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1731055887588Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1fa361bb: reopening flushed file at 1731055887603 (+15 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@43c5dd73: reopening flushed file at 1731055887612 (+9 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@713119a5: reopening flushed file at 1731055887622 (+10 ms)Finished flush of dataSize ~26.82 KB/27462, heapSize ~33.81 KB/34624, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 132ms, sequenceid=72, compaction requested=false at 1731055887633 (+11 ms)Writing region close event to WAL at 1731055887634 (+1 ms)Closed at 1731055887634 2024-11-08T08:51:27,635 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,635 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,635 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,635 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,636 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-11-08T08:51:27,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36041 is added to blk_1073741830_1006 (size=32665) 2024-11-08T08:51:27,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38231 is added to blk_1073741830_1006 (size=32665) 2024-11-08T08:51:27,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36421 is added to blk_1073741830_1006 (size=32665) 2024-11-08T08:51:27,639 INFO [M:0;6ac96558fc7f:46451 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-11-08T08:51:27,639 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-11-08T08:51:27,639 INFO [M:0;6ac96558fc7f:46451 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.3:46451 2024-11-08T08:51:27,640 INFO [M:0;6ac96558fc7f:46451 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-11-08T08:51:27,749 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,749 INFO [M:0;6ac96558fc7f:46451 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-11-08T08:51:27,750 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:46451-0x10119a494c90000, quorum=127.0.0.1:64730, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-11-08T08:51:27,753 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7261385{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:27,754 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@242a3163{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:27,754 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:27,754 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@15b92333{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:27,754 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6229ef54{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:27,757 WARN [BP-614830020-172.17.0.3-1731055881703 heartbeating to localhost/127.0.0.1:41273 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-08T08:51:27,757 WARN [BP-614830020-172.17.0.3-1731055881703 heartbeating to localhost/127.0.0.1:41273 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-614830020-172.17.0.3-1731055881703 (Datanode Uuid b9fc0088-a82b-4736-9c38-4f3597a87d67) service to localhost/127.0.0.1:41273 2024-11-08T08:51:27,759 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data5/current/BP-614830020-172.17.0.3-1731055881703 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:27,759 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data6/current/BP-614830020-172.17.0.3-1731055881703 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:27,759 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-08T08:51:27,760 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-08T08:51:27,760 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-08T08:51:27,762 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@47310287{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:27,762 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2b50a4e4{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:27,762 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:27,763 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@19cdd4f9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:27,763 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5ffbd1c8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:27,764 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-08T08:51:27,764 WARN [BP-614830020-172.17.0.3-1731055881703 heartbeating to localhost/127.0.0.1:41273 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-08T08:51:27,764 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-08T08:51:27,764 WARN [BP-614830020-172.17.0.3-1731055881703 heartbeating to localhost/127.0.0.1:41273 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-614830020-172.17.0.3-1731055881703 (Datanode Uuid 1de92640-8f18-4d49-b1a2-248977cbb17b) service to localhost/127.0.0.1:41273 2024-11-08T08:51:27,765 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data3/current/BP-614830020-172.17.0.3-1731055881703 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:27,765 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data4/current/BP-614830020-172.17.0.3-1731055881703 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:27,765 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-08T08:51:27,767 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6e45ef60{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-11-08T08:51:27,767 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6be33115{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:27,767 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:27,768 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@49cc577f{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:27,768 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@22e42287{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:27,769 WARN [BP-614830020-172.17.0.3-1731055881703 heartbeating to localhost/127.0.0.1:41273 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-11-08T08:51:27,769 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-11-08T08:51:27,769 WARN [BP-614830020-172.17.0.3-1731055881703 heartbeating to localhost/127.0.0.1:41273 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-614830020-172.17.0.3-1731055881703 (Datanode Uuid 9618e1bd-b775-4920-a84c-544796c2f2d0) service to localhost/127.0.0.1:41273 2024-11-08T08:51:27,769 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-11-08T08:51:27,769 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data1/current/BP-614830020-172.17.0.3-1731055881703 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:27,769 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/cluster_051708f3-47e7-ce8b-588d-c57e97a67aa7/data/data2/current/BP-614830020-172.17.0.3-1731055881703 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-11-08T08:51:27,769 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-11-08T08:51:27,774 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@26e4a54d{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-11-08T08:51:27,775 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@249bb0e5{HTTP/1.1, (http/1.1)}{localhost:0} 2024-11-08T08:51:27,775 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-11-08T08:51:27,775 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@778ab8c6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-11-08T08:51:27,775 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@642ad787{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-3/hbase-server/target/test-data/c2b7d31e-dc97-3939-8082-2e8fd4227400/hadoop.log.dir/,STOPPED} 2024-11-08T08:51:27,782 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-11-08T08:51:27,806 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-11-08T08:51:27,813 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=149 (was 89) - Thread LEAK? -, OpenFileDescriptor=518 (was 433) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=152 (was 121) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=9015 (was 9195)