2024-12-06 01:53:13,869 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-06 01:53:13,881 main DEBUG Took 0.010381 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-06 01:53:13,882 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-06 01:53:13,882 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-06 01:53:13,883 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-06 01:53:13,884 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,899 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-06 01:53:13,912 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,914 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,915 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,915 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,916 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,917 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,918 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,918 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,919 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,919 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,920 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,921 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,922 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,922 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,923 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,923 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,924 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,924 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,925 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,925 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,926 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,926 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,927 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,928 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-06 01:53:13,928 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,929 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-06 01:53:13,931 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-06 01:53:13,933 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-06 01:53:13,935 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-06 01:53:13,936 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-06 01:53:13,937 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-06 01:53:13,938 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-06 01:53:13,950 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-06 01:53:13,954 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-06 01:53:13,957 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-06 01:53:13,957 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-06 01:53:13,958 main DEBUG createAppenders(={Console}) 2024-12-06 01:53:13,959 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba initialized 2024-12-06 01:53:13,959 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba 2024-12-06 01:53:13,959 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@24569dba OK. 2024-12-06 01:53:13,960 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-06 01:53:13,961 main DEBUG OutputStream closed 2024-12-06 01:53:13,961 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-06 01:53:13,961 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-06 01:53:13,962 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@49c7b90e OK 2024-12-06 01:53:14,046 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-06 01:53:14,049 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-06 01:53:14,050 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-06 01:53:14,051 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-06 01:53:14,052 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-06 01:53:14,052 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-06 01:53:14,053 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-06 01:53:14,053 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-06 01:53:14,054 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-06 01:53:14,054 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-06 01:53:14,055 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-06 01:53:14,055 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-06 01:53:14,056 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-06 01:53:14,056 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-06 01:53:14,057 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-06 01:53:14,057 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-06 01:53:14,058 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-06 01:53:14,059 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-06 01:53:14,062 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-06 01:53:14,062 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-logging/target/hbase-logging-4.0.0-alpha-1-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@35432107) with optional ClassLoader: null 2024-12-06 01:53:14,062 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-06 01:53:14,063 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@35432107] started OK. 2024-12-06T01:53:14,081 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC timeout: 26 mins 2024-12-06 01:53:14,084 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-06 01:53:14,085 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-06T01:53:14,358 DEBUG [main {}] hbase.HBaseTestingUtil(323): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338 2024-12-06T01:53:14,384 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e, deleteOnExit=true 2024-12-06T01:53:14,385 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/test.cache.data in system properties and HBase conf 2024-12-06T01:53:14,386 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.tmp.dir in system properties and HBase conf 2024-12-06T01:53:14,386 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir in system properties and HBase conf 2024-12-06T01:53:14,387 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-06T01:53:14,387 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-06T01:53:14,388 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-06T01:53:14,497 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-06T01:53:14,632 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-06T01:53:14,637 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-06T01:53:14,638 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-06T01:53:14,639 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-06T01:53:14,639 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T01:53:14,640 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-06T01:53:14,641 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-06T01:53:14,641 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T01:53:14,642 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T01:53:14,643 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-06T01:53:14,643 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/nfs.dump.dir in system properties and HBase conf 2024-12-06T01:53:14,644 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/java.io.tmpdir in system properties and HBase conf 2024-12-06T01:53:14,644 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T01:53:14,645 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-06T01:53:14,646 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-06T01:53:15,533 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-06T01:53:15,615 INFO [Time-limited test {}] log.Log(170): Logging initialized @2502ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-06T01:53:15,696 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:15,762 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:15,786 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:15,787 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:15,788 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T01:53:15,805 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:15,808 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:15,810 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:16,012 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@76e4c45c{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/java.io.tmpdir/jetty-localhost-43113-hadoop-hdfs-3_4_1-tests_jar-_-any-15136824760723574990/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T01:53:16,019 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:43113} 2024-12-06T01:53:16,019 INFO [Time-limited test {}] server.Server(415): Started @2908ms 2024-12-06T01:53:16,431 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:16,440 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:16,442 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:16,442 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:16,442 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-06T01:53:16,443 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@32fec40a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:16,443 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@17f1c7fc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:16,573 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7e705dc8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/java.io.tmpdir/jetty-localhost-39431-hadoop-hdfs-3_4_1-tests_jar-_-any-3527022088986118832/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:16,573 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2ad1569e{HTTP/1.1, (http/1.1)}{localhost:39431} 2024-12-06T01:53:16,574 INFO [Time-limited test {}] server.Server(415): Started @3462ms 2024-12-06T01:53:16,633 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T01:53:16,771 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:16,778 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:16,796 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:16,797 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:16,797 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T01:53:16,798 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4a91ec1e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:16,799 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2c2c5be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:16,937 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@26b068f7{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/java.io.tmpdir/jetty-localhost-44583-hadoop-hdfs-3_4_1-tests_jar-_-any-15575543184642941648/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:16,937 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5739b847{HTTP/1.1, (http/1.1)}{localhost:44583} 2024-12-06T01:53:16,938 INFO [Time-limited test {}] server.Server(415): Started @3826ms 2024-12-06T01:53:16,940 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T01:53:16,976 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:16,982 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:16,984 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:16,984 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:16,984 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-06T01:53:16,986 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7e1f796{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:16,987 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1023f385{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:17,115 WARN [Thread-107 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data2/current/BP-1823148289-172.17.0.2-1733449995270/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:17,115 WARN [Thread-108 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data4/current/BP-1823148289-172.17.0.2-1733449995270/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:17,115 WARN [Thread-106 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data3/current/BP-1823148289-172.17.0.2-1733449995270/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:17,116 WARN [Thread-105 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data1/current/BP-1823148289-172.17.0.2-1733449995270/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:17,118 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7f750918{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/java.io.tmpdir/jetty-localhost-36559-hadoop-hdfs-3_4_1-tests_jar-_-any-14060305055533940423/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:17,118 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@86bf2a7{HTTP/1.1, (http/1.1)}{localhost:36559} 2024-12-06T01:53:17,119 INFO [Time-limited test {}] server.Server(415): Started @4007ms 2024-12-06T01:53:17,121 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T01:53:17,166 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T01:53:17,167 WARN [Thread-82 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T01:53:17,240 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data5/current/BP-1823148289-172.17.0.2-1733449995270/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:17,241 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data6/current/BP-1823148289-172.17.0.2-1733449995270/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:17,243 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9a3898dd47fd08ad with lease ID 0xd41642eba3d8fa70: Processing first storage report for DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e from datanode DatanodeRegistration(127.0.0.1:35771, datanodeUuid=c6fbe4bd-4bd6-437c-85cc-fe0f58012a14, infoPort=34441, infoSecurePort=0, ipcPort=42953, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270) 2024-12-06T01:53:17,244 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9a3898dd47fd08ad with lease ID 0xd41642eba3d8fa70: from storage DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e node DatanodeRegistration(127.0.0.1:35771, datanodeUuid=c6fbe4bd-4bd6-437c-85cc-fe0f58012a14, infoPort=34441, infoSecurePort=0, ipcPort=42953, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2024-12-06T01:53:17,245 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2048599d71d3a762 with lease ID 0xd41642eba3d8fa71: Processing first storage report for DS-249bbd57-6ebf-4f46-b94d-ed37654394bf from datanode DatanodeRegistration(127.0.0.1:33869, datanodeUuid=1ea54db9-a065-44f6-9d1e-08a154ee24c7, infoPort=39141, infoSecurePort=0, ipcPort=41933, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270) 2024-12-06T01:53:17,245 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2048599d71d3a762 with lease ID 0xd41642eba3d8fa71: from storage DS-249bbd57-6ebf-4f46-b94d-ed37654394bf node DatanodeRegistration(127.0.0.1:33869, datanodeUuid=1ea54db9-a065-44f6-9d1e-08a154ee24c7, infoPort=39141, infoSecurePort=0, ipcPort=41933, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-06T01:53:17,245 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9a3898dd47fd08ad with lease ID 0xd41642eba3d8fa70: Processing first storage report for DS-76ee940b-f1e5-42e5-ae60-59a7d491e832 from datanode DatanodeRegistration(127.0.0.1:35771, datanodeUuid=c6fbe4bd-4bd6-437c-85cc-fe0f58012a14, infoPort=34441, infoSecurePort=0, ipcPort=42953, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270) 2024-12-06T01:53:17,245 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9a3898dd47fd08ad with lease ID 0xd41642eba3d8fa70: from storage DS-76ee940b-f1e5-42e5-ae60-59a7d491e832 node DatanodeRegistration(127.0.0.1:35771, datanodeUuid=c6fbe4bd-4bd6-437c-85cc-fe0f58012a14, infoPort=34441, infoSecurePort=0, ipcPort=42953, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:17,246 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2048599d71d3a762 with lease ID 0xd41642eba3d8fa71: Processing first storage report for DS-47e7554b-e355-4040-bcff-ee80487334d1 from datanode DatanodeRegistration(127.0.0.1:33869, datanodeUuid=1ea54db9-a065-44f6-9d1e-08a154ee24c7, infoPort=39141, infoSecurePort=0, ipcPort=41933, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270) 2024-12-06T01:53:17,246 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2048599d71d3a762 with lease ID 0xd41642eba3d8fa71: from storage DS-47e7554b-e355-4040-bcff-ee80487334d1 node DatanodeRegistration(127.0.0.1:33869, datanodeUuid=1ea54db9-a065-44f6-9d1e-08a154ee24c7, infoPort=39141, infoSecurePort=0, ipcPort=41933, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:17,290 WARN [Thread-119 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T01:53:17,295 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x547c09f0e31335f5 with lease ID 0xd41642eba3d8fa72: Processing first storage report for DS-3a7455aa-2d6f-4683-9e30-be75498032d4 from datanode DatanodeRegistration(127.0.0.1:39971, datanodeUuid=31c47e50-3164-417d-b054-0b0affaffc70, infoPort=45655, infoSecurePort=0, ipcPort=45649, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270) 2024-12-06T01:53:17,296 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x547c09f0e31335f5 with lease ID 0xd41642eba3d8fa72: from storage DS-3a7455aa-2d6f-4683-9e30-be75498032d4 node DatanodeRegistration(127.0.0.1:39971, datanodeUuid=31c47e50-3164-417d-b054-0b0affaffc70, infoPort=45655, infoSecurePort=0, ipcPort=45649, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:17,296 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x547c09f0e31335f5 with lease ID 0xd41642eba3d8fa72: Processing first storage report for DS-e6a7515e-13b2-4c7e-879b-602837de3928 from datanode DatanodeRegistration(127.0.0.1:39971, datanodeUuid=31c47e50-3164-417d-b054-0b0affaffc70, infoPort=45655, infoSecurePort=0, ipcPort=45649, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270) 2024-12-06T01:53:17,296 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x547c09f0e31335f5 with lease ID 0xd41642eba3d8fa72: from storage DS-e6a7515e-13b2-4c7e-879b-602837de3928 node DatanodeRegistration(127.0.0.1:39971, datanodeUuid=31c47e50-3164-417d-b054-0b0affaffc70, infoPort=45655, infoSecurePort=0, ipcPort=45649, storageInfo=lv=-57;cid=testClusterID;nsid=1107911344;c=1733449995270), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:17,516 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338 2024-12-06T01:53:17,592 WARN [Time-limited test {}] erasurecode.ErasureCodeNative(55): ISA-L support is not available in your platform... using builtin-java codec where applicable 2024-12-06T01:53:17,646 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=158, OpenFileDescriptor=391, MaxFileDescriptor=1048576, SystemLoadAverage=323, ProcessCount=11, AvailableMemoryMB=10204 2024-12-06T01:53:17,648 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-06T01:53:17,658 INFO [Time-limited test {}] hbase.HBaseTestingUtil(821): NOT STARTING DFS 2024-12-06T01:53:17,777 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/zookeeper_0, clientPort=61613, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-06T01:53:17,790 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=61613 2024-12-06T01:53:17,806 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:17,809 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:17,901 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:17,902 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:17,957 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:35292 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775792_1001] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35292 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:17,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775792_1002 (size=7) 2024-12-06T01:53:18,380 WARN [Time-limited test {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:18,390 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72 with version=8 2024-12-06T01:53:18,390 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1139): Setting hbase.fs.tmp.dir to hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/hbase-staging 2024-12-06T01:53:18,488 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-06T01:53:18,747 INFO [Time-limited test {}] client.ConnectionUtils(128): master/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:18,760 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:18,761 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:18,767 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:18,768 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:18,768 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:18,955 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-06T01:53:19,016 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-06T01:53:19,024 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-06T01:53:19,028 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:19,056 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 11359 (auto-detected) 2024-12-06T01:53:19,057 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-06T01:53:19,076 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:36327 2024-12-06T01:53:19,101 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:36327 connecting to ZooKeeper ensemble=127.0.0.1:61613 2024-12-06T01:53:19,133 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:363270x0, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:19,135 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:36327-0x1006507969e0000 connected 2024-12-06T01:53:19,163 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,166 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,175 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:19,179 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72, hbase.cluster.distributed=false 2024-12-06T01:53:19,201 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:19,206 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36327 2024-12-06T01:53:19,206 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36327 2024-12-06T01:53:19,207 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36327 2024-12-06T01:53:19,208 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36327 2024-12-06T01:53:19,208 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36327 2024-12-06T01:53:19,317 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:19,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,319 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:19,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,319 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:19,322 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T01:53:19,324 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:19,325 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:44821 2024-12-06T01:53:19,327 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:44821 connecting to ZooKeeper ensemble=127.0.0.1:61613 2024-12-06T01:53:19,328 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,332 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,339 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:448210x0, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:19,340 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:44821-0x1006507969e0001 connected 2024-12-06T01:53:19,340 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:19,345 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T01:53:19,354 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T01:53:19,356 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T01:53:19,361 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:19,361 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=44821 2024-12-06T01:53:19,362 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=44821 2024-12-06T01:53:19,362 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=44821 2024-12-06T01:53:19,363 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=44821 2024-12-06T01:53:19,363 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=44821 2024-12-06T01:53:19,381 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:19,381 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,381 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,382 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:19,382 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,382 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:19,382 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T01:53:19,383 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:19,384 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39257 2024-12-06T01:53:19,385 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39257 connecting to ZooKeeper ensemble=127.0.0.1:61613 2024-12-06T01:53:19,386 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,391 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,399 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:392570x0, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:19,400 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39257-0x1006507969e0002 connected 2024-12-06T01:53:19,400 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:19,401 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T01:53:19,402 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T01:53:19,403 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T01:53:19,405 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:19,408 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39257 2024-12-06T01:53:19,411 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39257 2024-12-06T01:53:19,412 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39257 2024-12-06T01:53:19,414 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39257 2024-12-06T01:53:19,415 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39257 2024-12-06T01:53:19,434 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:19,434 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,435 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,435 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:19,435 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:19,435 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:19,435 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T01:53:19,436 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:19,437 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:35623 2024-12-06T01:53:19,439 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:35623 connecting to ZooKeeper ensemble=127.0.0.1:61613 2024-12-06T01:53:19,441 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,444 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,451 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:356230x0, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:19,452 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35623-0x1006507969e0003 connected 2024-12-06T01:53:19,452 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:19,453 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T01:53:19,456 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T01:53:19,457 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T01:53:19,459 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:19,459 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35623 2024-12-06T01:53:19,460 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35623 2024-12-06T01:53:19,460 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35623 2024-12-06T01:53:19,461 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35623 2024-12-06T01:53:19,461 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35623 2024-12-06T01:53:19,481 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;00c764f7c3aa:36327 2024-12-06T01:53:19,482 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:19,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,493 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,515 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,516 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-06T01:53:19,518 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/00c764f7c3aa,36327,1733449998544 from backup master directory 2024-12-06T01:53:19,522 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,522 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,522 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:19,522 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,522 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:19,523 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:19,523 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:19,526 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-06T01:53:19,527 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-06T01:53:19,591 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/hbase.id] with ID: 2e542351-decd-4532-b0e8-b1dc03fdf6e1 2024-12-06T01:53:19,591 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/.tmp/hbase.id 2024-12-06T01:53:19,598 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,598 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,601 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:59698 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775776_1003] {}] datanode.DataXceiver(331): 127.0.0.1:33869:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59698 dst: /127.0.0.1:33869 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:19,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775776_1004 (size=42) 2024-12-06T01:53:19,607 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:19,608 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/.tmp/hbase.id]:[hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/hbase.id] 2024-12-06T01:53:19,652 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:19,656 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-06T01:53:19,676 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 18ms. 2024-12-06T01:53:19,680 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,680 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,680 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,680 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:19,692 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,693 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,696 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:35318 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775760_1005] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35318 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:19,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775760_1006 (size=196) 2024-12-06T01:53:19,703 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:19,718 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T01:53:19,720 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-06T01:53:19,726 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T01:53:19,756 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,757 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,760 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:35332 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775744_1007] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35332 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:19,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775744_1008 (size=1189) 2024-12-06T01:53:19,766 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:19,791 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store 2024-12-06T01:53:19,809 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,809 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:19,813 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:35348 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775728_1009] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35348 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:19,817 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775728_1010 (size=34) 2024-12-06T01:53:19,818 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:19,823 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-06T01:53:19,826 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:19,827 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T01:53:19,827 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:19,827 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:19,829 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T01:53:19,829 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:19,829 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:19,830 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733449999827Disabling compacts and flushes for region at 1733449999827Disabling writes for close at 1733449999829 (+2 ms)Writing region close event to WAL at 1733449999829Closed at 1733449999829 2024-12-06T01:53:19,832 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/.initializing 2024-12-06T01:53:19,832 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/WALs/00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:19,841 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T01:53:19,856 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C36327%2C1733449998544, suffix=, logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/WALs/00c764f7c3aa,36327,1733449998544, archiveDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/oldWALs, maxLogs=10 2024-12-06T01:53:19,886 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/WALs/00c764f7c3aa,36327,1733449998544/00c764f7c3aa%2C36327%2C1733449998544.1733449999861, exclude list is [], retry=0 2024-12-06T01:53:19,907 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.initialize(FanOutOneBlockAsyncDFSOutputHelper.java:413) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:472) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper$5.operationComplete(FanOutOneBlockAsyncDFSOutputHelper.java:467) ~[hbase-asyncfs-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hadoop.hbase.util.NettyFutureUtils.lambda$addListener$0(NettyFutureUtils.java:56) ~[hbase-common-4.0.0-alpha-1-SNAPSHOT.jar:4.0.0-alpha-1-SNAPSHOT] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:590) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners0(DefaultPromise.java:583) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:559) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:492) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:636) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:625) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:105) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.fulfillConnectPromise(AbstractEpollChannel.java:658) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:696) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:567) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:491) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:399) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[hbase-shaded-netty-4.1.9.jar:?] at org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[hbase-shaded-netty-4.1.9.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:19,909 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35771,DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e,DISK] 2024-12-06T01:53:19,909 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33869,DS-249bbd57-6ebf-4f46-b94d-ed37654394bf,DISK] 2024-12-06T01:53:19,909 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39971,DS-3a7455aa-2d6f-4683-9e30-be75498032d4,DISK] 2024-12-06T01:53:19,913 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2024-12-06T01:53:19,955 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/WALs/00c764f7c3aa,36327,1733449998544/00c764f7c3aa%2C36327%2C1733449998544.1733449999861 2024-12-06T01:53:19,956 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34441:34441),(127.0.0.1/127.0.0.1:39141:39141),(127.0.0.1/127.0.0.1:45655:45655)] 2024-12-06T01:53:19,956 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-06T01:53:19,957 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:19,960 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:19,961 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,001 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,029 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-06T01:53:20,033 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,036 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:20,036 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,040 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-06T01:53:20,040 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,041 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:20,041 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,044 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-06T01:53:20,044 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,045 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:20,046 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,048 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-06T01:53:20,049 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,049 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:20,050 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,053 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,054 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,060 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,061 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,064 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T01:53:20,067 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:20,073 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T01:53:20,075 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64009832, jitterRate=-0.04617917537689209}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T01:53:20,081 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733449999974Initializing all the Stores at 1733449999977 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733449999977Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733449999978 (+1 ms)Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733449999978Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733449999978Cleaning up temporary data from old regions at 1733450000061 (+83 ms)Region opened successfully at 1733450000081 (+20 ms) 2024-12-06T01:53:20,083 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-06T01:53:20,118 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@20266662, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:20,152 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-06T01:53:20,164 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-06T01:53:20,164 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-06T01:53:20,167 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-06T01:53:20,169 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 1 msec 2024-12-06T01:53:20,175 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 6 msec 2024-12-06T01:53:20,175 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-06T01:53:20,206 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-06T01:53:20,214 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-06T01:53:20,217 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-06T01:53:20,220 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-06T01:53:20,221 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-06T01:53:20,223 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-06T01:53:20,226 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-06T01:53:20,229 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-06T01:53:20,231 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-06T01:53:20,232 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-06T01:53:20,234 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-06T01:53:20,252 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-06T01:53:20,254 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,259 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:20,260 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,263 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=00c764f7c3aa,36327,1733449998544, sessionid=0x1006507969e0000, setting cluster-up flag (Was=false) 2024-12-06T01:53:20,276 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,276 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,276 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,276 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,283 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-06T01:53:20,285 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:20,291 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,291 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,291 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,291 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:20,298 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-06T01:53:20,300 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:20,308 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-06T01:53:20,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775788_1002 (size=7) 2024-12-06T01:53:20,325 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775789_1002 (size=7) 2024-12-06T01:53:20,365 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(746): ClusterId : 2e542351-decd-4532-b0e8-b1dc03fdf6e1 2024-12-06T01:53:20,365 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(746): ClusterId : 2e542351-decd-4532-b0e8-b1dc03fdf6e1 2024-12-06T01:53:20,365 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(746): ClusterId : 2e542351-decd-4532-b0e8-b1dc03fdf6e1 2024-12-06T01:53:20,368 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T01:53:20,368 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T01:53:20,368 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T01:53:20,374 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T01:53:20,374 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T01:53:20,375 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T01:53:20,375 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T01:53:20,375 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T01:53:20,375 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T01:53:20,383 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T01:53:20,383 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T01:53:20,383 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T01:53:20,384 DEBUG [RS:2;00c764f7c3aa:35623 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@27a2e64d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:20,384 DEBUG [RS:0;00c764f7c3aa:44821 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2773c3e2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:20,384 DEBUG [RS:1;00c764f7c3aa:39257 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@18b3ffe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:20,394 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:20,403 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;00c764f7c3aa:39257 2024-12-06T01:53:20,404 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;00c764f7c3aa:35623 2024-12-06T01:53:20,405 DEBUG [RS:0;00c764f7c3aa:44821 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;00c764f7c3aa:44821 2024-12-06T01:53:20,407 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T01:53:20,407 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T01:53:20,407 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T01:53:20,407 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T01:53:20,407 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T01:53:20,407 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T01:53:20,408 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T01:53:20,408 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T01:53:20,408 DEBUG [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T01:53:20,408 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-06T01:53:20,411 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c764f7c3aa,36327,1733449998544 with port=39257, startcode=1733449999380 2024-12-06T01:53:20,411 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c764f7c3aa,36327,1733449998544 with port=35623, startcode=1733449999433 2024-12-06T01:53:20,411 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c764f7c3aa,36327,1733449998544 with port=44821, startcode=1733449999277 2024-12-06T01:53:20,418 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-06T01:53:20,425 DEBUG [RS:2;00c764f7c3aa:35623 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T01:53:20,425 DEBUG [RS:0;00c764f7c3aa:44821 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T01:53:20,425 DEBUG [RS:1;00c764f7c3aa:39257 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T01:53:20,426 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 00c764f7c3aa,36327,1733449998544 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-06T01:53:20,435 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:20,436 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:20,436 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:20,436 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:20,436 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/00c764f7c3aa:0, corePoolSize=10, maxPoolSize=10 2024-12-06T01:53:20,437 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,437 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:20,437 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,440 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733450030440 2024-12-06T01:53:20,442 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-06T01:53:20,443 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-06T01:53:20,448 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-06T01:53:20,449 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-06T01:53:20,449 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-06T01:53:20,449 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-06T01:53:20,450 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,455 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-06T01:53:20,456 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-06T01:53:20,457 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-06T01:53:20,459 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-06T01:53:20,460 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-06T01:53:20,462 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.large.0-1733450000461,5,FailOnTimeoutGroup] 2024-12-06T01:53:20,466 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.small.0-1733450000463,5,FailOnTimeoutGroup] 2024-12-06T01:53:20,467 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,467 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-06T01:53:20,468 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:20,468 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,468 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-06T01:53:20,469 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,477 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44295, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T01:53:20,477 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:39899, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T01:53:20,477 INFO [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56303, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T01:53:20,478 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,478 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-06T01:53:20,484 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36327 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:20,487 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36327 {}] master.ServerManager(517): Registering regionserver=00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:20,493 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:20,493 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:20,501 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36327 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:20,502 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36327 {}] master.ServerManager(517): Registering regionserver=00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:20,501 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:35376 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775712_1012] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35376 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:20,506 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72 2024-12-06T01:53:20,506 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42511 2024-12-06T01:53:20,506 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T01:53:20,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775712_1013 (size=1321) 2024-12-06T01:53:20,508 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36327 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:20,508 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=36327 {}] master.ServerManager(517): Registering regionserver=00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:20,508 DEBUG [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72 2024-12-06T01:53:20,508 DEBUG [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42511 2024-12-06T01:53:20,508 DEBUG [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T01:53:20,512 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72 2024-12-06T01:53:20,512 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:42511 2024-12-06T01:53:20,512 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T01:53:20,513 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:20,514 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:20,515 DEBUG [RS:0;00c764f7c3aa:44821 {}] zookeeper.ZKUtil(111): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:20,515 DEBUG [RS:1;00c764f7c3aa:39257 {}] zookeeper.ZKUtil(111): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:20,515 WARN [RS:0;00c764f7c3aa:44821 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:20,515 WARN [RS:1;00c764f7c3aa:39257 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:20,515 INFO [RS:0;00c764f7c3aa:44821 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T01:53:20,515 INFO [RS:1;00c764f7c3aa:39257 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T01:53:20,516 DEBUG [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:20,516 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:20,517 DEBUG [RS:2;00c764f7c3aa:35623 {}] zookeeper.ZKUtil(111): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:20,517 WARN [RS:2;00c764f7c3aa:35623 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:20,517 INFO [RS:2;00c764f7c3aa:35623 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T01:53:20,517 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:20,518 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c764f7c3aa,39257,1733449999380] 2024-12-06T01:53:20,519 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c764f7c3aa,35623,1733449999433] 2024-12-06T01:53:20,519 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c764f7c3aa,44821,1733449999277] 2024-12-06T01:53:20,519 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-06T01:53:20,519 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72 2024-12-06T01:53:20,538 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:20,538 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:20,543 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:59734 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775696_1014] {}] datanode.DataXceiver(331): 127.0.0.1:33869:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59734 dst: /127.0.0.1:33869 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:20,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775696_1015 (size=32) 2024-12-06T01:53:20,551 WARN [PEWorker-1 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:20,552 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:20,555 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T01:53:20,558 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T01:53:20,558 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,559 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:20,559 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T01:53:20,561 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T01:53:20,561 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T01:53:20,562 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T01:53:20,562 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,563 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:20,563 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T01:53:20,564 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T01:53:20,566 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T01:53:20,566 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,567 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:20,567 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T01:53:20,569 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T01:53:20,569 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:20,571 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:20,571 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T01:53:20,573 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740 2024-12-06T01:53:20,574 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740 2024-12-06T01:53:20,577 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T01:53:20,577 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T01:53:20,578 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T01:53:20,580 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T01:53:20,580 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T01:53:20,582 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T01:53:20,582 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T01:53:20,588 INFO [RS:2;00c764f7c3aa:35623 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T01:53:20,588 INFO [RS:0;00c764f7c3aa:44821 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T01:53:20,588 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,588 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,588 INFO [RS:1;00c764f7c3aa:39257 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T01:53:20,588 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,592 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T01:53:20,592 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T01:53:20,592 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T01:53:20,597 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T01:53:20,599 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67121147, jitterRate=1.8303096294403076E-4}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T01:53:20,599 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T01:53:20,600 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T01:53:20,601 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733450000552Initializing all the Stores at 1733450000554 (+2 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450000554Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450000554Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450000554Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450000555 (+1 ms)Cleaning up temporary data from old regions at 1733450000577 (+22 ms)Region opened successfully at 1733450000600 (+23 ms) 2024-12-06T01:53:20,601 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T01:53:20,601 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T01:53:20,601 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T01:53:20,601 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T01:53:20,601 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T01:53:20,601 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,601 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,602 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,602 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,602 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,602 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,602 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,602 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,602 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T01:53:20,602 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,603 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:20,603 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:20,603 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,603 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:20,604 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:20,604 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,604 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:20,604 DEBUG [RS:1;00c764f7c3aa:39257 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:20,604 DEBUG [RS:2;00c764f7c3aa:35623 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:20,604 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,605 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,605 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:20,605 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:20,605 DEBUG [RS:0;00c764f7c3aa:44821 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:20,605 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T01:53:20,606 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,606 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733450000601Disabling compacts and flushes for region at 1733450000601Disabling writes for close at 1733450000601Writing region close event to WAL at 1733450000605 (+4 ms)Closed at 1733450000606 (+1 ms) 2024-12-06T01:53:20,606 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,35623,1733449999433-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:20,606 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,607 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,608 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,39257,1733449999380-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:20,608 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,608 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,608 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,608 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,608 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,44821,1733449999277-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:20,610 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:20,610 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-06T01:53:20,619 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-06T01:53:20,631 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T01:53:20,635 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-06T01:53:20,637 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T01:53:20,637 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T01:53:20,638 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T01:53:20,640 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,35623,1733449999433-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,640 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,44821,1733449999277-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,640 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,39257,1733449999380-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,640 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,640 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.Replication(171): 00c764f7c3aa,35623,1733449999433 started 2024-12-06T01:53:20,640 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,640 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,641 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.Replication(171): 00c764f7c3aa,39257,1733449999380 started 2024-12-06T01:53:20,641 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.Replication(171): 00c764f7c3aa,44821,1733449999277 started 2024-12-06T01:53:20,662 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,662 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1482): Serving as 00c764f7c3aa,35623,1733449999433, RpcServer on 00c764f7c3aa/172.17.0.2:35623, sessionid=0x1006507969e0003 2024-12-06T01:53:20,663 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T01:53:20,663 DEBUG [RS:2;00c764f7c3aa:35623 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:20,663 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,35623,1733449999433' 2024-12-06T01:53:20,663 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T01:53:20,664 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T01:53:20,665 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T01:53:20,665 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T01:53:20,665 DEBUG [RS:2;00c764f7c3aa:35623 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:20,665 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,35623,1733449999433' 2024-12-06T01:53:20,665 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T01:53:20,666 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T01:53:20,667 DEBUG [RS:2;00c764f7c3aa:35623 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T01:53:20,667 INFO [RS:2;00c764f7c3aa:35623 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T01:53:20,667 INFO [RS:2;00c764f7c3aa:35623 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T01:53:20,668 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,668 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:20,668 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(1482): Serving as 00c764f7c3aa,44821,1733449999277, RpcServer on 00c764f7c3aa/172.17.0.2:44821, sessionid=0x1006507969e0001 2024-12-06T01:53:20,668 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1482): Serving as 00c764f7c3aa,39257,1733449999380, RpcServer on 00c764f7c3aa/172.17.0.2:39257, sessionid=0x1006507969e0002 2024-12-06T01:53:20,668 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T01:53:20,668 DEBUG [RS:0;00c764f7c3aa:44821 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:20,668 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T01:53:20,668 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,44821,1733449999277' 2024-12-06T01:53:20,668 DEBUG [RS:1;00c764f7c3aa:39257 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:20,668 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T01:53:20,668 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,39257,1733449999380' 2024-12-06T01:53:20,669 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T01:53:20,669 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T01:53:20,669 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T01:53:20,670 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T01:53:20,670 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T01:53:20,670 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T01:53:20,670 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T01:53:20,670 DEBUG [RS:0;00c764f7c3aa:44821 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:20,670 DEBUG [RS:1;00c764f7c3aa:39257 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:20,670 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,44821,1733449999277' 2024-12-06T01:53:20,670 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,39257,1733449999380' 2024-12-06T01:53:20,670 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T01:53:20,671 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T01:53:20,671 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T01:53:20,671 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T01:53:20,672 DEBUG [RS:1;00c764f7c3aa:39257 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T01:53:20,672 DEBUG [RS:0;00c764f7c3aa:44821 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T01:53:20,672 INFO [RS:1;00c764f7c3aa:39257 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T01:53:20,672 INFO [RS:0;00c764f7c3aa:44821 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T01:53:20,672 INFO [RS:1;00c764f7c3aa:39257 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T01:53:20,673 INFO [RS:0;00c764f7c3aa:44821 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T01:53:20,774 INFO [RS:0;00c764f7c3aa:44821 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T01:53:20,774 INFO [RS:1;00c764f7c3aa:39257 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T01:53:20,774 INFO [RS:2;00c764f7c3aa:35623 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2024-12-06T01:53:20,778 INFO [RS:2;00c764f7c3aa:35623 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C35623%2C1733449999433, suffix=, logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,35623,1733449999433, archiveDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs, maxLogs=32 2024-12-06T01:53:20,778 INFO [RS:0;00c764f7c3aa:44821 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C44821%2C1733449999277, suffix=, logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,44821,1733449999277, archiveDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs, maxLogs=32 2024-12-06T01:53:20,778 INFO [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C39257%2C1733449999380, suffix=, logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380, archiveDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs, maxLogs=32 2024-12-06T01:53:20,786 WARN [00c764f7c3aa:36327 {}] assignment.AssignmentManager(2451): No servers available; cannot place 1 unassigned regions. 2024-12-06T01:53:20,798 DEBUG [RS:1;00c764f7c3aa:39257 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380/00c764f7c3aa%2C39257%2C1733449999380.1733450000783, exclude list is [], retry=0 2024-12-06T01:53:20,798 DEBUG [RS:0;00c764f7c3aa:44821 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,44821,1733449999277/00c764f7c3aa%2C44821%2C1733449999277.1733450000783, exclude list is [], retry=0 2024-12-06T01:53:20,799 DEBUG [RS:2;00c764f7c3aa:35623 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,35623,1733449999433/00c764f7c3aa%2C35623%2C1733449999433.1733450000783, exclude list is [], retry=0 2024-12-06T01:53:20,804 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39971,DS-3a7455aa-2d6f-4683-9e30-be75498032d4,DISK] 2024-12-06T01:53:20,804 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33869,DS-249bbd57-6ebf-4f46-b94d-ed37654394bf,DISK] 2024-12-06T01:53:20,805 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39971,DS-3a7455aa-2d6f-4683-9e30-be75498032d4,DISK] 2024-12-06T01:53:20,806 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35771,DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e,DISK] 2024-12-06T01:53:20,806 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33869,DS-249bbd57-6ebf-4f46-b94d-ed37654394bf,DISK] 2024-12-06T01:53:20,806 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35771,DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e,DISK] 2024-12-06T01:53:20,807 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33869,DS-249bbd57-6ebf-4f46-b94d-ed37654394bf,DISK] 2024-12-06T01:53:20,853 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35771,DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e,DISK] 2024-12-06T01:53:20,854 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39971,DS-3a7455aa-2d6f-4683-9e30-be75498032d4,DISK] 2024-12-06T01:53:20,857 INFO [RS:0;00c764f7c3aa:44821 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,44821,1733449999277/00c764f7c3aa%2C44821%2C1733449999277.1733450000783 2024-12-06T01:53:20,862 DEBUG [RS:0;00c764f7c3aa:44821 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39141:39141),(127.0.0.1/127.0.0.1:45655:45655),(127.0.0.1/127.0.0.1:34441:34441)] 2024-12-06T01:53:20,867 INFO [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380/00c764f7c3aa%2C39257%2C1733449999380.1733450000783 2024-12-06T01:53:20,867 INFO [RS:2;00c764f7c3aa:35623 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,35623,1733449999433/00c764f7c3aa%2C35623%2C1733449999433.1733450000783 2024-12-06T01:53:20,868 DEBUG [RS:2;00c764f7c3aa:35623 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:39141:39141),(127.0.0.1/127.0.0.1:45655:45655),(127.0.0.1/127.0.0.1:34441:34441)] 2024-12-06T01:53:20,869 DEBUG [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:45655:45655),(127.0.0.1/127.0.0.1:39141:39141),(127.0.0.1/127.0.0.1:34441:34441)] 2024-12-06T01:53:21,039 DEBUG [00c764f7c3aa:36327 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=3, allServersCount=3 2024-12-06T01:53:21,047 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(204): Hosts are {00c764f7c3aa=0} racks are {/default-rack=0} 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T01:53:21,054 INFO [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T01:53:21,054 INFO [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T01:53:21,054 INFO [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T01:53:21,054 DEBUG [00c764f7c3aa:36327 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T01:53:21,061 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:21,068 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c764f7c3aa,39257,1733449999380, state=OPENING 2024-12-06T01:53:21,073 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-06T01:53:21,075 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:21,075 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:21,075 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:21,075 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:21,076 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,076 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,076 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,076 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,078 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T01:53:21,080 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=00c764f7c3aa,39257,1733449999380}] 2024-12-06T01:53:21,256 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T01:53:21,259 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56533, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T01:53:21,270 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-06T01:53:21,271 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2024-12-06T01:53:21,271 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2024-12-06T01:53:21,275 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C39257%2C1733449999380.meta, suffix=.meta, logDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380, archiveDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs, maxLogs=32 2024-12-06T01:53:21,292 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(524): When create output stream for /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380/00c764f7c3aa%2C39257%2C1733449999380.meta.1733450001277.meta, exclude list is [], retry=0 2024-12-06T01:53:21,325 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:35771,DS-2612d9a5-b655-40dd-b81e-96c5ab4a340e,DISK] 2024-12-06T01:53:21,326 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:39971,DS-3a7455aa-2d6f-4683-9e30-be75498032d4,DISK] 2024-12-06T01:53:21,326 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:33869,DS-249bbd57-6ebf-4f46-b94d-ed37654394bf,DISK] 2024-12-06T01:53:21,338 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/WALs/00c764f7c3aa,39257,1733449999380/00c764f7c3aa%2C39257%2C1733449999380.meta.1733450001277.meta 2024-12-06T01:53:21,340 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:34441:34441),(127.0.0.1/127.0.0.1:39141:39141),(127.0.0.1/127.0.0.1:45655:45655)] 2024-12-06T01:53:21,340 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-06T01:53:21,342 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-06T01:53:21,345 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-06T01:53:21,351 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-06T01:53:21,361 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-06T01:53:21,362 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:21,362 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-06T01:53:21,362 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-06T01:53:21,366 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T01:53:21,368 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T01:53:21,369 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:21,370 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:21,370 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T01:53:21,372 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T01:53:21,372 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:21,373 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:21,373 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T01:53:21,375 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T01:53:21,375 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:21,376 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:21,376 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T01:53:21,378 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T01:53:21,378 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:21,379 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:21,379 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T01:53:21,381 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740 2024-12-06T01:53:21,384 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740 2024-12-06T01:53:21,388 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T01:53:21,388 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T01:53:21,389 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T01:53:21,393 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T01:53:21,395 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=58948011, jitterRate=-0.12160618603229523}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T01:53:21,395 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-06T01:53:21,397 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733450001363Writing region info on filesystem at 1733450001363Initializing all the Stores at 1733450001366 (+3 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450001366Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450001366Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450001366Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450001366Cleaning up temporary data from old regions at 1733450001388 (+22 ms)Running coprocessor post-open hooks at 1733450001395 (+7 ms)Region opened successfully at 1733450001397 (+2 ms) 2024-12-06T01:53:21,406 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733450001247 2024-12-06T01:53:21,421 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-06T01:53:21,422 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-06T01:53:21,424 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:21,429 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c764f7c3aa,39257,1733449999380, state=OPEN 2024-12-06T01:53:21,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:21,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:21,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:21,432 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:21,433 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:21,434 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:21,442 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-06T01:53:21,442 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=00c764f7c3aa,39257,1733449999380 in 355 msec 2024-12-06T01:53:21,449 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-06T01:53:21,449 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 826 msec 2024-12-06T01:53:21,451 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:21,451 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-06T01:53:21,472 DEBUG [PEWorker-1 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T01:53:21,473 DEBUG [PEWorker-1 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c764f7c3aa,39257,1733449999380, seqNum=-1] 2024-12-06T01:53:21,493 DEBUG [PEWorker-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T01:53:21,496 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45099, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T01:53:21,518 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 1.1730 sec 2024-12-06T01:53:21,518 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733450001518, completionTime=-1 2024-12-06T01:53:21,521 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-06T01:53:21,521 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-06T01:53:21,553 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-06T01:53:21,553 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733450061553 2024-12-06T01:53:21,553 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733450121553 2024-12-06T01:53:21,553 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 32 msec 2024-12-06T01:53:21,555 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(159): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-06T01:53:21,562 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,36327,1733449998544-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,562 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,36327,1733449998544-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,562 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,36327,1733449998544-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,564 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-00c764f7c3aa:36327, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,564 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,565 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,572 DEBUG [master/00c764f7c3aa:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-06T01:53:21,593 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 2.070sec 2024-12-06T01:53:21,595 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-06T01:53:21,596 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-06T01:53:21,597 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-06T01:53:21,597 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-06T01:53:21,597 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-06T01:53:21,598 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,36327,1733449998544-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:21,599 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,36327,1733449998544-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-06T01:53:21,603 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-06T01:53:21,604 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-06T01:53:21,604 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,36327,1733449998544-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:21,676 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7b3bf62c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T01:53:21,680 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-06T01:53:21,680 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-06T01:53:21,684 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 00c764f7c3aa,36327,-1 for getting cluster id 2024-12-06T01:53:21,687 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-06T01:53:21,696 DEBUG [HMaster-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '2e542351-decd-4532-b0e8-b1dc03fdf6e1' 2024-12-06T01:53:21,699 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-06T01:53:21,699 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "2e542351-decd-4532-b0e8-b1dc03fdf6e1" 2024-12-06T01:53:21,700 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4cb195e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T01:53:21,700 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [00c764f7c3aa,36327,-1] 2024-12-06T01:53:21,703 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-06T01:53:21,705 DEBUG [RPCClient-NioEventLoopGroup-6-1 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:21,706 INFO [HMaster-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55554, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-06T01:53:21,709 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1891862c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T01:53:21,709 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T01:53:21,716 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c764f7c3aa,39257,1733449999380, seqNum=-1] 2024-12-06T01:53:21,717 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T01:53:21,723 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43440, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T01:53:21,751 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:21,758 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-06T01:53:21,763 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.AsyncConnectionImpl(321): The fetched master address is 00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:21,767 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@58c42e76 2024-12-06T01:53:21,768 DEBUG [RPCClient-NioEventLoopGroup-6-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-06T01:53:21,771 INFO [HMaster-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55570, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-06T01:53:21,776 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T01:53:21,785 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-06T01:53:21,788 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-06T01:53:21,790 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-06T01:53:21,791 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:21,794 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-06T01:53:21,796 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:21,805 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:21,805 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:21,813 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:59778 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775680_1020] {}] datanode.DataXceiver(331): 127.0.0.1:33869:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59778 dst: /127.0.0.1:33869 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:21,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775680_1021 (size=392) 2024-12-06T01:53:21,820 WARN [PEWorker-3 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:21,823 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 95976c4da0d10d7e3a3e7ef44f43c019, NAME => 'TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72 2024-12-06T01:53:21,829 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:21,830 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:21,837 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:59802 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775664_1022] {}] datanode.DataXceiver(331): 127.0.0.1:33869:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59802 dst: /127.0.0.1:33869 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:21,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775664_1023 (size=51) 2024-12-06T01:53:21,846 WARN [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:21,847 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:21,847 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 95976c4da0d10d7e3a3e7ef44f43c019, disabling compactions & flushes 2024-12-06T01:53:21,847 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:21,847 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:21,847 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. after waiting 0 ms 2024-12-06T01:53:21,847 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:21,847 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:21,847 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 95976c4da0d10d7e3a3e7ef44f43c019: Waiting for close lock at 1733450001847Disabling compacts and flushes for region at 1733450001847Disabling writes for close at 1733450001847Writing region close event to WAL at 1733450001847Closed at 1733450001847 2024-12-06T01:53:21,850 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-06T01:53:21,855 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733450001850"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733450001850"}]},"ts":"1733450001850"} 2024-12-06T01:53:21,864 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-06T01:53:21,866 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-06T01:53:21,870 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733450001867"}]},"ts":"1733450001867"} 2024-12-06T01:53:21,876 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-06T01:53:21,877 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {00c764f7c3aa=0} racks are {/default-rack=0} 2024-12-06T01:53:21,878 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T01:53:21,878 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T01:53:21,879 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T01:53:21,879 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T01:53:21,879 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T01:53:21,879 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T01:53:21,879 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T01:53:21,879 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T01:53:21,879 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T01:53:21,879 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T01:53:21,881 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95976c4da0d10d7e3a3e7ef44f43c019, ASSIGN}] 2024-12-06T01:53:21,884 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95976c4da0d10d7e3a3e7ef44f43c019, ASSIGN 2024-12-06T01:53:21,885 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95976c4da0d10d7e3a3e7ef44f43c019, ASSIGN; state=OFFLINE, location=00c764f7c3aa,35623,1733449999433; forceNewPlan=false, retain=false 2024-12-06T01:53:21,903 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:22,038 INFO [00c764f7c3aa:36327 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-06T01:53:22,039 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=95976c4da0d10d7e3a3e7ef44f43c019, regionState=OPENING, regionLocation=00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:22,043 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95976c4da0d10d7e3a3e7ef44f43c019, ASSIGN because future has completed 2024-12-06T01:53:22,044 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 95976c4da0d10d7e3a3e7ef44f43c019, server=00c764f7c3aa,35623,1733449999433}] 2024-12-06T01:53:22,113 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:22,199 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T01:53:22,201 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59429, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T01:53:22,207 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:22,208 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 95976c4da0d10d7e3a3e7ef44f43c019, NAME => 'TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019.', STARTKEY => '', ENDKEY => ''} 2024-12-06T01:53:22,208 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,208 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:22,209 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,209 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,211 INFO [StoreOpener-95976c4da0d10d7e3a3e7ef44f43c019-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,213 INFO [StoreOpener-95976c4da0d10d7e3a3e7ef44f43c019-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 95976c4da0d10d7e3a3e7ef44f43c019 columnFamilyName cf 2024-12-06T01:53:22,213 DEBUG [StoreOpener-95976c4da0d10d7e3a3e7ef44f43c019-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:22,214 INFO [StoreOpener-95976c4da0d10d7e3a3e7ef44f43c019-1 {}] regionserver.HStore(327): Store=95976c4da0d10d7e3a3e7ef44f43c019/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:22,214 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,215 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,216 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,217 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,217 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,219 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,225 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T01:53:22,225 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 95976c4da0d10d7e3a3e7ef44f43c019; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61448653, jitterRate=-0.08434371650218964}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-06T01:53:22,225 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:22,226 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 95976c4da0d10d7e3a3e7ef44f43c019: Running coprocessor pre-open hook at 1733450002209Writing region info on filesystem at 1733450002209Initializing all the Stores at 1733450002210 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450002210Cleaning up temporary data from old regions at 1733450002217 (+7 ms)Running coprocessor post-open hooks at 1733450002225 (+8 ms)Region opened successfully at 1733450002226 (+1 ms) 2024-12-06T01:53:22,228 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019., pid=6, masterSystemTime=1733450002199 2024-12-06T01:53:22,232 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:22,232 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:22,233 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=95976c4da0d10d7e3a3e7ef44f43c019, regionState=OPEN, openSeqNum=2, regionLocation=00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:22,237 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 95976c4da0d10d7e3a3e7ef44f43c019, server=00c764f7c3aa,35623,1733449999433 because future has completed 2024-12-06T01:53:22,243 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-06T01:53:22,243 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 95976c4da0d10d7e3a3e7ef44f43c019, server=00c764f7c3aa,35623,1733449999433 in 195 msec 2024-12-06T01:53:22,248 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-06T01:53:22,248 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95976c4da0d10d7e3a3e7ef44f43c019, ASSIGN in 362 msec 2024-12-06T01:53:22,250 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-06T01:53:22,250 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733450002250"}]},"ts":"1733450002250"} 2024-12-06T01:53:22,253 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-06T01:53:22,255 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-06T01:53:22,258 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 475 msec 2024-12-06T01:53:22,424 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:22,425 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T01:53:22,425 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-06T01:53:22,427 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T01:53:22,434 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-06T01:53:22,434 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T01:53:22,435 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-06T01:53:22,446 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019., hostname=00c764f7c3aa,35623,1733449999433, seqNum=2] 2024-12-06T01:53:22,448 DEBUG [RPCClient-NioEventLoopGroup-6-3 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T01:53:22,451 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50248, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T01:53:22,461 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-06T01:53:22,467 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-06T01:53:22,469 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:22,469 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-06T01:53:22,472 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-06T01:53:22,474 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-06T01:53:22,573 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:22,638 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=35623 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-06T01:53:22,638 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:22,642 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 95976c4da0d10d7e3a3e7ef44f43c019 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-06T01:53:22,708 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/.tmp/cf/fe9125f9692244818c2707059e93b5f7 is 36, key is row/cf:cq/1733450002452/Put/seqid=0 2024-12-06T01:53:22,715 WARN [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:22,715 WARN [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:22,719 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1240345278_22 at /127.0.0.1:49196 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775648_1024] {}] datanode.DataXceiver(331): 127.0.0.1:33869:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49196 dst: /127.0.0.1:33869 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:22,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775648_1025 (size=4787) 2024-12-06T01:53:22,725 WARN [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:22,726 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/.tmp/cf/fe9125f9692244818c2707059e93b5f7 2024-12-06T01:53:22,774 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/.tmp/cf/fe9125f9692244818c2707059e93b5f7 as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/cf/fe9125f9692244818c2707059e93b5f7 2024-12-06T01:53:22,784 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:22,787 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/cf/fe9125f9692244818c2707059e93b5f7, entries=1, sequenceid=5, filesize=4.7 K 2024-12-06T01:53:22,797 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 95976c4da0d10d7e3a3e7ef44f43c019 in 152ms, sequenceid=5, compaction requested=false 2024-12-06T01:53:22,799 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestHBaseWalOnEC' 2024-12-06T01:53:22,802 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 95976c4da0d10d7e3a3e7ef44f43c019: 2024-12-06T01:53:22,802 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:22,804 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-06T01:53:22,806 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-06T01:53:22,814 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-06T01:53:22,814 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 336 msec 2024-12-06T01:53:22,818 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 353 msec 2024-12-06T01:53:23,093 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36327 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:23,093 INFO [RPCClient-NioEventLoopGroup-6-4 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T01:53:23,109 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-06T01:53:23,110 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T01:53:23,110 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:23,116 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,117 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,117 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-06T01:53:23,117 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-06T01:53:23,117 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=990235322, stopped=false 2024-12-06T01:53:23,118 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=00c764f7c3aa,36327,1733449998544 2024-12-06T01:53:23,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:23,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:23,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:23,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:23,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:23,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:23,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:23,120 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T01:53:23,120 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T01:53:23,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:23,120 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:23,121 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,121 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:23,121 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:23,121 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:23,121 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c764f7c3aa,44821,1733449999277' ***** 2024-12-06T01:53:23,121 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T01:53:23,121 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:23,122 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c764f7c3aa,39257,1733449999380' ***** 2024-12-06T01:53:23,122 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T01:53:23,122 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T01:53:23,122 INFO [regionserver/00c764f7c3aa:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:23,122 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c764f7c3aa,35623,1733449999433' ***** 2024-12-06T01:53:23,122 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T01:53:23,122 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T01:53:23,122 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T01:53:23,122 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T01:53:23,123 INFO [RS:2;00c764f7c3aa:35623 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T01:53:23,123 INFO [RS:1;00c764f7c3aa:39257 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T01:53:23,123 INFO [RS:2;00c764f7c3aa:35623 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T01:53:23,123 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T01:53:23,123 INFO [RS:1;00c764f7c3aa:39257 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T01:53:23,123 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T01:53:23,123 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(959): stopping server 00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:23,122 INFO [RS:0;00c764f7c3aa:44821 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T01:53:23,123 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(3091): Received CLOSE for 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:23,123 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:23,123 INFO [RS:0;00c764f7c3aa:44821 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T01:53:23,123 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(959): stopping server 00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:23,123 INFO [RS:1;00c764f7c3aa:39257 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;00c764f7c3aa:39257. 2024-12-06T01:53:23,123 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:23,124 INFO [RS:0;00c764f7c3aa:44821 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;00c764f7c3aa:44821. 2024-12-06T01:53:23,124 DEBUG [RS:1;00c764f7c3aa:39257 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:23,124 DEBUG [RS:0;00c764f7c3aa:44821 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:23,124 DEBUG [RS:1;00c764f7c3aa:39257 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,124 DEBUG [RS:0;00c764f7c3aa:44821 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,124 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T01:53:23,124 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(976): stopping server 00c764f7c3aa,44821,1733449999277; all regions closed. 2024-12-06T01:53:23,124 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T01:53:23,124 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T01:53:23,124 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-06T01:53:23,125 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(959): stopping server 00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:23,125 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:23,125 INFO [RS:2;00c764f7c3aa:35623 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;00c764f7c3aa:35623. 2024-12-06T01:53:23,125 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-06T01:53:23,125 DEBUG [RS:2;00c764f7c3aa:35623 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:23,125 DEBUG [RS:2;00c764f7c3aa:35623 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,125 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-06T01:53:23,126 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 95976c4da0d10d7e3a3e7ef44f43c019, disabling compactions & flushes 2024-12-06T01:53:23,126 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1325): Online Regions={95976c4da0d10d7e3a3e7ef44f43c019=TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019.} 2024-12-06T01:53:23,126 INFO [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:23,126 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-06T01:53:23,126 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. after waiting 0 ms 2024-12-06T01:53:23,126 DEBUG [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1351): Waiting on 95976c4da0d10d7e3a3e7ef44f43c019 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T01:53:23,126 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T01:53:23,126 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-06T01:53:23,131 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_1073741826_1016 (size=93) 2024-12-06T01:53:23,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741826_1016 (size=93) 2024-12-06T01:53:23,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_1073741826_1016 (size=93) 2024-12-06T01:53:23,140 DEBUG [RS:0;00c764f7c3aa:44821 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs 2024-12-06T01:53:23,141 INFO [RS:0;00c764f7c3aa:44821 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c764f7c3aa%2C44821%2C1733449999277:(num 1733450000783) 2024-12-06T01:53:23,141 DEBUG [RS:0;00c764f7c3aa:44821 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,141 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:23,141 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:23,141 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.ChoreService(370): Chore service for: regionserver/00c764f7c3aa:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:23,142 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T01:53:23,142 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T01:53:23,142 INFO [regionserver/00c764f7c3aa:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:23,142 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T01:53:23,142 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:23,142 INFO [RS:0;00c764f7c3aa:44821 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:44821 2024-12-06T01:53:23,146 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c764f7c3aa,44821,1733449999277 2024-12-06T01:53:23,147 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:23,147 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:23,148 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c764f7c3aa,44821,1733449999277] 2024-12-06T01:53:23,150 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c764f7c3aa,44821,1733449999277 already deleted, retry=false 2024-12-06T01:53:23,150 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c764f7c3aa,44821,1733449999277 expired; onlineServers=2 2024-12-06T01:53:23,151 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/default/TestHBaseWalOnEC/95976c4da0d10d7e3a3e7ef44f43c019/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-06T01:53:23,154 INFO [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:23,154 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 95976c4da0d10d7e3a3e7ef44f43c019: Waiting for close lock at 1733450003125Running coprocessor pre-close hooks at 1733450003125Disabling compacts and flushes for region at 1733450003125Disabling writes for close at 1733450003126 (+1 ms)Writing region close event to WAL at 1733450003129 (+3 ms)Running coprocessor post-close hooks at 1733450003152 (+23 ms)Closed at 1733450003154 (+2 ms) 2024-12-06T01:53:23,154 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019. 2024-12-06T01:53:23,169 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/info/e728719002914af6ab39a8d07ef26fe1 is 153, key is TestHBaseWalOnEC,,1733450001772.95976c4da0d10d7e3a3e7ef44f43c019./info:regioninfo/1733450002233/Put/seqid=0 2024-12-06T01:53:23,172 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,172 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,176 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1835267660_22 at /127.0.0.1:49218 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775632_1026] {}] datanode.DataXceiver(331): 127.0.0.1:33869:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49218 dst: /127.0.0.1:33869 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775632_1027 (size=6637) 2024-12-06T01:53:23,182 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:23,182 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/info/e728719002914af6ab39a8d07ef26fe1 2024-12-06T01:53:23,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775773_1004 (size=42) 2024-12-06T01:53:23,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775772_1004 (size=42) 2024-12-06T01:53:23,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775692_1015 (size=32) 2024-12-06T01:53:23,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775693_1015 (size=32) 2024-12-06T01:53:23,219 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/ns/85eb4b3b1927476b9c9cdc1e706e4196 is 43, key is default/ns:d/1733450001500/Put/seqid=0 2024-12-06T01:53:23,221 INFO [regionserver/00c764f7c3aa:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:23,221 INFO [regionserver/00c764f7c3aa:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:23,222 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,222 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,228 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1835267660_22 at /127.0.0.1:54114 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775616_1028] {}] datanode.DataXceiver(331): 127.0.0.1:39971:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54114 dst: /127.0.0.1:39971 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,231 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775616_1029 (size=5153) 2024-12-06T01:53:23,232 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:23,233 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/ns/85eb4b3b1927476b9c9cdc1e706e4196 2024-12-06T01:53:23,250 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:23,250 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44821-0x1006507969e0001, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:23,251 INFO [RS:0;00c764f7c3aa:44821 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:23,251 INFO [RS:0;00c764f7c3aa:44821 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c764f7c3aa,44821,1733449999277; zookeeper connection closed. 2024-12-06T01:53:23,251 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6968515e {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6968515e 2024-12-06T01:53:23,268 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/table/a69b55487d454a2a8c35959a3d23f8ad is 52, key is TestHBaseWalOnEC/table:state/1733450002250/Put/seqid=0 2024-12-06T01:53:23,270 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,271 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,274 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1835267660_22 at /127.0.0.1:37736 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775600_1030] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37736 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775600_1031 (size=5249) 2024-12-06T01:53:23,279 WARN [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:23,279 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/table/a69b55487d454a2a8c35959a3d23f8ad 2024-12-06T01:53:23,289 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/info/e728719002914af6ab39a8d07ef26fe1 as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/info/e728719002914af6ab39a8d07ef26fe1 2024-12-06T01:53:23,300 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/info/e728719002914af6ab39a8d07ef26fe1, entries=10, sequenceid=11, filesize=6.5 K 2024-12-06T01:53:23,304 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/ns/85eb4b3b1927476b9c9cdc1e706e4196 as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/ns/85eb4b3b1927476b9c9cdc1e706e4196 2024-12-06T01:53:23,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775757_1006 (size=196) 2024-12-06T01:53:23,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775756_1006 (size=196) 2024-12-06T01:53:23,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775741_1008 (size=1189) 2024-12-06T01:53:23,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_-9223372036854775740_1008 (size=1189) 2024-12-06T01:53:23,314 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/ns/85eb4b3b1927476b9c9cdc1e706e4196, entries=2, sequenceid=11, filesize=5.0 K 2024-12-06T01:53:23,316 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/.tmp/table/a69b55487d454a2a8c35959a3d23f8ad as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/table/a69b55487d454a2a8c35959a3d23f8ad 2024-12-06T01:53:23,326 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/table/a69b55487d454a2a8c35959a3d23f8ad, entries=2, sequenceid=11, filesize=5.1 K 2024-12-06T01:53:23,326 DEBUG [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-06T01:53:23,326 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(976): stopping server 00c764f7c3aa,35623,1733449999433; all regions closed. 2024-12-06T01:53:23,328 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 202ms, sequenceid=11, compaction requested=false 2024-12-06T01:53:23,328 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-06T01:53:23,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_1073741828_1018 (size=1298) 2024-12-06T01:53:23,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_1073741828_1018 (size=1298) 2024-12-06T01:53:23,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741828_1018 (size=1298) 2024-12-06T01:53:23,336 DEBUG [RS:2;00c764f7c3aa:35623 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs 2024-12-06T01:53:23,336 INFO [RS:2;00c764f7c3aa:35623 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c764f7c3aa%2C35623%2C1733449999433:(num 1733450000783) 2024-12-06T01:53:23,336 DEBUG [RS:2;00c764f7c3aa:35623 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,336 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:23,336 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:23,336 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.ChoreService(370): Chore service for: regionserver/00c764f7c3aa:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:23,337 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T01:53:23,337 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T01:53:23,337 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T01:53:23,337 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:23,337 INFO [RS:2;00c764f7c3aa:35623 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:35623 2024-12-06T01:53:23,338 INFO [regionserver/00c764f7c3aa:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:23,339 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c764f7c3aa,35623,1733449999433 2024-12-06T01:53:23,339 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:23,340 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:23,340 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c764f7c3aa,35623,1733449999433] 2024-12-06T01:53:23,342 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c764f7c3aa,35623,1733449999433 already deleted, retry=false 2024-12-06T01:53:23,343 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c764f7c3aa,35623,1733449999433 expired; onlineServers=1 2024-12-06T01:53:23,348 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-06T01:53:23,349 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-06T01:53:23,349 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T01:53:23,349 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733450003125Running coprocessor pre-close hooks at 1733450003126 (+1 ms)Disabling compacts and flushes for region at 1733450003126Disabling writes for close at 1733450003126Obtaining lock to block concurrent updates at 1733450003126Preparing flush snapshotting stores in 1588230740 at 1733450003126Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733450003127 (+1 ms)Flushing stores of hbase:meta,,1.1588230740 at 1733450003129 (+2 ms)Flushing 1588230740/info: creating writer at 1733450003129Flushing 1588230740/info: appending metadata at 1733450003165 (+36 ms)Flushing 1588230740/info: closing flushed file at 1733450003165Flushing 1588230740/ns: creating writer at 1733450003195 (+30 ms)Flushing 1588230740/ns: appending metadata at 1733450003217 (+22 ms)Flushing 1588230740/ns: closing flushed file at 1733450003218 (+1 ms)Flushing 1588230740/table: creating writer at 1733450003242 (+24 ms)Flushing 1588230740/table: appending metadata at 1733450003266 (+24 ms)Flushing 1588230740/table: closing flushed file at 1733450003266Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@16a0169f: reopening flushed file at 1733450003288 (+22 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@239fe0b1: reopening flushed file at 1733450003300 (+12 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@1617c44b: reopening flushed file at 1733450003314 (+14 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 202ms, sequenceid=11, compaction requested=false at 1733450003328 (+14 ms)Writing region close event to WAL at 1733450003333 (+5 ms)Running coprocessor post-close hooks at 1733450003349 (+16 ms)Closed at 1733450003349 2024-12-06T01:53:23,349 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-06T01:53:23,442 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:23,442 INFO [RS:2;00c764f7c3aa:35623 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:23,442 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35623-0x1006507969e0003, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:23,442 INFO [RS:2;00c764f7c3aa:35623 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c764f7c3aa,35623,1733449999433; zookeeper connection closed. 2024-12-06T01:53:23,442 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@40bb3312 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@40bb3312 2024-12-06T01:53:23,526 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(976): stopping server 00c764f7c3aa,39257,1733449999380; all regions closed. 2024-12-06T01:53:23,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_1073741829_1019 (size=2751) 2024-12-06T01:53:23,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_1073741829_1019 (size=2751) 2024-12-06T01:53:23,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741829_1019 (size=2751) 2024-12-06T01:53:23,533 DEBUG [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs 2024-12-06T01:53:23,533 INFO [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c764f7c3aa%2C39257%2C1733449999380.meta:.meta(num 1733450001277) 2024-12-06T01:53:23,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_1073741827_1017 (size=93) 2024-12-06T01:53:23,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741827_1017 (size=93) 2024-12-06T01:53:23,537 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_1073741827_1017 (size=93) 2024-12-06T01:53:23,542 DEBUG [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/oldWALs 2024-12-06T01:53:23,542 INFO [RS:1;00c764f7c3aa:39257 {}] wal.AbstractFSWAL(1259): Closed WAL: AsyncFSWAL 00c764f7c3aa%2C39257%2C1733449999380:(num 1733450000783) 2024-12-06T01:53:23,542 DEBUG [RS:1;00c764f7c3aa:39257 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:23,542 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:23,542 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:23,542 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.ChoreService(370): Chore service for: regionserver/00c764f7c3aa:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:23,542 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:23,543 INFO [regionserver/00c764f7c3aa:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:23,543 INFO [RS:1;00c764f7c3aa:39257 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39257 2024-12-06T01:53:23,545 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c764f7c3aa,39257,1733449999380 2024-12-06T01:53:23,545 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:23,545 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:23,547 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c764f7c3aa,39257,1733449999380] 2024-12-06T01:53:23,548 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c764f7c3aa,39257,1733449999380 already deleted, retry=false 2024-12-06T01:53:23,548 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c764f7c3aa,39257,1733449999380 expired; onlineServers=0 2024-12-06T01:53:23,549 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '00c764f7c3aa,36327,1733449998544' ***** 2024-12-06T01:53:23,549 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-06T01:53:23,549 INFO [M:0;00c764f7c3aa:36327 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:23,549 INFO [M:0;00c764f7c3aa:36327 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:23,549 DEBUG [M:0;00c764f7c3aa:36327 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-06T01:53:23,549 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-06T01:53:23,549 DEBUG [M:0;00c764f7c3aa:36327 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-06T01:53:23,549 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.small.0-1733450000463 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.small.0-1733450000463,5,FailOnTimeoutGroup] 2024-12-06T01:53:23,549 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.large.0-1733450000461 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.large.0-1733450000461,5,FailOnTimeoutGroup] 2024-12-06T01:53:23,550 INFO [M:0;00c764f7c3aa:36327 {}] hbase.ChoreService(370): Chore service for: master/00c764f7c3aa:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:23,550 INFO [M:0;00c764f7c3aa:36327 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:23,550 DEBUG [M:0;00c764f7c3aa:36327 {}] master.HMaster(1795): Stopping service threads 2024-12-06T01:53:23,550 INFO [M:0;00c764f7c3aa:36327 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-06T01:53:23,550 INFO [M:0;00c764f7c3aa:36327 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T01:53:23,550 INFO [M:0;00c764f7c3aa:36327 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-06T01:53:23,551 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-06T01:53:23,551 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:23,551 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:23,551 DEBUG [M:0;00c764f7c3aa:36327 {}] zookeeper.ZKUtil(347): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-06T01:53:23,551 WARN [M:0;00c764f7c3aa:36327 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-06T01:53:23,552 INFO [M:0;00c764f7c3aa:36327 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/.lastflushedseqids 2024-12-06T01:53:23,561 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,561 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,564 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:37766 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775584_1032] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37766 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,569 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775584_1033 (size=127) 2024-12-06T01:53:23,569 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:23,569 INFO [M:0;00c764f7c3aa:36327 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-06T01:53:23,570 INFO [M:0;00c764f7c3aa:36327 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-06T01:53:23,570 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T01:53:23,570 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:23,570 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:23,570 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T01:53:23,570 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:23,570 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-06T01:53:23,591 DEBUG [M:0;00c764f7c3aa:36327 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/636fdd72c5f64dc5aa21a9e81ac8c491 is 82, key is hbase:meta,,1/info:regioninfo/1733450001423/Put/seqid=0 2024-12-06T01:53:23,593 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,593 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,596 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:37782 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775568_1034] {}] datanode.DataXceiver(331): 127.0.0.1:35771:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37782 dst: /127.0.0.1:35771 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,599 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_-9223372036854775568_1035 (size=5672) 2024-12-06T01:53:23,600 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:23,600 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/636fdd72c5f64dc5aa21a9e81ac8c491 2024-12-06T01:53:23,627 DEBUG [M:0;00c764f7c3aa:36327 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/4afb833dddc344ecb23bc26ec33c247f is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733450002257/Put/seqid=0 2024-12-06T01:53:23,629 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,629 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,633 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:54150 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775552_1036] {}] datanode.DataXceiver(331): 127.0.0.1:39971:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54150 dst: /127.0.0.1:39971 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775552_1037 (size=6440) 2024-12-06T01:53:23,638 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:23,638 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.15 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/4afb833dddc344ecb23bc26ec33c247f 2024-12-06T01:53:23,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:23,647 INFO [RS:1;00c764f7c3aa:39257 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:23,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39257-0x1006507969e0002, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:23,647 INFO [RS:1;00c764f7c3aa:39257 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c764f7c3aa,39257,1733449999380; zookeeper connection closed. 2024-12-06T01:53:23,647 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4c0f4445 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4c0f4445 2024-12-06T01:53:23,648 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-06T01:53:23,663 DEBUG [M:0;00c764f7c3aa:36327 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/18d592360c0a4e18917f66198769c7a6 is 69, key is 00c764f7c3aa,35623,1733449999433/rs:state/1733450000508/Put/seqid=0 2024-12-06T01:53:23,664 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=3, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,665 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(531): Cannot allocate parity block(index=4, policy=RS-3-2-1024k). Exclude nodes=[]. There may not be enough datanodes or racks. You can check if the cluster topology supports the enabled erasure coding policies by running the command 'hdfs ec -verifyClusterSetup'. 2024-12-06T01:53:23,668 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1115649555_22 at /127.0.0.1:54160 [Receiving block BP-1823148289-172.17.0.2-1733449995270:blk_-9223372036854775536_1038] {}] datanode.DataXceiver(331): 127.0.0.1:39971:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:54160 dst: /127.0.0.1:39971 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-06T01:53:23,672 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_-9223372036854775536_1039 (size=5294) 2024-12-06T01:53:24,073 WARN [M:0;00c764f7c3aa:36327 {}] hdfs.DFSStripedOutputStream(1367): Block group <1> failed to write 2 blocks. It's at high risk of losing data. 2024-12-06T01:53:24,073 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/18d592360c0a4e18917f66198769c7a6 2024-12-06T01:53:24,086 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/636fdd72c5f64dc5aa21a9e81ac8c491 as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/636fdd72c5f64dc5aa21a9e81ac8c491 2024-12-06T01:53:24,094 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/636fdd72c5f64dc5aa21a9e81ac8c491, entries=8, sequenceid=72, filesize=5.5 K 2024-12-06T01:53:24,095 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/4afb833dddc344ecb23bc26ec33c247f as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/4afb833dddc344ecb23bc26ec33c247f 2024-12-06T01:53:24,103 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/4afb833dddc344ecb23bc26ec33c247f, entries=8, sequenceid=72, filesize=6.3 K 2024-12-06T01:53:24,104 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/18d592360c0a4e18917f66198769c7a6 as hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/18d592360c0a4e18917f66198769c7a6 2024-12-06T01:53:24,111 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/18d592360c0a4e18917f66198769c7a6, entries=3, sequenceid=72, filesize=5.2 K 2024-12-06T01:53:24,113 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 543ms, sequenceid=72, compaction requested=false 2024-12-06T01:53:24,115 INFO [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:24,115 DEBUG [M:0;00c764f7c3aa:36327 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733450003570Disabling compacts and flushes for region at 1733450003570Disabling writes for close at 1733450003570Obtaining lock to block concurrent updates at 1733450003570Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733450003570Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27480, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733450003571 (+1 ms)Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733450003572 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733450003572Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733450003590 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733450003590Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733450003609 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733450003626 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733450003626Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733450003646 (+20 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733450003662 (+16 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733450003662Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3fa3fbbd: reopening flushed file at 1733450004083 (+421 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@54ea9a3b: reopening flushed file at 1733450004094 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3ac0eb0: reopening flushed file at 1733450004103 (+9 ms)Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 543ms, sequenceid=72, compaction requested=false at 1733450004113 (+10 ms)Writing region close event to WAL at 1733450004115 (+2 ms)Closed at 1733450004115 2024-12-06T01:53:24,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33869 is added to blk_1073741825_1011 (size=32683) 2024-12-06T01:53:24,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39971 is added to blk_1073741825_1011 (size=32683) 2024-12-06T01:53:24,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741825_1011 (size=32683) 2024-12-06T01:53:24,120 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:24,120 INFO [M:0;00c764f7c3aa:36327 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-06T01:53:24,121 INFO [M:0;00c764f7c3aa:36327 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:36327 2024-12-06T01:53:24,121 INFO [M:0;00c764f7c3aa:36327 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:24,223 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:24,223 INFO [M:0;00c764f7c3aa:36327 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:24,223 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:36327-0x1006507969e0000, quorum=127.0.0.1:61613, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:24,227 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7f750918{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:24,230 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@86bf2a7{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:24,230 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:24,230 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1023f385{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:24,230 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7e1f796{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:24,236 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T01:53:24,236 WARN [BP-1823148289-172.17.0.2-1733449995270 heartbeating to localhost/127.0.0.1:42511 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T01:53:24,236 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T01:53:24,236 WARN [BP-1823148289-172.17.0.2-1733449995270 heartbeating to localhost/127.0.0.1:42511 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1823148289-172.17.0.2-1733449995270 (Datanode Uuid 31c47e50-3164-417d-b054-0b0affaffc70) service to localhost/127.0.0.1:42511 2024-12-06T01:53:24,238 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data5/current/BP-1823148289-172.17.0.2-1733449995270 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:24,238 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data6/current/BP-1823148289-172.17.0.2-1733449995270 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:24,239 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T01:53:24,242 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@26b068f7{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:24,242 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5739b847{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:24,242 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:24,242 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2c2c5be{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:24,242 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4a91ec1e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:24,243 WARN [BP-1823148289-172.17.0.2-1733449995270 heartbeating to localhost/127.0.0.1:42511 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T01:53:24,243 WARN [BP-1823148289-172.17.0.2-1733449995270 heartbeating to localhost/127.0.0.1:42511 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1823148289-172.17.0.2-1733449995270 (Datanode Uuid c6fbe4bd-4bd6-437c-85cc-fe0f58012a14) service to localhost/127.0.0.1:42511 2024-12-06T01:53:24,244 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data3/current/BP-1823148289-172.17.0.2-1733449995270 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:24,244 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data4/current/BP-1823148289-172.17.0.2-1733449995270 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:24,245 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T01:53:24,246 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7e705dc8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:24,247 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2ad1569e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:24,247 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:24,247 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@17f1c7fc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:24,247 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@32fec40a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:24,248 WARN [BP-1823148289-172.17.0.2-1733449995270 heartbeating to localhost/127.0.0.1:42511 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T01:53:24,248 WARN [BP-1823148289-172.17.0.2-1733449995270 heartbeating to localhost/127.0.0.1:42511 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1823148289-172.17.0.2-1733449995270 (Datanode Uuid 1ea54db9-a065-44f6-9d1e-08a154ee24c7) service to localhost/127.0.0.1:42511 2024-12-06T01:53:24,248 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T01:53:24,248 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T01:53:24,249 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data1/current/BP-1823148289-172.17.0.2-1733449995270 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:24,249 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/cluster_ae096910-94f6-6192-e7a3-175aad41c42e/data/data2/current/BP-1823148289-172.17.0.2-1733449995270 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:24,249 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T01:53:24,256 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T01:53:24,256 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T01:53:24,258 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@76e4c45c{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T01:53:24,259 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4637aff6{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:24,259 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:24,259 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@383d55e4{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:24,259 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21b7d177{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:24,268 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-06T01:53:24,296 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-06T01:53:24,305 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[0] Thread=88 (was 158), OpenFileDescriptor=439 (was 391) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=337 (was 323) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=9910 (was 10204) 2024-12-06T01:53:24,312 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=88, OpenFileDescriptor=439, MaxFileDescriptor=1048576, SystemLoadAverage=337, ProcessCount=11, AvailableMemoryMB=9910 2024-12-06T01:53:24,313 INFO [Time-limited test {}] hbase.HBaseTestingUtil(805): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-06T01:53:24,313 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.log.dir so I do NOT create it in target/test-data/243d5e56-0d61-8f83-caa4-734437509e15 2024-12-06T01:53:24,313 INFO [Time-limited test {}] hbase.HBaseTestingUtil(401): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/1a89f218-e24a-516a-af2a-36f42be76338/hadoop.tmp.dir so I do NOT create it in target/test-data/243d5e56-0d61-8f83-caa4-734437509e15 2024-12-06T01:53:24,313 INFO [Time-limited test {}] hbase.HBaseZKTestingUtil(84): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9, deleteOnExit=true 2024-12-06T01:53:24,313 INFO [Time-limited test {}] hbase.HBaseTestingUtil(818): STARTING DFS 2024-12-06T01:53:24,313 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/test.cache.data in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.tmp.dir in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(738): read short circuit is OFF 2024-12-06T01:53:24,314 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-06T01:53:24,314 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/nfs.dump.dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/java.io.tmpdir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-06T01:53:24,315 INFO [Time-limited test {}] hbase.HBaseTestingUtil(751): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-06T01:53:24,407 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:24,412 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:24,413 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:24,413 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:24,413 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T01:53:24,414 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:24,414 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2baf02db{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:24,415 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@18ee857{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:24,531 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1a4689e1{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/java.io.tmpdir/jetty-localhost-45663-hadoop-hdfs-3_4_1-tests_jar-_-any-14091258160001852131/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T01:53:24,532 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@53c85678{HTTP/1.1, (http/1.1)}{localhost:45663} 2024-12-06T01:53:24,532 INFO [Time-limited test {}] server.Server(415): Started @11421ms 2024-12-06T01:53:24,633 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:24,637 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:24,638 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:24,638 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:24,638 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-06T01:53:24,639 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1841cfc0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:24,639 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6d1c78c1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:24,758 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@760f4a1c{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/java.io.tmpdir/jetty-localhost-45689-hadoop-hdfs-3_4_1-tests_jar-_-any-6134627259415476660/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:24,759 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@e494f88{HTTP/1.1, (http/1.1)}{localhost:45689} 2024-12-06T01:53:24,759 INFO [Time-limited test {}] server.Server(415): Started @11648ms 2024-12-06T01:53:24,760 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T01:53:24,798 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:24,802 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:24,803 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:24,803 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:24,803 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-06T01:53:24,803 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@55fefad1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:24,804 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@564d8641{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:24,855 WARN [Thread-520 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data1/current/BP-803844786-172.17.0.2-1733450004348/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:24,855 WARN [Thread-521 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data2/current/BP-803844786-172.17.0.2-1733450004348/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:24,879 WARN [Thread-499 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T01:53:24,882 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x28291564d1594191 with lease ID 0x30de0eeb23083e2: Processing first storage report for DS-02b778df-1b83-4552-9385-cd8b1ba3b47a from datanode DatanodeRegistration(127.0.0.1:45079, datanodeUuid=f0ebc5d8-587b-460d-8288-27f072d5ac52, infoPort=42271, infoSecurePort=0, ipcPort=38017, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348) 2024-12-06T01:53:24,882 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x28291564d1594191 with lease ID 0x30de0eeb23083e2: from storage DS-02b778df-1b83-4552-9385-cd8b1ba3b47a node DatanodeRegistration(127.0.0.1:45079, datanodeUuid=f0ebc5d8-587b-460d-8288-27f072d5ac52, infoPort=42271, infoSecurePort=0, ipcPort=38017, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:24,883 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x28291564d1594191 with lease ID 0x30de0eeb23083e2: Processing first storage report for DS-62490638-b4d6-4fcc-80c8-18cdfbfc5b63 from datanode DatanodeRegistration(127.0.0.1:45079, datanodeUuid=f0ebc5d8-587b-460d-8288-27f072d5ac52, infoPort=42271, infoSecurePort=0, ipcPort=38017, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348) 2024-12-06T01:53:24,883 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x28291564d1594191 with lease ID 0x30de0eeb23083e2: from storage DS-62490638-b4d6-4fcc-80c8-18cdfbfc5b63 node DatanodeRegistration(127.0.0.1:45079, datanodeUuid=f0ebc5d8-587b-460d-8288-27f072d5ac52, infoPort=42271, infoSecurePort=0, ipcPort=38017, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:24,928 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@25106b03{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/java.io.tmpdir/jetty-localhost-40841-hadoop-hdfs-3_4_1-tests_jar-_-any-17619864595771299915/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:24,928 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@ae51624{HTTP/1.1, (http/1.1)}{localhost:40841} 2024-12-06T01:53:24,928 INFO [Time-limited test {}] server.Server(415): Started @11817ms 2024-12-06T01:53:24,930 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T01:53:24,962 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-06T01:53:24,966 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-06T01:53:24,968 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-06T01:53:24,968 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-06T01:53:24,968 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-06T01:53:24,969 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@e56045e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,AVAILABLE} 2024-12-06T01:53:24,969 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@328032c9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-06T01:53:25,027 WARN [Thread-556 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data4/current/BP-803844786-172.17.0.2-1733450004348/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:25,027 WARN [Thread-555 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data3/current/BP-803844786-172.17.0.2-1733450004348/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:25,051 WARN [Thread-535 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T01:53:25,055 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3e28365b6403153c with lease ID 0x30de0eeb23083e3: Processing first storage report for DS-6f311a83-aa2d-464d-9872-9080178174fe from datanode DatanodeRegistration(127.0.0.1:33671, datanodeUuid=1d7909ec-71be-4281-bec0-349cd85878e2, infoPort=45017, infoSecurePort=0, ipcPort=42869, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348) 2024-12-06T01:53:25,055 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3e28365b6403153c with lease ID 0x30de0eeb23083e3: from storage DS-6f311a83-aa2d-464d-9872-9080178174fe node DatanodeRegistration(127.0.0.1:33671, datanodeUuid=1d7909ec-71be-4281-bec0-349cd85878e2, infoPort=45017, infoSecurePort=0, ipcPort=42869, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:25,055 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x3e28365b6403153c with lease ID 0x30de0eeb23083e3: Processing first storage report for DS-e52022fb-ba4d-4cb8-a3ae-42c57acedded from datanode DatanodeRegistration(127.0.0.1:33671, datanodeUuid=1d7909ec-71be-4281-bec0-349cd85878e2, infoPort=45017, infoSecurePort=0, ipcPort=42869, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348) 2024-12-06T01:53:25,055 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x3e28365b6403153c with lease ID 0x30de0eeb23083e3: from storage DS-e52022fb-ba4d-4cb8-a3ae-42c57acedded node DatanodeRegistration(127.0.0.1:33671, datanodeUuid=1d7909ec-71be-4281-bec0-349cd85878e2, infoPort=45017, infoSecurePort=0, ipcPort=42869, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:25,102 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6ff8a873{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/java.io.tmpdir/jetty-localhost-42809-hadoop-hdfs-3_4_1-tests_jar-_-any-9107551762144426106/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:25,103 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1ec581d{HTTP/1.1, (http/1.1)}{localhost:42809} 2024-12-06T01:53:25,103 INFO [Time-limited test {}] server.Server(415): Started @11992ms 2024-12-06T01:53:25,105 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-06T01:53:25,196 WARN [Thread-582 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data6/current/BP-803844786-172.17.0.2-1733450004348/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:25,196 WARN [Thread-581 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data5/current/BP-803844786-172.17.0.2-1733450004348/current, will proceed with Du for space computation calculation, 2024-12-06T01:53:25,224 WARN [Thread-570 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-06T01:53:25,228 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2b57570bcf92706d with lease ID 0x30de0eeb23083e4: Processing first storage report for DS-641cd887-09b4-46f8-a074-73c27b07bab1 from datanode DatanodeRegistration(127.0.0.1:45831, datanodeUuid=3fa84b4d-de66-4c7c-9324-c0d5ea5bf184, infoPort=45427, infoSecurePort=0, ipcPort=41927, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348) 2024-12-06T01:53:25,228 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2b57570bcf92706d with lease ID 0x30de0eeb23083e4: from storage DS-641cd887-09b4-46f8-a074-73c27b07bab1 node DatanodeRegistration(127.0.0.1:45831, datanodeUuid=3fa84b4d-de66-4c7c-9324-c0d5ea5bf184, infoPort=45427, infoSecurePort=0, ipcPort=41927, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:25,228 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2b57570bcf92706d with lease ID 0x30de0eeb23083e4: Processing first storage report for DS-f30fb7f2-2b97-4c61-9b14-764379232bcd from datanode DatanodeRegistration(127.0.0.1:45831, datanodeUuid=3fa84b4d-de66-4c7c-9324-c0d5ea5bf184, infoPort=45427, infoSecurePort=0, ipcPort=41927, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348) 2024-12-06T01:53:25,228 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2b57570bcf92706d with lease ID 0x30de0eeb23083e4: from storage DS-f30fb7f2-2b97-4c61-9b14-764379232bcd node DatanodeRegistration(127.0.0.1:45831, datanodeUuid=3fa84b4d-de66-4c7c-9324-c0d5ea5bf184, infoPort=45427, infoSecurePort=0, ipcPort=41927, storageInfo=lv=-57;cid=testClusterID;nsid=32892756;c=1733450004348), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-06T01:53:25,241 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(631): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15 2024-12-06T01:53:25,244 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(261): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/zookeeper_0, clientPort=57244, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-06T01:53:25,245 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(286): Started MiniZooKeeperCluster and ran 'stat' on client port=57244 2024-12-06T01:53:25,246 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,248 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741825_1001 (size=7) 2024-12-06T01:53:25,264 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741825_1001 (size=7) 2024-12-06T01:53:25,264 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741825_1001 (size=7) 2024-12-06T01:53:25,266 INFO [Time-limited test {}] util.FSUtils(489): Created version file at hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c with version=8 2024-12-06T01:53:25,266 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1141): The hbase.fs.tmp.dir is set to hdfs://localhost:42511/user/jenkins/test-data/c2aebf7b-aa67-2d81-1023-c44963cb8c72/hbase-staging 2024-12-06T01:53:25,268 INFO [Time-limited test {}] client.ConnectionUtils(128): master/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:25,268 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,268 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,268 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:25,268 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,268 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:25,268 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.AdminService 2024-12-06T01:53:25,269 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:25,269 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:32985 2024-12-06T01:53:25,271 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=master:32985 connecting to ZooKeeper ensemble=127.0.0.1:57244 2024-12-06T01:53:25,275 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:329850x0, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:25,276 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:32985-0x1006507b3e20000 connected 2024-12-06T01:53:25,295 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,296 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,299 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:25,299 INFO [Time-limited test {}] master.HMaster(525): hbase.rootdir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c, hbase.cluster.distributed=false 2024-12-06T01:53:25,301 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:25,301 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=32985 2024-12-06T01:53:25,301 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=32985 2024-12-06T01:53:25,302 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=32985 2024-12-06T01:53:25,302 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=32985 2024-12-06T01:53:25,302 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=32985 2024-12-06T01:53:25,317 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:25,317 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,317 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,318 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:25,318 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,318 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:25,318 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T01:53:25,318 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:25,318 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39091 2024-12-06T01:53:25,320 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39091 connecting to ZooKeeper ensemble=127.0.0.1:57244 2024-12-06T01:53:25,321 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,323 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,327 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:390910x0, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:25,328 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:25,327 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39091-0x1006507b3e20001 connected 2024-12-06T01:53:25,328 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T01:53:25,328 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T01:53:25,329 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T01:53:25,330 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:25,331 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39091 2024-12-06T01:53:25,331 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39091 2024-12-06T01:53:25,331 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39091 2024-12-06T01:53:25,332 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39091 2024-12-06T01:53:25,332 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39091 2024-12-06T01:53:25,348 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T01:53:25,348 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:25,349 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:39077 2024-12-06T01:53:25,350 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:39077 connecting to ZooKeeper ensemble=127.0.0.1:57244 2024-12-06T01:53:25,350 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,352 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:390770x0, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:25,357 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39077-0x1006507b3e20002 connected 2024-12-06T01:53:25,357 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:25,358 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T01:53:25,358 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T01:53:25,359 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T01:53:25,360 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:25,360 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39077 2024-12-06T01:53:25,361 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39077 2024-12-06T01:53:25,361 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39077 2024-12-06T01:53:25,361 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39077 2024-12-06T01:53:25,361 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39077 2024-12-06T01:53:25,376 INFO [Time-limited test {}] client.ConnectionUtils(128): regionserver/00c764f7c3aa:0 server-side Connection retries=45 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-06T01:53:25,377 INFO [Time-limited test {}] ipc.NettyRpcServer(309): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-06T01:53:25,378 INFO [Time-limited test {}] ipc.NettyRpcServer(191): Bind to /172.17.0.2:43335 2024-12-06T01:53:25,379 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(137): Process identifier=regionserver:43335 connecting to ZooKeeper ensemble=127.0.0.1:57244 2024-12-06T01:53:25,380 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,381 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,385 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:433350x0, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-06T01:53:25,386 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43335-0x1006507b3e20003 connected 2024-12-06T01:53:25,386 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:25,386 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-06T01:53:25,387 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-06T01:53:25,388 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-06T01:53:25,389 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-06T01:53:25,390 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43335 2024-12-06T01:53:25,390 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43335 2024-12-06T01:53:25,390 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43335 2024-12-06T01:53:25,391 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43335 2024-12-06T01:53:25,391 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43335 2024-12-06T01:53:25,403 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;00c764f7c3aa:32985 2024-12-06T01:53:25,403 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(2510): Adding backup master ZNode /hbase/backup-masters/00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,405 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,405 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,405 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,405 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,406 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:25,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,410 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-06T01:53:25,410 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/00c764f7c3aa,32985,1733450005268 from backup master directory 2024-12-06T01:53:25,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,412 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-06T01:53:25,412 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:25,412 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,419 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSUtils(620): Create cluster ID file [hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/hbase.id] with ID: 0f55070f-63d8-461c-bbf6-4c4ed6d2794d 2024-12-06T01:53:25,419 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSUtils(625): Write the cluster ID file to a temporary location: hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/.tmp/hbase.id 2024-12-06T01:53:25,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741826_1002 (size=42) 2024-12-06T01:53:25,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741826_1002 (size=42) 2024-12-06T01:53:25,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741826_1002 (size=42) 2024-12-06T01:53:25,429 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSUtils(634): Move the temporary cluster ID file to its target location [hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/.tmp/hbase.id]:[hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/hbase.id] 2024-12-06T01:53:25,446 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-06T01:53:25,446 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSTableDescriptors(270): Fetching table descriptors from the filesystem. 2024-12-06T01:53:25,447 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] util.FSTableDescriptors(299): Fetched table descriptors(size=0) cost 1ms. 2024-12-06T01:53:25,449 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,449 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,450 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,460 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741827_1003 (size=196) 2024-12-06T01:53:25,460 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741827_1003 (size=196) 2024-12-06T01:53:25,460 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741827_1003 (size=196) 2024-12-06T01:53:25,461 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegion(370): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T01:53:25,462 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-06T01:53:25,463 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T01:53:25,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741828_1004 (size=1189) 2024-12-06T01:53:25,474 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741828_1004 (size=1189) 2024-12-06T01:53:25,474 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741828_1004 (size=1189) 2024-12-06T01:53:25,475 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7590): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store 2024-12-06T01:53:25,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741829_1005 (size=34) 2024-12-06T01:53:25,486 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741829_1005 (size=34) 2024-12-06T01:53:25,487 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741829_1005 (size=34) 2024-12-06T01:53:25,487 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:25,487 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T01:53:25,488 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:25,488 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:25,488 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T01:53:25,488 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:25,488 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:25,488 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733450005487Disabling compacts and flushes for region at 1733450005487Disabling writes for close at 1733450005488 (+1 ms)Writing region close event to WAL at 1733450005488Closed at 1733450005488 2024-12-06T01:53:25,489 WARN [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/.initializing 2024-12-06T01:53:25,489 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/WALs/00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,494 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C32985%2C1733450005268, suffix=, logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/WALs/00c764f7c3aa,32985,1733450005268, archiveDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/oldWALs, maxLogs=10 2024-12-06T01:53:25,494 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c764f7c3aa%2C32985%2C1733450005268.1733450005494 2024-12-06T01:53:25,504 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/WALs/00c764f7c3aa,32985,1733450005268/00c764f7c3aa%2C32985%2C1733450005268.1733450005494 2024-12-06T01:53:25,506 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45427:45427),(127.0.0.1/127.0.0.1:42271:42271),(127.0.0.1/127.0.0.1:45017:45017)] 2024-12-06T01:53:25,507 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7752): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-06T01:53:25,507 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(898): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:25,507 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7794): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,507 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(7797): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,509 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,512 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-06T01:53:25,512 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,512 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:25,513 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,514 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-06T01:53:25,515 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,515 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:25,515 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,518 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-06T01:53:25,518 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,519 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:25,519 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,520 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-06T01:53:25,520 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,521 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:25,521 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1038): replaying wal for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,522 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,522 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,524 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1048): stopping wal replay for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,524 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1060): Cleaning up temporary data for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,525 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T01:53:25,526 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1093): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-06T01:53:25,529 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T01:53:25,530 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1114): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73848888, jitterRate=0.10043418407440186}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T01:53:25,530 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] regionserver.HRegion(1006): Region open journal for 1595e783b53d99cd5eef43b6debb2682: Writing region info on filesystem at 1733450005508Initializing all the Stores at 1733450005509 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450005509Instantiating store for column family {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450005509Instantiating store for column family {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450005509Instantiating store for column family {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450005509Cleaning up temporary data from old regions at 1733450005524 (+15 ms)Region opened successfully at 1733450005530 (+6 ms) 2024-12-06T01:53:25,531 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-06T01:53:25,535 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4d33bdbb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:25,536 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(912): No meta location available on zookeeper, skip migrating... 2024-12-06T01:53:25,536 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-06T01:53:25,536 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(626): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-06T01:53:25,536 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-06T01:53:25,537 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(676): Recovered RegionProcedureStore lease in 0 msec 2024-12-06T01:53:25,537 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(690): Loaded RegionProcedureStore in 0 msec 2024-12-06T01:53:25,537 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-06T01:53:25,540 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-06T01:53:25,541 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-06T01:53:25,542 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/balancer already deleted, retry=false 2024-12-06T01:53:25,543 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-06T01:53:25,543 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-06T01:53:25,545 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/normalizer already deleted, retry=false 2024-12-06T01:53:25,545 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-06T01:53:25,546 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-06T01:53:25,547 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/split already deleted, retry=false 2024-12-06T01:53:25,548 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-06T01:53:25,549 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/switch/merge already deleted, retry=false 2024-12-06T01:53:25,551 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-06T01:53:25,553 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,555 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,556 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(856): Active/primary master=00c764f7c3aa,32985,1733450005268, sessionid=0x1006507b3e20000, setting cluster-up flag (Was=false) 2024-12-06T01:53:25,559 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,559 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,559 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,559 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,566 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-06T01:53:25,567 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,570 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,570 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,570 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,570 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,576 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-06T01:53:25,577 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:25,578 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ServerManager(1185): No .lastflushedseqids found at hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/.lastflushedseqids will record last flushed sequence id for regions by regionserver report all over again 2024-12-06T01:53:25,581 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1139): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=false; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:25,581 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(416): slop=0.2 2024-12-06T01:53:25,581 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(272): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, CPRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-06T01:53:25,582 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] balancer.RegionHDFSBlockLocationFinder(133): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 00c764f7c3aa,32985,1733450005268 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=5, maxPoolSize=5 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/00c764f7c3aa:0, corePoolSize=10, maxPoolSize=10 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:25,583 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,585 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733450035585 2024-12-06T01:53:25,585 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-06T01:53:25,585 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-06T01:53:25,585 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-06T01:53:25,586 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-06T01:53:25,586 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-06T01:53:25,586 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-06T01:53:25,586 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,586 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:25,586 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(76): BOOTSTRAP: creating hbase:meta region 2024-12-06T01:53:25,586 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-06T01:53:25,586 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-06T01:53:25,587 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-06T01:53:25,587 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-06T01:53:25,587 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-06T01:53:25,587 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.large.0-1733450005587,5,FailOnTimeoutGroup] 2024-12-06T01:53:25,587 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.small.0-1733450005587,5,FailOnTimeoutGroup] 2024-12-06T01:53:25,587 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,587 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1741): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-06T01:53:25,588 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,588 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,588 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,588 INFO [PEWorker-1 {}] util.FSTableDescriptors(156): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-06T01:53:25,594 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(746): ClusterId : 0f55070f-63d8-461c-bbf6-4c4ed6d2794d 2024-12-06T01:53:25,594 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T01:53:25,595 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(746): ClusterId : 0f55070f-63d8-461c-bbf6-4c4ed6d2794d 2024-12-06T01:53:25,595 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T01:53:25,598 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(746): ClusterId : 0f55070f-63d8-461c-bbf6-4c4ed6d2794d 2024-12-06T01:53:25,598 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-06T01:53:25,598 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T01:53:25,598 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T01:53:25,599 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T01:53:25,599 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T01:53:25,602 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-06T01:53:25,602 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-06T01:53:25,602 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T01:53:25,602 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T01:53:25,603 DEBUG [RS:1;00c764f7c3aa:39077 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5ec068e8, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:25,603 DEBUG [RS:0;00c764f7c3aa:39091 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@69e99680, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:25,604 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-06T01:53:25,605 DEBUG [RS:2;00c764f7c3aa:43335 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@37c5dc2e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=00c764f7c3aa/172.17.0.2:0 2024-12-06T01:53:25,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741831_1007 (size=1321) 2024-12-06T01:53:25,613 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741831_1007 (size=1321) 2024-12-06T01:53:25,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741831_1007 (size=1321) 2024-12-06T01:53:25,615 INFO [PEWorker-1 {}] util.FSTableDescriptors(163): Updated hbase:meta table descriptor to hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1321 2024-12-06T01:53:25,615 INFO [PEWorker-1 {}] regionserver.HRegion(7572): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c 2024-12-06T01:53:25,623 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;00c764f7c3aa:43335 2024-12-06T01:53:25,623 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T01:53:25,623 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T01:53:25,623 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T01:53:25,624 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c764f7c3aa,32985,1733450005268 with port=43335, startcode=1733450005376 2024-12-06T01:53:25,624 DEBUG [RS:2;00c764f7c3aa:43335 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T01:53:25,627 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;00c764f7c3aa:39091 2024-12-06T01:53:25,627 DEBUG [RS:1;00c764f7c3aa:39077 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;00c764f7c3aa:39077 2024-12-06T01:53:25,627 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T01:53:25,627 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2024-12-06T01:53:25,627 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T01:53:25,627 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2024-12-06T01:53:25,627 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T01:53:25,627 DEBUG [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(832): About to register with Master. 2024-12-06T01:53:25,628 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c764f7c3aa,32985,1733450005268 with port=39091, startcode=1733450005317 2024-12-06T01:53:25,628 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(2659): reportForDuty to master=00c764f7c3aa,32985,1733450005268 with port=39077, startcode=1733450005347 2024-12-06T01:53:25,628 DEBUG [RS:1;00c764f7c3aa:39077 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T01:53:25,630 DEBUG [RS:0;00c764f7c3aa:39091 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-06T01:53:25,630 INFO [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43487, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T01:53:25,631 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=32985 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,632 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=32985 {}] master.ServerManager(517): Registering regionserver=00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,632 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49275, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T01:53:25,632 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60223, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-06T01:53:25,634 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=32985 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:25,634 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=32985 {}] master.ServerManager(517): Registering regionserver=00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:25,634 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c 2024-12-06T01:53:25,635 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40977 2024-12-06T01:53:25,635 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T01:53:25,636 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=32985 {}] master.ServerManager(363): Checking decommissioned status of RegionServer 00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:25,636 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=32985 {}] master.ServerManager(517): Registering regionserver=00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:25,636 DEBUG [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c 2024-12-06T01:53:25,636 DEBUG [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40977 2024-12-06T01:53:25,637 DEBUG [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T01:53:25,637 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:25,637 DEBUG [RS:2;00c764f7c3aa:43335 {}] zookeeper.ZKUtil(111): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,637 WARN [RS:2;00c764f7c3aa:43335 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:25,638 INFO [RS:2;00c764f7c3aa:43335 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T01:53:25,638 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,639 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1440): Config from master: hbase.rootdir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c 2024-12-06T01:53:25,639 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1440): Config from master: fs.defaultFS=hdfs://localhost:40977 2024-12-06T01:53:25,639 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1440): Config from master: hbase.master.info.port=-1 2024-12-06T01:53:25,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741832_1008 (size=32) 2024-12-06T01:53:25,642 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c764f7c3aa,39077,1733450005347] 2024-12-06T01:53:25,643 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c764f7c3aa,43335,1733450005376] 2024-12-06T01:53:25,643 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:25,643 DEBUG [RS:1;00c764f7c3aa:39077 {}] zookeeper.ZKUtil(111): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:25,643 WARN [RS:1;00c764f7c3aa:39077 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:25,643 INFO [RS:1;00c764f7c3aa:39077 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T01:53:25,643 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741832_1008 (size=32) 2024-12-06T01:53:25,643 DEBUG [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:25,644 DEBUG [RS:0;00c764f7c3aa:39091 {}] zookeeper.ZKUtil(111): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:25,644 WARN [RS:0;00c764f7c3aa:39091 {}] hbase.ZNodeClearer(65): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-06T01:53:25,644 INFO [RS:0;00c764f7c3aa:39091 {}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T01:53:25,644 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [00c764f7c3aa,39091,1733450005317] 2024-12-06T01:53:25,644 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1793): logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:25,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741832_1008 (size=32) 2024-12-06T01:53:25,645 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T01:53:25,645 DEBUG [PEWorker-1 {}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:25,648 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T01:53:25,648 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T01:53:25,649 INFO [RS:2;00c764f7c3aa:43335 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T01:53:25,649 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,650 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T01:53:25,650 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T01:53:25,650 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,651 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T01:53:25,651 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,651 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,651 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,651 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,651 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,652 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:25,653 DEBUG [RS:2;00c764f7c3aa:43335 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:25,653 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,653 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,654 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,654 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,654 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,654 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,43335,1733450005376-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:25,655 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-06T01:53:25,655 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T01:53:25,656 INFO [RS:1;00c764f7c3aa:39077 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T01:53:25,656 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,656 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T01:53:25,656 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,657 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:25,657 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T01:53:25,659 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T01:53:25,660 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,660 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:25,660 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T01:53:25,662 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T01:53:25,663 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:25,663 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:25,664 DEBUG [PEWorker-1 {}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T01:53:25,664 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T01:53:25,665 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.MemStoreFlusher(131): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-06T01:53:25,665 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740 2024-12-06T01:53:25,665 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T01:53:25,665 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,665 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,665 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:25,666 DEBUG [RS:1;00c764f7c3aa:39077 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:25,667 DEBUG [PEWorker-1 {}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740 2024-12-06T01:53:25,668 INFO [RS:0;00c764f7c3aa:39091 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-06T01:53:25,668 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,669 DEBUG [PEWorker-1 {}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T01:53:25,669 DEBUG [PEWorker-1 {}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T01:53:25,670 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T01:53:25,671 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer$CompactionChecker(1680): CompactionChecker runs every PT1S 2024-12-06T01:53:25,672 DEBUG [PEWorker-1 {}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T01:53:25,672 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,672 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,672 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,672 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,672 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,673 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ExecutorStatusChore(48): ExecutorStatusChore runs every 1mins, 0sec 2024-12-06T01:53:25,673 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,39077,1733450005347-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:25,673 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,673 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,673 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,673 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,673 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,673 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,673 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/00c764f7c3aa:0, corePoolSize=2, maxPoolSize=2 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_REPLAY_SYNC_REPLICATION_WAL-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/00c764f7c3aa:0, corePoolSize=1, maxPoolSize=1 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:25,674 DEBUG [RS:0;00c764f7c3aa:39091 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0, corePoolSize=3, maxPoolSize=3 2024-12-06T01:53:25,677 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=ExecutorStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T01:53:25,677 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T01:53:25,677 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,39091,1733450005317-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,43335,1733450005376-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,677 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.Replication(171): 00c764f7c3aa,43335,1733450005376 started 2024-12-06T01:53:25,678 INFO [PEWorker-1 {}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74950001, jitterRate=0.11684204638004303}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T01:53:25,679 DEBUG [PEWorker-1 {}] regionserver.HRegion(1006): Region open journal for 1588230740: Writing region info on filesystem at 1733450005646Initializing all the Stores at 1733450005647 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450005647Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450005648 (+1 ms)Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450005648Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450005648Cleaning up temporary data from old regions at 1733450005669 (+21 ms)Region opened successfully at 1733450005679 (+10 ms) 2024-12-06T01:53:25,680 DEBUG [PEWorker-1 {}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T01:53:25,680 INFO [PEWorker-1 {}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T01:53:25,680 DEBUG [PEWorker-1 {}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T01:53:25,680 DEBUG [PEWorker-1 {}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T01:53:25,680 DEBUG [PEWorker-1 {}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T01:53:25,684 INFO [PEWorker-1 {}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T01:53:25,684 DEBUG [PEWorker-1 {}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733450005679Disabling compacts and flushes for region at 1733450005679Disabling writes for close at 1733450005680 (+1 ms)Writing region close event to WAL at 1733450005684 (+4 ms)Closed at 1733450005684 2024-12-06T01:53:25,687 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:25,687 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(108): Going to assign meta 2024-12-06T01:53:25,687 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-06T01:53:25,690 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T01:53:25,692 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(269): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-06T01:53:25,695 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T01:53:25,695 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,39077,1733450005347-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,696 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,696 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.Replication(171): 00c764f7c3aa,39077,1733450005347 started 2024-12-06T01:53:25,699 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,699 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1482): Serving as 00c764f7c3aa,43335,1733450005376, RpcServer on 00c764f7c3aa/172.17.0.2:43335, sessionid=0x1006507b3e20003 2024-12-06T01:53:25,699 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T01:53:25,699 DEBUG [RS:2;00c764f7c3aa:43335 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,699 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,43335,1733450005376' 2024-12-06T01:53:25,700 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T01:53:25,701 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T01:53:25,701 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T01:53:25,701 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T01:53:25,701 DEBUG [RS:2;00c764f7c3aa:43335 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,702 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,43335,1733450005376' 2024-12-06T01:53:25,702 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-06T01:53:25,702 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T01:53:25,702 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,39091,1733450005317-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,702 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,702 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.Replication(171): 00c764f7c3aa,39091,1733450005317 started 2024-12-06T01:53:25,702 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T01:53:25,703 DEBUG [RS:2;00c764f7c3aa:43335 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T01:53:25,703 INFO [RS:2;00c764f7c3aa:43335 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T01:53:25,703 INFO [RS:2;00c764f7c3aa:43335 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T01:53:25,712 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,712 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(1482): Serving as 00c764f7c3aa,39077,1733450005347, RpcServer on 00c764f7c3aa/172.17.0.2:39077, sessionid=0x1006507b3e20002 2024-12-06T01:53:25,712 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T01:53:25,712 DEBUG [RS:1;00c764f7c3aa:39077 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:25,712 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,39077,1733450005347' 2024-12-06T01:53:25,712 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T01:53:25,713 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T01:53:25,713 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T01:53:25,713 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T01:53:25,714 DEBUG [RS:1;00c764f7c3aa:39077 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:25,714 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,39077,1733450005347' 2024-12-06T01:53:25,714 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T01:53:25,714 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T01:53:25,715 DEBUG [RS:1;00c764f7c3aa:39077 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T01:53:25,715 INFO [RS:1;00c764f7c3aa:39077 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T01:53:25,715 INFO [RS:1;00c764f7c3aa:39077 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T01:53:25,722 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:25,722 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1482): Serving as 00c764f7c3aa,39091,1733450005317, RpcServer on 00c764f7c3aa/172.17.0.2:39091, sessionid=0x1006507b3e20001 2024-12-06T01:53:25,722 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-06T01:53:25,723 DEBUG [RS:0;00c764f7c3aa:39091 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:25,723 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,39091,1733450005317' 2024-12-06T01:53:25,723 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-06T01:53:25,723 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-06T01:53:25,724 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-06T01:53:25,724 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-06T01:53:25,724 DEBUG [RS:0;00c764f7c3aa:39091 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:25,724 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '00c764f7c3aa,39091,1733450005317' 2024-12-06T01:53:25,724 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-06T01:53:25,724 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-06T01:53:25,725 DEBUG [RS:0;00c764f7c3aa:39091 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-06T01:53:25,725 INFO [RS:0;00c764f7c3aa:39091 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-06T01:53:25,725 INFO [RS:0;00c764f7c3aa:39091 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-06T01:53:25,806 INFO [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C43335%2C1733450005376, suffix=, logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,43335,1733450005376, archiveDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs, maxLogs=32 2024-12-06T01:53:25,809 INFO [RS:2;00c764f7c3aa:43335 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c764f7c3aa%2C43335%2C1733450005376.1733450005809 2024-12-06T01:53:25,818 INFO [RS:1;00c764f7c3aa:39077 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C39077%2C1733450005347, suffix=, logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,39077,1733450005347, archiveDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs, maxLogs=32 2024-12-06T01:53:25,819 INFO [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,43335,1733450005376/00c764f7c3aa%2C43335%2C1733450005376.1733450005809 2024-12-06T01:53:25,820 DEBUG [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45017:45017),(127.0.0.1/127.0.0.1:42271:42271),(127.0.0.1/127.0.0.1:45427:45427)] 2024-12-06T01:53:25,820 INFO [RS:1;00c764f7c3aa:39077 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c764f7c3aa%2C39077%2C1733450005347.1733450005820 2024-12-06T01:53:25,827 INFO [RS:0;00c764f7c3aa:39091 {}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C39091%2C1733450005317, suffix=, logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,39091,1733450005317, archiveDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs, maxLogs=32 2024-12-06T01:53:25,828 INFO [RS:0;00c764f7c3aa:39091 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c764f7c3aa%2C39091%2C1733450005317.1733450005828 2024-12-06T01:53:25,833 INFO [RS:1;00c764f7c3aa:39077 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,39077,1733450005347/00c764f7c3aa%2C39077%2C1733450005347.1733450005820 2024-12-06T01:53:25,834 DEBUG [RS:1;00c764f7c3aa:39077 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45017:45017),(127.0.0.1/127.0.0.1:45427:45427),(127.0.0.1/127.0.0.1:42271:42271)] 2024-12-06T01:53:25,841 INFO [RS:0;00c764f7c3aa:39091 {}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,39091,1733450005317/00c764f7c3aa%2C39091%2C1733450005317.1733450005828 2024-12-06T01:53:25,842 DEBUG [RS:0;00c764f7c3aa:39091 {}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:42271:42271),(127.0.0.1/127.0.0.1:45017:45017),(127.0.0.1/127.0.0.1:45427:45427)] 2024-12-06T01:53:25,842 DEBUG [00c764f7c3aa:32985 {}] assignment.AssignmentManager(2472): Processing assignQueue; systemServersCount=2, allServersCount=2 2024-12-06T01:53:25,843 DEBUG [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(204): Hosts are {00c764f7c3aa=0} racks are {/default-rack=0} 2024-12-06T01:53:25,845 DEBUG [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T01:53:25,845 DEBUG [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T01:53:25,845 DEBUG [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T01:53:25,845 DEBUG [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T01:53:25,845 INFO [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T01:53:25,845 INFO [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T01:53:25,845 DEBUG [00c764f7c3aa:32985 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T01:53:25,846 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:25,848 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c764f7c3aa,43335,1733450005376, state=OPENING 2024-12-06T01:53:25,850 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-06T01:53:25,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,851 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:25,852 DEBUG [PEWorker-3 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-06T01:53:25,852 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:25,852 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:25,852 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE, hasLock=false; OpenRegionProcedure 1588230740, server=00c764f7c3aa,43335,1733450005376}] 2024-12-06T01:53:25,852 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:25,852 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:26,007 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T01:53:26,009 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50521, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T01:53:26,013 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(132): Open hbase:meta,,1.1588230740 2024-12-06T01:53:26,014 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(196): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-06T01:53:26,017 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(613): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=00c764f7c3aa%2C43335%2C1733450005376.meta, suffix=.meta, logDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,43335,1733450005376, archiveDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs, maxLogs=32 2024-12-06T01:53:26,018 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 00c764f7c3aa%2C43335%2C1733450005376.meta.1733450006018.meta 2024-12-06T01:53:26,028 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(991): New WAL /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/WALs/00c764f7c3aa,43335,1733450005376/00c764f7c3aa%2C43335%2C1733450005376.meta.1733450006018.meta 2024-12-06T01:53:26,029 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(1109): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:42271:42271),(127.0.0.1/127.0.0.1:45427:45427),(127.0.0.1/127.0.0.1:45017:45017)] 2024-12-06T01:53:26,030 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7752): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-06T01:53:26,030 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-06T01:53:26,030 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(8280): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-06T01:53:26,031 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(434): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-06T01:53:26,031 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-06T01:53:26,031 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(898): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:26,031 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7794): checking encryption for 1588230740 2024-12-06T01:53:26,031 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7797): checking classloading for 1588230740 2024-12-06T01:53:26,033 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-06T01:53:26,034 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-06T01:53:26,034 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:26,035 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:26,035 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family ns of region 1588230740 2024-12-06T01:53:26,036 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName ns 2024-12-06T01:53:26,036 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:26,037 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/ns, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:26,037 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-06T01:53:26,038 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-06T01:53:26,038 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:26,038 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:26,039 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-06T01:53:26,039 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-06T01:53:26,039 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:26,040 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-06T01:53:26,040 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1038): replaying wal for 1588230740 2024-12-06T01:53:26,041 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740 2024-12-06T01:53:26,043 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740 2024-12-06T01:53:26,045 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1048): stopping wal replay for 1588230740 2024-12-06T01:53:26,045 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1060): Cleaning up temporary data for 1588230740 2024-12-06T01:53:26,046 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-06T01:53:26,048 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1093): writing seq id for 1588230740 2024-12-06T01:53:26,049 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1114): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67359866, jitterRate=0.0037402212619781494}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-06T01:53:26,049 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 1588230740 2024-12-06T01:53:26,051 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1006): Region open journal for 1588230740: Running coprocessor pre-open hook at 1733450006031Writing region info on filesystem at 1733450006031Initializing all the Stores at 1733450006032 (+1 ms)Instantiating store for column family {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450006033 (+1 ms)Instantiating store for column family {NAME => 'ns', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450006033Instantiating store for column family {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450006033Instantiating store for column family {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} at 1733450006033Cleaning up temporary data from old regions at 1733450006045 (+12 ms)Running coprocessor post-open hooks at 1733450006049 (+4 ms)Region opened successfully at 1733450006050 (+1 ms) 2024-12-06T01:53:26,052 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2236): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733450006007 2024-12-06T01:53:26,056 DEBUG [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2266): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-06T01:53:26,056 INFO [RS_OPEN_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(153): Opened hbase:meta,,1.1588230740 2024-12-06T01:53:26,057 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:26,058 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 00c764f7c3aa,43335,1733450005376, state=OPEN 2024-12-06T01:53:26,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:26,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:26,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:26,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-06T01:53:26,060 DEBUG [PEWorker-5 {}] procedure2.ProcedureFutureUtil(75): The future has completed while adding callback, give up suspending procedure pid=3, ppid=2, state=RUNNABLE, hasLock=true; OpenRegionProcedure 1588230740, server=00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:26,060 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:26,060 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:26,060 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:26,060 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-06T01:53:26,065 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=3, resume processing ppid=2 2024-12-06T01:53:26,065 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=3, ppid=2, state=SUCCESS, hasLock=false; OpenRegionProcedure 1588230740, server=00c764f7c3aa,43335,1733450005376 in 208 msec 2024-12-06T01:53:26,070 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=2, resume processing ppid=1 2024-12-06T01:53:26,070 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=2, ppid=1, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 378 msec 2024-12-06T01:53:26,072 DEBUG [PEWorker-2 {}] procedure.InitMetaProcedure(97): Execute pid=1, state=RUNNABLE:INIT_META_CREATE_NAMESPACES, hasLock=true; InitMetaProcedure table=hbase:meta 2024-12-06T01:53:26,072 INFO [PEWorker-2 {}] procedure.InitMetaProcedure(114): Going to create {NAME => 'default'} and {NAME => 'hbase'} namespaces 2024-12-06T01:53:26,074 DEBUG [PEWorker-2 {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T01:53:26,074 DEBUG [PEWorker-2 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c764f7c3aa,43335,1733450005376, seqNum=-1] 2024-12-06T01:53:26,075 DEBUG [PEWorker-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T01:53:26,077 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34237, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T01:53:26,086 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1521): Finished pid=1, state=SUCCESS, hasLock=false; InitMetaProcedure table=hbase:meta in 503 msec 2024-12-06T01:53:26,086 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1123): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733450006086, completionTime=-1 2024-12-06T01:53:26,086 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ServerManager(903): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2024-12-06T01:53:26,086 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] assignment.AssignmentManager(1764): Joining cluster... 2024-12-06T01:53:26,088 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] assignment.AssignmentManager(1776): Number of RegionServers=3 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733450066088 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(82): ADDED pid=-1, state=WAITING_TIMEOUT, hasLock=false; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733450126089 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] assignment.AssignmentManager(1783): Joined the cluster in 2 msec 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,32985,1733450005268-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,32985,1733450005268-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,32985,1733450005268-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-00c764f7c3aa:32985, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,089 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,091 DEBUG [master/00c764f7c3aa:0.Chore.1 {}] janitor.CatalogJanitor(180): 2024-12-06T01:53:26,094 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1239): Master has completed initialization 0.681sec 2024-12-06T01:53:26,094 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-06T01:53:26,095 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-06T01:53:26,095 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-06T01:53:26,095 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-06T01:53:26,095 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-06T01:53:26,095 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,32985,1733450005268-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-06T01:53:26,095 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,32985,1733450005268-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-06T01:53:26,098 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4d5d2159, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T01:53:26,098 DEBUG [Time-limited test {}] client.ClusterIdFetcher(90): Going to request 00c764f7c3aa,32985,-1 for getting cluster id 2024-12-06T01:53:26,098 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ConnectionRegistryService, sasl=false 2024-12-06T01:53:26,098 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster {}] master.HMaster(1374): Balancer post startup initialization complete, took 0 seconds 2024-12-06T01:53:26,098 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-06T01:53:26,098 INFO [master/00c764f7c3aa:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=00c764f7c3aa,32985,1733450005268-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-06T01:53:26,099 DEBUG [HMaster-EventLoopGroup-7-2 {}] ipc.ServerRpcConnection(714): Response connection registry, clusterId = '0f55070f-63d8-461c-bbf6-4c4ed6d2794d' 2024-12-06T01:53:26,100 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(424): process preamble call response with response type GetConnectionRegistryResponse 2024-12-06T01:53:26,100 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ClusterIdFetcher$1(103): Got connection registry info: cluster_id: "0f55070f-63d8-461c-bbf6-4c4ed6d2794d" 2024-12-06T01:53:26,100 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@17444283, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T01:53:26,100 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] client.ConnectionRegistryRpcStubHolder(93): Going to use new servers to create stubs: [00c764f7c3aa,32985,-1] 2024-12-06T01:53:26,100 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientMetaService, sasl=false 2024-12-06T01:53:26,101 DEBUG [RPCClient-NioEventLoopGroup-6-6 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:26,102 INFO [HMaster-EventLoopGroup-7-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40994, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientMetaService 2024-12-06T01:53:26,103 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1f519320, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-06T01:53:26,104 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching meta region location from registry 2024-12-06T01:53:26,105 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched meta region location is [region=hbase:meta,,1.1588230740, hostname=00c764f7c3aa,43335,1733450005376, seqNum=-1] 2024-12-06T01:53:26,105 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T01:53:26,107 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60792, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T01:53:26,110 INFO [Time-limited test {}] hbase.HBaseTestingUtil(877): Minicluster is up; activeMaster=00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:26,111 DEBUG [Time-limited test {}] client.ConnectionUtils(547): Start fetching master stub from registry 2024-12-06T01:53:26,112 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.AsyncConnectionImpl(321): The fetched master address is 00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:26,112 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] client.ConnectionUtils(555): The fetched master stub is org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$MasterService$Stub@49e01724 2024-12-06T01:53:26,112 DEBUG [RPCClient-NioEventLoopGroup-6-7 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-06T01:53:26,114 INFO [HMaster-EventLoopGroup-7-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41000, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-06T01:53:26,115 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.HMaster$4(2454): Client=jenkins//172.17.0.2 create 'TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1'}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-06T01:53:26,117 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] procedure2.ProcedureExecutor(1139): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC 2024-12-06T01:53:26,119 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_PRE_OPERATION 2024-12-06T01:53:26,119 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:26,119 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(787): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestHBaseWalOnEC" procId is: 4 2024-12-06T01:53:26,121 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:26,121 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-06T01:53:26,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741837_1013 (size=392) 2024-12-06T01:53:26,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741837_1013 (size=392) 2024-12-06T01:53:26,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741837_1013 (size=392) 2024-12-06T01:53:26,138 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(7572): creating {ENCODED => 95ba767f4b30c4ffbe0b46e001145a71, NAME => 'TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestHBaseWalOnEC', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c 2024-12-06T01:53:26,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741838_1014 (size=51) 2024-12-06T01:53:26,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741838_1014 (size=51) 2024-12-06T01:53:26,148 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741838_1014 (size=51) 2024-12-06T01:53:26,149 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:26,149 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1722): Closing 95ba767f4b30c4ffbe0b46e001145a71, disabling compactions & flushes 2024-12-06T01:53:26,149 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,150 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,150 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. after waiting 0 ms 2024-12-06T01:53:26,150 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,150 INFO [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,150 DEBUG [RegionOpenAndInit-TestHBaseWalOnEC-pool-0 {}] regionserver.HRegion(1676): Region close journal for 95ba767f4b30c4ffbe0b46e001145a71: Waiting for close lock at 1733450006149Disabling compacts and flushes for region at 1733450006149Disabling writes for close at 1733450006150 (+1 ms)Writing region close event to WAL at 1733450006150Closed at 1733450006150 2024-12-06T01:53:26,152 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ADD_TO_META 2024-12-06T01:53:26,152 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":2,"row":"TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71.","families":{"info":[{"qualifier":"regioninfo","vlen":50,"tag":[],"timestamp":"1733450006152"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733450006152"}]},"ts":"1733450006152"} 2024-12-06T01:53:26,156 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(832): Added 1 regions to meta. 2024-12-06T01:53:26,158 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-06T01:53:26,158 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733450006158"}]},"ts":"1733450006158"} 2024-12-06T01:53:26,161 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLING in hbase:meta 2024-12-06T01:53:26,162 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(204): Hosts are {00c764f7c3aa=0} racks are {/default-rack=0} 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 has 0 regions 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 has 0 regions 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 has 0 regions 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 0 is on host 0 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 1 is on host 0 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(310): server 2 is on host 0 2024-12-06T01:53:26,163 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 0 is on rack 0 2024-12-06T01:53:26,163 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 1 is on rack 0 2024-12-06T01:53:26,163 INFO [PEWorker-3 {}] balancer.BalancerClusterState(321): server 2 is on rack 0 2024-12-06T01:53:26,163 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(326): Number of tables=1, number of hosts=1, number of racks=1 2024-12-06T01:53:26,163 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95ba767f4b30c4ffbe0b46e001145a71, ASSIGN}] 2024-12-06T01:53:26,165 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(851): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95ba767f4b30c4ffbe0b46e001145a71, ASSIGN 2024-12-06T01:53:26,167 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(269): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95ba767f4b30c4ffbe0b46e001145a71, ASSIGN; state=OFFLINE, location=00c764f7c3aa,39091,1733450005317; forceNewPlan=false, retain=false 2024-12-06T01:53:26,233 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:26,318 INFO [00c764f7c3aa:32985 {}] balancer.BaseLoadBalancer(388): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-06T01:53:26,318 INFO [PEWorker-5 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=95ba767f4b30c4ffbe0b46e001145a71, regionState=OPENING, regionLocation=00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:26,323 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_OPEN, hasLock=true; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95ba767f4b30c4ffbe0b46e001145a71, ASSIGN because future has completed 2024-12-06T01:53:26,324 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 95ba767f4b30c4ffbe0b46e001145a71, server=00c764f7c3aa,39091,1733450005317}] 2024-12-06T01:53:26,443 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:26,479 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-06T01:53:26,481 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33989, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-06T01:53:26,486 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(132): Open TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,486 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7752): Opening region: {ENCODED => 95ba767f4b30c4ffbe0b46e001145a71, NAME => 'TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71.', STARTKEY => '', ENDKEY => ''} 2024-12-06T01:53:26,487 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestHBaseWalOnEC 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,487 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(898): Instantiated TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-06T01:53:26,487 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7794): checking encryption for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,487 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(7797): checking classloading for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,489 INFO [StoreOpener-95ba767f4b30c4ffbe0b46e001145a71-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,491 INFO [StoreOpener-95ba767f4b30c4ffbe0b46e001145a71-1 {}] compactions.CompactionConfiguration(183): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 95ba767f4b30c4ffbe0b46e001145a71 columnFamilyName cf 2024-12-06T01:53:26,491 DEBUG [StoreOpener-95ba767f4b30c4ffbe0b46e001145a71-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-06T01:53:26,491 INFO [StoreOpener-95ba767f4b30c4ffbe0b46e001145a71-1 {}] regionserver.HStore(327): Store=95ba767f4b30c4ffbe0b46e001145a71/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-06T01:53:26,491 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1038): replaying wal for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,492 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,493 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(5546): Found 0 recovered edits file(s) under hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,493 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1048): stopping wal replay for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,493 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1060): Cleaning up temporary data for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,495 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1093): writing seq id for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,498 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-06T01:53:26,499 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1114): Opened 95ba767f4b30c4ffbe0b46e001145a71; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=67136055, jitterRate=4.051774740219116E-4}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-06T01:53:26,499 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1122): Running coprocessor post-open hooks for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:26,500 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegion(1006): Region open journal for 95ba767f4b30c4ffbe0b46e001145a71: Running coprocessor pre-open hook at 1733450006487Writing region info on filesystem at 1733450006487Initializing all the Stores at 1733450006488 (+1 ms)Instantiating store for column family {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} at 1733450006488Cleaning up temporary data from old regions at 1733450006493 (+5 ms)Running coprocessor post-open hooks at 1733450006499 (+6 ms)Region opened successfully at 1733450006500 (+1 ms) 2024-12-06T01:53:26,501 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2236): Post open deploy tasks for TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71., pid=6, masterSystemTime=1733450006479 2024-12-06T01:53:26,505 DEBUG [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] regionserver.HRegionServer(2266): Finished post open deploy task for TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,505 INFO [RS_OPEN_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_OPEN_REGION, pid=6}] handler.AssignRegionHandler(153): Opened TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,506 INFO [PEWorker-3 {}] assignment.RegionStateStore(223): pid=5 updating hbase:meta row=95ba767f4b30c4ffbe0b46e001145a71, regionState=OPEN, openSeqNum=2, regionLocation=00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:26,511 DEBUG [MiniHBaseClusterRegionServer-EventLoopGroup-10-1 {}] procedure2.ProcedureFutureUtil(82): Going to wake up procedure pid=6, ppid=5, state=RUNNABLE, hasLock=false; OpenRegionProcedure 95ba767f4b30c4ffbe0b46e001145a71, server=00c764f7c3aa,39091,1733450005317 because future has completed 2024-12-06T01:53:26,517 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=6, resume processing ppid=5 2024-12-06T01:53:26,517 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=6, ppid=5, state=SUCCESS, hasLock=false; OpenRegionProcedure 95ba767f4b30c4ffbe0b46e001145a71, server=00c764f7c3aa,39091,1733450005317 in 189 msec 2024-12-06T01:53:26,521 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=5, resume processing ppid=4 2024-12-06T01:53:26,521 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=5, ppid=4, state=SUCCESS, hasLock=false; TransitRegionStateProcedure table=TestHBaseWalOnEC, region=95ba767f4b30c4ffbe0b46e001145a71, ASSIGN in 354 msec 2024-12-06T01:53:26,522 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-06T01:53:26,523 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(964): Put {"totalColumns":1,"row":"TestHBaseWalOnEC","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733450006522"}]},"ts":"1733450006522"} 2024-12-06T01:53:26,526 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(843): Updated tableName=TestHBaseWalOnEC, state=ENABLED in hbase:meta 2024-12-06T01:53:26,528 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, hasLock=true; CreateTableProcedure table=TestHBaseWalOnEC execute state=CREATE_TABLE_POST_OPERATION 2024-12-06T01:53:26,531 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1521): Finished pid=4, state=SUCCESS, hasLock=false; CreateTableProcedure table=TestHBaseWalOnEC in 413 msec 2024-12-06T01:53:26,753 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=4 2024-12-06T01:53:26,754 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: CREATE, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T01:53:26,754 DEBUG [Time-limited test {}] hbase.HBaseTestingUtil(3046): Waiting until all regions of table TestHBaseWalOnEC get assigned. Timeout = 60000ms 2024-12-06T01:53:26,754 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T01:53:26,758 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3100): All regions for table TestHBaseWalOnEC assigned to meta. Checking AM states. 2024-12-06T01:53:26,759 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2024-12-06T01:53:26,759 INFO [Time-limited test {}] hbase.HBaseTestingUtil(3120): All regions for table TestHBaseWalOnEC assigned. 2024-12-06T01:53:26,762 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] client.AsyncNonMetaRegionLocator(310): The fetched location of 'TestHBaseWalOnEC', row='row', locateType=CURRENT is [region=TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71., hostname=00c764f7c3aa,39091,1733450005317, seqNum=2] 2024-12-06T01:53:26,763 DEBUG [RPCClient-NioEventLoopGroup-6-8 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-06T01:53:26,765 INFO [MiniHBaseClusterRegionServer-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:35928, version=4.0.0-alpha-1-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-06T01:53:26,769 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.HMaster$22(4506): Client=jenkins//172.17.0.2 flush TestHBaseWalOnEC 2024-12-06T01:53:26,770 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] procedure2.ProcedureExecutor(1139): Stored pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC 2024-12-06T01:53:26,772 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:26,772 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_PREPARE, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_PREPARE 2024-12-06T01:53:26,773 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=7, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, hasLock=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-06T01:53:26,774 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1860): Initialized subprocedures=[{pid=8, ppid=7, state=RUNNABLE, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-06T01:53:26,866 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-06T01:53:26,871 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-06T01:53:26,883 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:26,920 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-06T01:53:26,920 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-06T01:53:26,929 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39091 {}] regionserver.RSRpcServices(3929): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=8 2024-12-06T01:53:26,929 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(51): Starting region operation on TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,930 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2902): Flushing 95ba767f4b30c4ffbe0b46e001145a71 1/1 column families, dataSize=32 B heapSize=360 B 2024-12-06T01:53:26,949 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/.tmp/cf/832a3865ff7e4585a1012e2754932249 is 36, key is row/cf:cq/1733450006766/Put/seqid=0 2024-12-06T01:53:26,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741839_1015 (size=4787) 2024-12-06T01:53:26,957 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741839_1015 (size=4787) 2024-12-06T01:53:26,957 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741839_1015 (size=4787) 2024-12-06T01:53:26,958 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=32 B at sequenceid=5 (bloomFilter=false), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/.tmp/cf/832a3865ff7e4585a1012e2754932249 2024-12-06T01:53:26,968 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/.tmp/cf/832a3865ff7e4585a1012e2754932249 as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/cf/832a3865ff7e4585a1012e2754932249 2024-12-06T01:53:26,977 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/cf/832a3865ff7e4585a1012e2754932249, entries=1, sequenceid=5, filesize=4.7 K 2024-12-06T01:53:26,979 INFO [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(3140): Finished flush of dataSize ~32 B/32, heapSize ~344 B/344, currentSize=0 B/0 for 95ba767f4b30c4ffbe0b46e001145a71 in 50ms, sequenceid=5, compaction requested=false 2024-12-06T01:53:26,979 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.HRegion(2603): Flush status journal for 95ba767f4b30c4ffbe0b46e001145a71: 2024-12-06T01:53:26,979 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.FlushRegionCallable(64): Closing region operation on TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:26,979 DEBUG [RS_FLUSH_OPERATIONS-regionserver/00c764f7c3aa:0-0 {event_type=RS_FLUSH_REGIONS, pid=8}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=8 2024-12-06T01:53:26,980 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.HMaster(4169): Remote procedure done, pid=8 2024-12-06T01:53:26,987 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(2017): Finished subprocedure pid=8, resume processing ppid=7 2024-12-06T01:53:26,987 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1521): Finished pid=8, ppid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 208 msec 2024-12-06T01:53:26,991 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1521): Finished pid=7, state=SUCCESS, hasLock=false; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=7, table=TestHBaseWalOnEC in 219 msec 2024-12-06T01:53:27,093 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32985 {}] master.MasterRpcServices(1377): Checking to see if procedure is done pid=7 2024-12-06T01:53:27,094 INFO [RPCClient-NioEventLoopGroup-6-9 {}] client.RawAsyncHBaseAdmin$TableProcedureBiConsumer(2721): Operation: FLUSH, Table Name: default:TestHBaseWalOnEC completed 2024-12-06T01:53:27,099 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1019): Shutting down minicluster 2024-12-06T01:53:27,099 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T01:53:27,099 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hbase.thirdparty.com.google.common.io.Closeables.close(Closeables.java:79) at org.apache.hadoop.hbase.HBaseTestingUtil.closeConnection(HBaseTestingUtil.java:2611) at org.apache.hadoop.hbase.HBaseTestingUtil.cleanup(HBaseTestingUtil.java:1065) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1034) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:27,099 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,099 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,100 INFO [Registry-endpoints-refresh-end-points {}] client.RegistryEndpointsRefresher(78): Registry end points refresher loop exited. 2024-12-06T01:53:27,100 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-06T01:53:27,100 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=91922108, stopped=false 2024-12-06T01:53:27,100 INFO [Time-limited test {}] master.ServerManager(983): Cluster shutdown requested of master=00c764f7c3aa,32985,1733450005268 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-06T01:53:27,103 INFO [Time-limited test {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:27,103 INFO [Time-limited test {}] client.AsyncConnectionImpl(233): Connection has been closed by Time-limited test. 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:27,103 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:27,103 DEBUG [Time-limited test {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.master.HMaster.lambda$shutdown$17(HMaster.java:3306) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.master.HMaster.shutdown(HMaster.java:3277) at org.apache.hadoop.hbase.util.JVMClusterUtil.shutdown(JVMClusterUtil.java:265) at org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:416) at org.apache.hadoop.hbase.SingleProcessHBaseCluster.shutdown(SingleProcessHBaseCluster.java:676) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniHBaseCluster(HBaseTestingUtil.java:1036) at org.apache.hadoop.hbase.HBaseTestingUtil.shutdownMiniCluster(HBaseTestingUtil.java:1020) at org.apache.hadoop.hbase.regionserver.wal.TestHBaseWalOnEC.tearDown(TestHBaseWalOnEC.java:101) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:27,104 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,104 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c764f7c3aa,39091,1733450005317' ***** 2024-12-06T01:53:27,104 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T01:53:27,104 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:27,104 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c764f7c3aa,39077,1733450005347' ***** 2024-12-06T01:53:27,104 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T01:53:27,104 INFO [Time-limited test {}] regionserver.HRegionServer(2196): ***** STOPPING region server '00c764f7c3aa,43335,1733450005376' ***** 2024-12-06T01:53:27,104 INFO [Time-limited test {}] regionserver.HRegionServer(2210): STOPPED: Shutdown requested 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T01:53:27,105 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T01:53:27,105 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(959): stopping server 00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:1;00c764f7c3aa:39077. 2024-12-06T01:53:27,105 DEBUG [RS:1;00c764f7c3aa:39077 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:27,105 DEBUG [RS:1;00c764f7c3aa:39077 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,105 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:27,105 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(976): stopping server 00c764f7c3aa,39077,1733450005347; all regions closed. 2024-12-06T01:53:27,105 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T01:53:27,106 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(959): stopping server 00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:27,106 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:2;00c764f7c3aa:43335. 2024-12-06T01:53:27,106 INFO [RS:0;00c764f7c3aa:39091 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-06T01:53:27,106 DEBUG [RS:2;00c764f7c3aa:43335 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:27,106 INFO [RS:0;00c764f7c3aa:39091 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-06T01:53:27,106 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(374): MemStoreFlusher.0 exiting 2024-12-06T01:53:27,106 DEBUG [RS:2;00c764f7c3aa:43335 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,106 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(3091): Received CLOSE for 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:27,106 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T01:53:27,106 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T01:53:27,107 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(3091): Received CLOSE for 1588230740 2024-12-06T01:53:27,107 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(959): stopping server 00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:27,107 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:27,107 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,107 INFO [RS:0;00c764f7c3aa:39091 {}] client.AsyncConnectionImpl(233): Connection has been closed by RS:0;00c764f7c3aa:39091. 2024-12-06T01:53:27,107 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-06T01:53:27,107 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1325): Online Regions={1588230740=hbase:meta,,1.1588230740} 2024-12-06T01:53:27,107 DEBUG [RS:0;00c764f7c3aa:39091 {}] client.AsyncConnectionImpl(264): Call stack: at java.base/java.lang.Thread.getStackTrace(Thread.java:1619) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.lambda$close$5(AsyncConnectionImpl.java:235) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) at org.apache.hadoop.hbase.client.AsyncConnectionImpl.close(AsyncConnectionImpl.java:229) at org.apache.hadoop.hbase.HBaseServerBase.closeClusterConnection(HBaseServerBase.java:457) at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:962) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(SingleProcessHBaseCluster.java:171) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer$1.run(SingleProcessHBaseCluster.java:155) at java.base/java.security.AccessController.doPrivileged(AccessController.java:399) at java.base/javax.security.auth.Subject.doAs(Subject.java:376) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1930) at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:322) at org.apache.hadoop.hbase.SingleProcessHBaseCluster$MiniHBaseClusterRegionServer.run(SingleProcessHBaseCluster.java:152) at java.base/java.lang.Thread.run(Thread.java:840) 2024-12-06T01:53:27,107 DEBUG [RS:0;00c764f7c3aa:39091 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,107 DEBUG [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1351): Waiting on 1588230740 2024-12-06T01:53:27,107 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,107 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1321): Waiting on 1 regions to close 2024-12-06T01:53:27,107 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1325): Online Regions={95ba767f4b30c4ffbe0b46e001145a71=TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71.} 2024-12-06T01:53:27,107 DEBUG [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1351): Waiting on 95ba767f4b30c4ffbe0b46e001145a71 2024-12-06T01:53:27,107 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1722): Closing 1588230740, disabling compactions & flushes 2024-12-06T01:53:27,107 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1722): Closing 95ba767f4b30c4ffbe0b46e001145a71, disabling compactions & flushes 2024-12-06T01:53:27,107 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,107 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1755): Closing region hbase:meta,,1.1588230740 2024-12-06T01:53:27,107 INFO [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1755): Closing region TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:27,107 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1776): Time limited wait for close lock on hbase:meta,,1.1588230740 2024-12-06T01:53:27,107 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1776): Time limited wait for close lock on TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:27,108 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1843): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-06T01:53:27,108 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1843): Acquired close lock on TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. after waiting 0 ms 2024-12-06T01:53:27,108 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,108 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1853): Updates disabled for region hbase:meta,,1.1588230740 2024-12-06T01:53:27,108 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1853): Updates disabled for region TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:27,108 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2902): Flushing 1588230740 4/4 column families, dataSize=1.34 KB heapSize=3.38 KB 2024-12-06T01:53:27,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741834_1010 (size=93) 2024-12-06T01:53:27,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741834_1010 (size=93) 2024-12-06T01:53:27,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741834_1010 (size=93) 2024-12-06T01:53:27,117 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/default/TestHBaseWalOnEC/95ba767f4b30c4ffbe0b46e001145a71/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2024-12-06T01:53:27,119 DEBUG [RS:1;00c764f7c3aa:39077 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs 2024-12-06T01:53:27,119 INFO [RS:1;00c764f7c3aa:39077 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c764f7c3aa%2C39077%2C1733450005347:(num 1733450005820) 2024-12-06T01:53:27,119 DEBUG [RS:1;00c764f7c3aa:39077 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,119 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:27,119 INFO [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1973): Closed TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:27,119 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1676): Region close journal for 95ba767f4b30c4ffbe0b46e001145a71: Waiting for close lock at 1733450007107Running coprocessor pre-close hooks at 1733450007107Disabling compacts and flushes for region at 1733450007107Disabling writes for close at 1733450007108 (+1 ms)Writing region close event to WAL at 1733450007108Running coprocessor post-close hooks at 1733450007119 (+11 ms)Closed at 1733450007119 2024-12-06T01:53:27,119 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:27,119 DEBUG [RS_CLOSE_REGION-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71. 2024-12-06T01:53:27,120 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.ChoreService(370): Chore service for: regionserver/00c764f7c3aa:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:27,120 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T01:53:27,120 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T01:53:27,120 INFO [regionserver/00c764f7c3aa:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:27,120 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T01:53:27,120 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:27,120 INFO [RS:1;00c764f7c3aa:39077 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39077 2024-12-06T01:53:27,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c764f7c3aa,39077,1733450005347 2024-12-06T01:53:27,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:27,123 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:27,125 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c764f7c3aa,39077,1733450005347] 2024-12-06T01:53:27,127 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c764f7c3aa,39077,1733450005347 already deleted, retry=false 2024-12-06T01:53:27,127 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c764f7c3aa,39077,1733450005347 expired; onlineServers=2 2024-12-06T01:53:27,133 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/info/02bb74ec539d44fabdd9dd085fa2dd99 is 153, key is TestHBaseWalOnEC,,1733450006115.95ba767f4b30c4ffbe0b46e001145a71./info:regioninfo/1733450006506/Put/seqid=0 2024-12-06T01:53:27,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741840_1016 (size=6637) 2024-12-06T01:53:27,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741840_1016 (size=6637) 2024-12-06T01:53:27,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741840_1016 (size=6637) 2024-12-06T01:53:27,142 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.18 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/info/02bb74ec539d44fabdd9dd085fa2dd99 2024-12-06T01:53:27,161 INFO [regionserver/00c764f7c3aa:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:27,166 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/ns/90c1b66c667547fbbf45127c0bed8cf5 is 43, key is default/ns:d/1733450006077/Put/seqid=0 2024-12-06T01:53:27,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741841_1017 (size=5153) 2024-12-06T01:53:27,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741841_1017 (size=5153) 2024-12-06T01:53:27,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741841_1017 (size=5153) 2024-12-06T01:53:27,175 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=74 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/ns/90c1b66c667547fbbf45127c0bed8cf5 2024-12-06T01:53:27,178 INFO [regionserver/00c764f7c3aa:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:27,181 INFO [regionserver/00c764f7c3aa:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:27,200 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/table/91962273db854e3a8fcb29ea86ebd9b4 is 52, key is TestHBaseWalOnEC/table:state/1733450006522/Put/seqid=0 2024-12-06T01:53:27,207 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741842_1018 (size=5249) 2024-12-06T01:53:27,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741842_1018 (size=5249) 2024-12-06T01:53:27,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741842_1018 (size=5249) 2024-12-06T01:53:27,209 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=96 B at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/table/91962273db854e3a8fcb29ea86ebd9b4 2024-12-06T01:53:27,217 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/info/02bb74ec539d44fabdd9dd085fa2dd99 as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/info/02bb74ec539d44fabdd9dd085fa2dd99 2024-12-06T01:53:27,225 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,225 INFO [RS:1;00c764f7c3aa:39077 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:27,225 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39077-0x1006507b3e20002, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,225 INFO [RS:1;00c764f7c3aa:39077 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c764f7c3aa,39077,1733450005347; zookeeper connection closed. 2024-12-06T01:53:27,225 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@5f95ac96 {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@5f95ac96 2024-12-06T01:53:27,227 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/info/02bb74ec539d44fabdd9dd085fa2dd99, entries=10, sequenceid=11, filesize=6.5 K 2024-12-06T01:53:27,228 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/ns/90c1b66c667547fbbf45127c0bed8cf5 as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/ns/90c1b66c667547fbbf45127c0bed8cf5 2024-12-06T01:53:27,237 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/ns/90c1b66c667547fbbf45127c0bed8cf5, entries=2, sequenceid=11, filesize=5.0 K 2024-12-06T01:53:27,238 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/.tmp/table/91962273db854e3a8fcb29ea86ebd9b4 as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/table/91962273db854e3a8fcb29ea86ebd9b4 2024-12-06T01:53:27,246 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/table/91962273db854e3a8fcb29ea86ebd9b4, entries=2, sequenceid=11, filesize=5.1 K 2024-12-06T01:53:27,247 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3140): Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 139ms, sequenceid=11, compaction requested=false 2024-12-06T01:53:27,258 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(410): Wrote file=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/data/hbase/meta/1588230740/recovered.edits/14.seqid, newMaxSeqId=14, maxSeqId=1 2024-12-06T01:53:27,259 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-06T01:53:27,259 INFO [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1973): Closed hbase:meta,,1.1588230740 2024-12-06T01:53:27,259 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1676): Region close journal for 1588230740: Waiting for close lock at 1733450007107Running coprocessor pre-close hooks at 1733450007107Disabling compacts and flushes for region at 1733450007107Disabling writes for close at 1733450007108 (+1 ms)Obtaining lock to block concurrent updates at 1733450007108Preparing flush snapshotting stores in 1588230740 at 1733450007108Finished memstore snapshotting hbase:meta,,1.1588230740, syncing WAL and waiting on mvcc, flushsize=dataSize=1377, getHeapSize=3392, getOffHeapSize=0, getCellsCount=14 at 1733450007108Flushing stores of hbase:meta,,1.1588230740 at 1733450007110 (+2 ms)Flushing 1588230740/info: creating writer at 1733450007110Flushing 1588230740/info: appending metadata at 1733450007133 (+23 ms)Flushing 1588230740/info: closing flushed file at 1733450007133Flushing 1588230740/ns: creating writer at 1733450007150 (+17 ms)Flushing 1588230740/ns: appending metadata at 1733450007166 (+16 ms)Flushing 1588230740/ns: closing flushed file at 1733450007166Flushing 1588230740/table: creating writer at 1733450007183 (+17 ms)Flushing 1588230740/table: appending metadata at 1733450007199 (+16 ms)Flushing 1588230740/table: closing flushed file at 1733450007200 (+1 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@3135af3: reopening flushed file at 1733450007216 (+16 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@f7d124c: reopening flushed file at 1733450007227 (+11 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@5549f8b2: reopening flushed file at 1733450007237 (+10 ms)Finished flush of dataSize ~1.34 KB/1377, heapSize ~3.08 KB/3152, currentSize=0 B/0 for 1588230740 in 139ms, sequenceid=11, compaction requested=false at 1733450007247 (+10 ms)Writing region close event to WAL at 1733450007253 (+6 ms)Running coprocessor post-close hooks at 1733450007259 (+6 ms)Closed at 1733450007259 2024-12-06T01:53:27,259 DEBUG [RS_CLOSE_META-regionserver/00c764f7c3aa:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-06T01:53:27,307 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(976): stopping server 00c764f7c3aa,43335,1733450005376; all regions closed. 2024-12-06T01:53:27,307 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(976): stopping server 00c764f7c3aa,39091,1733450005317; all regions closed. 2024-12-06T01:53:27,308 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,308 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,309 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,309 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741836_1012 (size=2751) 2024-12-06T01:53:27,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741835_1011 (size=1298) 2024-12-06T01:53:27,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741835_1011 (size=1298) 2024-12-06T01:53:27,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741836_1012 (size=2751) 2024-12-06T01:53:27,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741836_1012 (size=2751) 2024-12-06T01:53:27,316 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741835_1011 (size=1298) 2024-12-06T01:53:27,319 DEBUG [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs 2024-12-06T01:53:27,319 INFO [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c764f7c3aa%2C43335%2C1733450005376.meta:.meta(num 1733450006018) 2024-12-06T01:53:27,319 DEBUG [RS:0;00c764f7c3aa:39091 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs 2024-12-06T01:53:27,319 INFO [RS:0;00c764f7c3aa:39091 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c764f7c3aa%2C39091%2C1733450005317:(num 1733450005828) 2024-12-06T01:53:27,319 DEBUG [RS:0;00c764f7c3aa:39091 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,319 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:27,319 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:27,319 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.ChoreService(370): Chore service for: regionserver/00c764f7c3aa:0 had [ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:27,320 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,320 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-06T01:53:27,320 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-06T01:53:27,320 INFO [regionserver/00c764f7c3aa:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:27,320 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,320 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-06T01:53:27,320 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:27,321 INFO [RS:0;00c764f7c3aa:39091 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:39091 2024-12-06T01:53:27,321 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,321 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,322 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,323 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c764f7c3aa,39091,1733450005317 2024-12-06T01:53:27,323 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:27,323 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:27,325 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c764f7c3aa,39091,1733450005317] 2024-12-06T01:53:27,327 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c764f7c3aa,39091,1733450005317 already deleted, retry=false 2024-12-06T01:53:27,327 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c764f7c3aa,39091,1733450005317 expired; onlineServers=1 2024-12-06T01:53:27,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741833_1009 (size=93) 2024-12-06T01:53:27,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741833_1009 (size=93) 2024-12-06T01:53:27,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741833_1009 (size=93) 2024-12-06T01:53:27,335 DEBUG [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(1256): Moved 1 WAL file(s) to /user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/oldWALs 2024-12-06T01:53:27,335 INFO [RS:2;00c764f7c3aa:43335 {}] wal.AbstractFSWAL(1259): Closed WAL: FSHLog 00c764f7c3aa%2C43335%2C1733450005376:(num 1733450005809) 2024-12-06T01:53:27,335 DEBUG [RS:2;00c764f7c3aa:43335 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-06T01:53:27,335 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.LeaseManager(133): Closed leases 2024-12-06T01:53:27,335 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:27,335 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.ChoreService(370): Chore service for: regionserver/00c764f7c3aa:0 had [ScheduledChore name=ReplicationSinkStatistics, period=300000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=ReplicationSourceStatistics, period=300000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:27,335 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:27,336 INFO [RS:2;00c764f7c3aa:43335 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:43335 2024-12-06T01:53:27,336 INFO [regionserver/00c764f7c3aa:0.logRoller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:27,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/00c764f7c3aa,43335,1733450005376 2024-12-06T01:53:27,338 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:27,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-06T01:53:27,340 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [00c764f7c3aa,43335,1733450005376] 2024-12-06T01:53:27,341 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(212): Node /hbase/draining/00c764f7c3aa,43335,1733450005376 already deleted, retry=false 2024-12-06T01:53:27,341 INFO [RegionServerTracker-0 {}] master.ServerManager(688): Cluster shutdown set; 00c764f7c3aa,43335,1733450005376 expired; onlineServers=0 2024-12-06T01:53:27,341 INFO [RegionServerTracker-0 {}] master.HMaster(3321): ***** STOPPING master '00c764f7c3aa,32985,1733450005268' ***** 2024-12-06T01:53:27,341 INFO [RegionServerTracker-0 {}] master.HMaster(3323): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-06T01:53:27,342 INFO [M:0;00c764f7c3aa:32985 {}] hbase.HBaseServerBase(455): Close async cluster connection 2024-12-06T01:53:27,342 INFO [M:0;00c764f7c3aa:32985 {}] hbase.HBaseServerBase(438): Shutdown chores and chore service 2024-12-06T01:53:27,342 DEBUG [M:0;00c764f7c3aa:32985 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-06T01:53:27,342 DEBUG [M:0;00c764f7c3aa:32985 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-06T01:53:27,342 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-06T01:53:27,342 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.large.0-1733450005587 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.large.0-1733450005587,5,FailOnTimeoutGroup] 2024-12-06T01:53:27,342 DEBUG [master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.small.0-1733450005587 {}] cleaner.HFileCleaner(306): Exit Thread[master/00c764f7c3aa:0:becomeActiveMaster-HFileCleaner.small.0-1733450005587,5,FailOnTimeoutGroup] 2024-12-06T01:53:27,342 INFO [M:0;00c764f7c3aa:32985 {}] hbase.ChoreService(370): Chore service for: master/00c764f7c3aa:0 had [ScheduledChore name=FlushedSequenceIdFlusher, period=10800000, unit=MILLISECONDS] on shutdown 2024-12-06T01:53:27,342 INFO [M:0;00c764f7c3aa:32985 {}] hbase.HBaseServerBase(448): Shutdown executor service 2024-12-06T01:53:27,342 DEBUG [M:0;00c764f7c3aa:32985 {}] master.HMaster(1795): Stopping service threads 2024-12-06T01:53:27,342 INFO [M:0;00c764f7c3aa:32985 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-06T01:53:27,343 INFO [M:0;00c764f7c3aa:32985 {}] procedure2.ProcedureExecutor(723): Stopping 2024-12-06T01:53:27,343 INFO [M:0;00c764f7c3aa:32985 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-06T01:53:27,343 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-06T01:53:27,344 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-06T01:53:27,344 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-06T01:53:27,344 DEBUG [M:0;00c764f7c3aa:32985 {}] zookeeper.ZKUtil(347): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-06T01:53:27,344 WARN [M:0;00c764f7c3aa:32985 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-06T01:53:27,345 INFO [M:0;00c764f7c3aa:32985 {}] master.ServerManager(1139): Writing .lastflushedseqids file at: hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/.lastflushedseqids 2024-12-06T01:53:27,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741843_1019 (size=127) 2024-12-06T01:53:27,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741843_1019 (size=127) 2024-12-06T01:53:27,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741843_1019 (size=127) 2024-12-06T01:53:27,355 INFO [M:0;00c764f7c3aa:32985 {}] assignment.AssignmentManager(395): Stopping assignment manager 2024-12-06T01:53:27,355 INFO [M:0;00c764f7c3aa:32985 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-06T01:53:27,355 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1722): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-06T01:53:27,355 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1755): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:27,355 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1776): Time limited wait for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:27,355 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1843): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-06T01:53:27,355 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1853): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:27,356 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(2902): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=26.84 KB heapSize=34.13 KB 2024-12-06T01:53:27,375 DEBUG [M:0;00c764f7c3aa:32985 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/db8e4e8bf5844d29b1c443c284062742 is 82, key is hbase:meta,,1/info:regioninfo/1733450006057/Put/seqid=0 2024-12-06T01:53:27,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741844_1020 (size=5672) 2024-12-06T01:53:27,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741844_1020 (size=5672) 2024-12-06T01:53:27,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741844_1020 (size=5672) 2024-12-06T01:53:27,387 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/db8e4e8bf5844d29b1c443c284062742 2024-12-06T01:53:27,411 DEBUG [M:0;00c764f7c3aa:32985 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/e2722753ce1c454ba5b809800af12e3d is 748, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733450006530/Put/seqid=0 2024-12-06T01:53:27,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741845_1021 (size=6440) 2024-12-06T01:53:27,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741845_1021 (size=6440) 2024-12-06T01:53:27,422 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=26.15 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/e2722753ce1c454ba5b809800af12e3d 2024-12-06T01:53:27,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741845_1021 (size=6440) 2024-12-06T01:53:27,425 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,425 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39091-0x1006507b3e20001, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,425 INFO [RS:0;00c764f7c3aa:39091 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:27,425 INFO [RS:0;00c764f7c3aa:39091 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c764f7c3aa,39091,1733450005317; zookeeper connection closed. 2024-12-06T01:53:27,426 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@56e7670f {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@56e7670f 2024-12-06T01:53:27,440 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,440 INFO [RS:2;00c764f7c3aa:43335 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:27,440 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43335-0x1006507b3e20003, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,440 INFO [RS:2;00c764f7c3aa:43335 {}] regionserver.HRegionServer(1031): Exiting; stopping=00c764f7c3aa,43335,1733450005376; zookeeper connection closed. 2024-12-06T01:53:27,441 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@4b9b10ef {}] hbase.SingleProcessHBaseCluster$SingleFileSystemShutdownThread(211): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@4b9b10ef 2024-12-06T01:53:27,441 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2024-12-06T01:53:27,448 DEBUG [M:0;00c764f7c3aa:32985 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1432961d4020468e84605e4bf0a9245a is 69, key is 00c764f7c3aa,39077,1733450005347/rs:state/1733450005635/Put/seqid=0 2024-12-06T01:53:27,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741846_1022 (size=5294) 2024-12-06T01:53:27,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741846_1022 (size=5294) 2024-12-06T01:53:27,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741846_1022 (size=5294) 2024-12-06T01:53:27,459 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=195 B at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1432961d4020468e84605e4bf0a9245a 2024-12-06T01:53:27,466 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/db8e4e8bf5844d29b1c443c284062742 as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/db8e4e8bf5844d29b1c443c284062742 2024-12-06T01:53:27,475 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/db8e4e8bf5844d29b1c443c284062742, entries=8, sequenceid=72, filesize=5.5 K 2024-12-06T01:53:27,477 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/e2722753ce1c454ba5b809800af12e3d as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/e2722753ce1c454ba5b809800af12e3d 2024-12-06T01:53:27,485 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/e2722753ce1c454ba5b809800af12e3d, entries=8, sequenceid=72, filesize=6.3 K 2024-12-06T01:53:27,486 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/1432961d4020468e84605e4bf0a9245a as hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1432961d4020468e84605e4bf0a9245a 2024-12-06T01:53:27,494 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HStore$StoreFlusherImpl(1990): Added hdfs://localhost:40977/user/jenkins/test-data/80789da6-5f98-db60-6756-dd33c9084d2c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/1432961d4020468e84605e4bf0a9245a, entries=3, sequenceid=72, filesize=5.2 K 2024-12-06T01:53:27,496 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(3140): Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 140ms, sequenceid=72, compaction requested=false 2024-12-06T01:53:27,498 INFO [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1973): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-06T01:53:27,498 DEBUG [M:0;00c764f7c3aa:32985 {}] regionserver.HRegion(1676): Region close journal for 1595e783b53d99cd5eef43b6debb2682: Waiting for close lock at 1733450007355Disabling compacts and flushes for region at 1733450007355Disabling writes for close at 1733450007355Obtaining lock to block concurrent updates at 1733450007356 (+1 ms)Preparing flush snapshotting stores in 1595e783b53d99cd5eef43b6debb2682 at 1733450007356Finished memstore snapshotting master:store,,1.1595e783b53d99cd5eef43b6debb2682., syncing WAL and waiting on mvcc, flushsize=dataSize=27480, getHeapSize=34880, getOffHeapSize=0, getCellsCount=85 at 1733450007356Flushing stores of master:store,,1.1595e783b53d99cd5eef43b6debb2682. at 1733450007357 (+1 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: creating writer at 1733450007357Flushing 1595e783b53d99cd5eef43b6debb2682/info: appending metadata at 1733450007375 (+18 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/info: closing flushed file at 1733450007375Flushing 1595e783b53d99cd5eef43b6debb2682/proc: creating writer at 1733450007394 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: appending metadata at 1733450007411 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/proc: closing flushed file at 1733450007411Flushing 1595e783b53d99cd5eef43b6debb2682/rs: creating writer at 1733450007430 (+19 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: appending metadata at 1733450007447 (+17 ms)Flushing 1595e783b53d99cd5eef43b6debb2682/rs: closing flushed file at 1733450007447Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@43e883eb: reopening flushed file at 1733450007465 (+18 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@162527cb: reopening flushed file at 1733450007475 (+10 ms)Flushing org.apache.hadoop.hbase.regionserver.HStore$StoreFlusherImpl@727de284: reopening flushed file at 1733450007485 (+10 ms)Finished flush of dataSize ~26.84 KB/27480, heapSize ~33.83 KB/34640, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 140ms, sequenceid=72, compaction requested=false at 1733450007496 (+11 ms)Writing region close event to WAL at 1733450007498 (+2 ms)Closed at 1733450007498 2024-12-06T01:53:27,499 INFO [sync.0 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,499 INFO [sync.1 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,499 INFO [sync.2 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,499 INFO [sync.3 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,499 INFO [sync.4 {}] wal.FSHLog$SyncRunner(477): interrupted 2024-12-06T01:53:27,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33671 is added to blk_1073741830_1006 (size=32683) 2024-12-06T01:53:27,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45831 is added to blk_1073741830_1006 (size=32683) 2024-12-06T01:53:27,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45079 is added to blk_1073741830_1006 (size=32683) 2024-12-06T01:53:27,504 INFO [M:0;00c764f7c3aa:32985 {}] flush.MasterFlushTableProcedureManager(90): stop: server shutting down. 2024-12-06T01:53:27,504 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(249): LogRoller exiting. 2024-12-06T01:53:27,505 INFO [M:0;00c764f7c3aa:32985 {}] ipc.NettyRpcServer(345): Stopping server on /172.17.0.2:32985 2024-12-06T01:53:27,506 INFO [M:0;00c764f7c3aa:32985 {}] hbase.HBaseServerBase(479): Close zookeeper 2024-12-06T01:53:27,608 INFO [M:0;00c764f7c3aa:32985 {}] hbase.HBaseServerBase(486): Close table descriptors 2024-12-06T01:53:27,608 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,608 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:32985-0x1006507b3e20000, quorum=127.0.0.1:57244, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-06T01:53:27,610 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6ff8a873{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:27,611 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1ec581d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:27,611 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:27,611 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@328032c9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:27,611 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@e56045e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:27,613 WARN [BP-803844786-172.17.0.2-1733450004348 heartbeating to localhost/127.0.0.1:40977 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T01:53:27,613 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T01:53:27,613 WARN [BP-803844786-172.17.0.2-1733450004348 heartbeating to localhost/127.0.0.1:40977 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-803844786-172.17.0.2-1733450004348 (Datanode Uuid 3fa84b4d-de66-4c7c-9324-c0d5ea5bf184) service to localhost/127.0.0.1:40977 2024-12-06T01:53:27,613 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T01:53:27,614 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data5/current/BP-803844786-172.17.0.2-1733450004348 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:27,614 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data6/current/BP-803844786-172.17.0.2-1733450004348 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:27,614 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T01:53:27,616 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@25106b03{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:27,616 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@ae51624{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:27,617 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:27,617 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@564d8641{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:27,617 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@55fefad1{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:27,618 WARN [BP-803844786-172.17.0.2-1733450004348 heartbeating to localhost/127.0.0.1:40977 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T01:53:27,618 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T01:53:27,618 WARN [BP-803844786-172.17.0.2-1733450004348 heartbeating to localhost/127.0.0.1:40977 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-803844786-172.17.0.2-1733450004348 (Datanode Uuid 1d7909ec-71be-4281-bec0-349cd85878e2) service to localhost/127.0.0.1:40977 2024-12-06T01:53:27,618 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T01:53:27,619 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data3/current/BP-803844786-172.17.0.2-1733450004348 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:27,619 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data4/current/BP-803844786-172.17.0.2-1733450004348 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:27,620 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T01:53:27,624 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@760f4a1c{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-06T01:53:27,624 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@e494f88{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:27,624 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:27,624 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6d1c78c1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:27,625 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1841cfc0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:27,626 WARN [BP-803844786-172.17.0.2-1733450004348 heartbeating to localhost/127.0.0.1:40977 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-06T01:53:27,626 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-06T01:53:27,626 WARN [BP-803844786-172.17.0.2-1733450004348 heartbeating to localhost/127.0.0.1:40977 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-803844786-172.17.0.2-1733450004348 (Datanode Uuid f0ebc5d8-587b-460d-8288-27f072d5ac52) service to localhost/127.0.0.1:40977 2024-12-06T01:53:27,626 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-06T01:53:27,627 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data1/current/BP-803844786-172.17.0.2-1733450004348 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:27,627 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/cluster_c35f71cc-fc38-2ec0-b50c-f67f714d4ae9/data/data2/current/BP-803844786-172.17.0.2-1733450004348 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-06T01:53:27,627 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-06T01:53:27,642 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1a4689e1{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-06T01:53:27,642 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@53c85678{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-06T01:53:27,642 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-06T01:53:27,643 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@18ee857{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-06T01:53:27,643 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2baf02db{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_master/hbase-server/target/test-data/243d5e56-0d61-8f83-caa4-734437509e15/hadoop.log.dir/,STOPPED} 2024-12-06T01:53:27,651 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(347): Shutdown MiniZK cluster with all ZK servers 2024-12-06T01:53:27,677 INFO [Time-limited test {}] hbase.HBaseTestingUtil(1026): Minicluster is down 2024-12-06T01:53:27,685 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestHBaseWalOnEC#testReadWrite[1] Thread=148 (was 88) - Thread LEAK? -, OpenFileDescriptor=519 (was 439) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=366 (was 337) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=9701 (was 9910)