2024-12-02 01:06:33,502 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc 2024-12-02 01:06:33,517 main DEBUG Took 0.012535 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-02 01:06:33,517 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-02 01:06:33,518 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-02 01:06:33,519 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-02 01:06:33,520 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,526 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-02 01:06:33,539 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,540 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,541 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,542 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,542 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,543 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,544 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,544 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,545 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,545 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,546 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,547 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,547 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,548 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,548 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,549 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,549 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,549 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,550 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,550 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,550 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,550 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,551 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,551 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-02 01:06:33,552 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,552 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-02 01:06:33,553 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-02 01:06:33,554 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-02 01:06:33,556 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-02 01:06:33,557 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-02 01:06:33,558 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-02 01:06:33,559 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-02 01:06:33,566 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-02 01:06:33,568 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-02 01:06:33,570 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-02 01:06:33,570 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-02 01:06:33,571 main DEBUG createAppenders(={Console}) 2024-12-02 01:06:33,571 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc initialized 2024-12-02 01:06:33,572 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc 2024-12-02 01:06:33,572 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc OK. 2024-12-02 01:06:33,572 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-02 01:06:33,572 main DEBUG OutputStream closed 2024-12-02 01:06:33,573 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-02 01:06:33,573 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-02 01:06:33,573 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@53ce1329 OK 2024-12-02 01:06:33,634 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-02 01:06:33,636 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-02 01:06:33,637 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-02 01:06:33,637 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-02 01:06:33,638 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-02 01:06:33,638 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-02 01:06:33,639 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-02 01:06:33,639 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-02 01:06:33,639 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-02 01:06:33,639 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-02 01:06:33,640 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-02 01:06:33,640 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-02 01:06:33,640 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-02 01:06:33,640 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-02 01:06:33,641 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-02 01:06:33,641 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-02 01:06:33,641 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-02 01:06:33,642 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-02 01:06:33,644 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-02 01:06:33,644 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-logging/target/hbase-logging-2.7.0-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@6dab9b6d) with optional ClassLoader: null 2024-12-02 01:06:33,644 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-02 01:06:33,645 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@6dab9b6d] started OK. 2024-12-02T01:06:33,873 DEBUG [main {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44 2024-12-02 01:06:33,875 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-02 01:06:33,875 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-02T01:06:33,882 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestLogRolling timeout: 13 mins 2024-12-02T01:06:33,909 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testSlowSyncLogRolling Thread=12, OpenFileDescriptor=286, MaxFileDescriptor=1048576, SystemLoadAverage=206, ProcessCount=11, AvailableMemoryMB=6409 2024-12-02T01:06:33,911 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T01:06:33,913 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68, deleteOnExit=true 2024-12-02T01:06:33,913 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-02T01:06:33,914 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/test.cache.data in system properties and HBase conf 2024-12-02T01:06:33,914 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T01:06:33,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir in system properties and HBase conf 2024-12-02T01:06:33,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T01:06:33,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T01:06:33,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-02T01:06:33,986 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-02T01:06:34,059 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T01:06:34,063 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:06:34,063 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:06:34,064 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T01:06:34,064 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:06:34,064 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T01:06:34,065 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T01:06:34,065 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:06:34,066 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:06:34,066 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T01:06:34,066 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/nfs.dump.dir in system properties and HBase conf 2024-12-02T01:06:34,067 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/java.io.tmpdir in system properties and HBase conf 2024-12-02T01:06:34,067 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:06:34,067 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T01:06:34,068 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T01:06:34,466 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:06:34,969 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-02T01:06:35,030 INFO [Time-limited test {}] log.Log(170): Logging initialized @2157ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-02T01:06:35,087 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:06:35,143 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:06:35,159 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:06:35,159 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:06:35,161 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:06:35,172 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:06:35,174 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@37896107{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:06:35,175 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7c7fe1c5{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:06:35,330 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4cd532bc{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/java.io.tmpdir/jetty-localhost-38159-hadoop-hdfs-3_4_1-tests_jar-_-any-3791426087368808570/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:06:35,336 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@619b4309{HTTP/1.1, (http/1.1)}{localhost:38159} 2024-12-02T01:06:35,336 INFO [Time-limited test {}] server.Server(415): Started @2464ms 2024-12-02T01:06:35,357 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:06:35,814 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:06:35,820 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:06:35,821 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:06:35,821 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:06:35,821 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:06:35,822 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4239ce1e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:06:35,822 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4c2208{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:06:35,915 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@9f8c8e3{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/java.io.tmpdir/jetty-localhost-43505-hadoop-hdfs-3_4_1-tests_jar-_-any-8594746004713917784/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:06:35,915 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@49793231{HTTP/1.1, (http/1.1)}{localhost:43505} 2024-12-02T01:06:35,915 INFO [Time-limited test {}] server.Server(415): Started @3044ms 2024-12-02T01:06:35,964 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:06:36,069 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:06:36,073 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:06:36,074 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:06:36,075 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:06:36,075 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:06:36,075 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2170f3b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:06:36,076 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@49fab02b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:06:36,173 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@63d8281c{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/java.io.tmpdir/jetty-localhost-40177-hadoop-hdfs-3_4_1-tests_jar-_-any-2730553454713741201/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:06:36,174 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@55b489f0{HTTP/1.1, (http/1.1)}{localhost:40177} 2024-12-02T01:06:36,174 INFO [Time-limited test {}] server.Server(415): Started @3302ms 2024-12-02T01:06:36,176 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:06:37,002 WARN [Thread-98 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data1/current/BP-1841578127-172.17.0.2-1733101594540/current, will proceed with Du for space computation calculation, 2024-12-02T01:06:37,002 WARN [Thread-99 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data2/current/BP-1841578127-172.17.0.2-1733101594540/current, will proceed with Du for space computation calculation, 2024-12-02T01:06:37,002 WARN [Thread-100 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data3/current/BP-1841578127-172.17.0.2-1733101594540/current, will proceed with Du for space computation calculation, 2024-12-02T01:06:37,002 WARN [Thread-101 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data4/current/BP-1841578127-172.17.0.2-1733101594540/current, will proceed with Du for space computation calculation, 2024-12-02T01:06:37,037 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:06:37,037 WARN [Thread-81 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:06:37,083 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdb88af39fe932823 with lease ID 0x22a19b635f3b5549: Processing first storage report for DS-d0537616-f695-40c0-b0ec-af4dc2f1af95 from datanode DatanodeRegistration(127.0.0.1:44461, datanodeUuid=46dad73d-aeaa-4d89-aa02-b0f83b51536f, infoPort=37523, infoSecurePort=0, ipcPort=45299, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540) 2024-12-02T01:06:37,084 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdb88af39fe932823 with lease ID 0x22a19b635f3b5549: from storage DS-d0537616-f695-40c0-b0ec-af4dc2f1af95 node DatanodeRegistration(127.0.0.1:44461, datanodeUuid=46dad73d-aeaa-4d89-aa02-b0f83b51536f, infoPort=37523, infoSecurePort=0, ipcPort=45299, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T01:06:37,084 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9730c96fd50a308f with lease ID 0x22a19b635f3b554a: Processing first storage report for DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95 from datanode DatanodeRegistration(127.0.0.1:38321, datanodeUuid=348ded74-6a6b-4069-9a99-01aa7523fcc1, infoPort=46831, infoSecurePort=0, ipcPort=34767, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540) 2024-12-02T01:06:37,084 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9730c96fd50a308f with lease ID 0x22a19b635f3b554a: from storage DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95 node DatanodeRegistration(127.0.0.1:38321, datanodeUuid=348ded74-6a6b-4069-9a99-01aa7523fcc1, infoPort=46831, infoSecurePort=0, ipcPort=34767, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:06:37,084 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdb88af39fe932823 with lease ID 0x22a19b635f3b5549: Processing first storage report for DS-b3882179-28e1-4139-99a2-8fa58b2c12e5 from datanode DatanodeRegistration(127.0.0.1:44461, datanodeUuid=46dad73d-aeaa-4d89-aa02-b0f83b51536f, infoPort=37523, infoSecurePort=0, ipcPort=45299, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540) 2024-12-02T01:06:37,084 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdb88af39fe932823 with lease ID 0x22a19b635f3b5549: from storage DS-b3882179-28e1-4139-99a2-8fa58b2c12e5 node DatanodeRegistration(127.0.0.1:44461, datanodeUuid=46dad73d-aeaa-4d89-aa02-b0f83b51536f, infoPort=37523, infoSecurePort=0, ipcPort=45299, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T01:06:37,085 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9730c96fd50a308f with lease ID 0x22a19b635f3b554a: Processing first storage report for DS-b5e4058c-abad-4440-b3d0-11d97e909d6c from datanode DatanodeRegistration(127.0.0.1:38321, datanodeUuid=348ded74-6a6b-4069-9a99-01aa7523fcc1, infoPort=46831, infoSecurePort=0, ipcPort=34767, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540) 2024-12-02T01:06:37,085 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9730c96fd50a308f with lease ID 0x22a19b635f3b554a: from storage DS-b5e4058c-abad-4440-b3d0-11d97e909d6c node DatanodeRegistration(127.0.0.1:38321, datanodeUuid=348ded74-6a6b-4069-9a99-01aa7523fcc1, infoPort=46831, infoSecurePort=0, ipcPort=34767, storageInfo=lv=-57;cid=testClusterID;nsid=1104046758;c=1733101594540), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:06:37,181 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44 2024-12-02T01:06:37,240 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/zookeeper_0, clientPort=52426, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T01:06:37,248 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=52426 2024-12-02T01:06:37,256 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:37,258 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:37,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:06:37,453 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:06:37,863 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345 with version=8 2024-12-02T01:06:37,864 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1460): Setting hbase.fs.tmp.dir to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase-staging 2024-12-02T01:06:37,960 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-02T01:06:38,188 INFO [Time-limited test {}] client.ConnectionUtils(129): master/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:06:38,203 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:06:38,203 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:06:38,204 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:06:38,204 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:06:38,204 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:06:38,308 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:06:38,355 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-02T01:06:38,363 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-02T01:06:38,366 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:06:38,387 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 16233 (auto-detected) 2024-12-02T01:06:38,387 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-02T01:06:38,402 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:43051 2024-12-02T01:06:38,409 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:38,410 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:38,424 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:43051 connecting to ZooKeeper ensemble=127.0.0.1:52426 2024-12-02T01:06:38,536 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:430510x0, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:06:38,539 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:43051-0x10194dd549e0000 connected 2024-12-02T01:06:38,637 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:06:38,642 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:06:38,646 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:06:38,652 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43051 2024-12-02T01:06:38,653 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43051 2024-12-02T01:06:38,654 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43051 2024-12-02T01:06:38,654 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43051 2024-12-02T01:06:38,655 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43051 2024-12-02T01:06:38,663 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345, hbase.cluster.distributed=false 2024-12-02T01:06:38,726 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:06:38,727 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:06:38,727 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:06:38,727 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:06:38,727 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:06:38,727 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:06:38,729 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:06:38,731 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:06:38,732 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:35243 2024-12-02T01:06:38,733 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:06:38,738 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:06:38,739 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:38,742 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:38,746 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:35243 connecting to ZooKeeper ensemble=127.0.0.1:52426 2024-12-02T01:06:38,755 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:352430x0, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:06:38,755 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:352430x0, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:06:38,755 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35243-0x10194dd549e0001 connected 2024-12-02T01:06:38,757 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:06:38,758 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:06:38,759 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35243 2024-12-02T01:06:38,759 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35243 2024-12-02T01:06:38,760 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35243 2024-12-02T01:06:38,760 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35243 2024-12-02T01:06:38,761 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35243 2024-12-02T01:06:38,763 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/a82f277fe18c,43051,1733101597955 2024-12-02T01:06:38,776 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:06:38,776 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:06:38,776 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a82f277fe18c:43051 2024-12-02T01:06:38,777 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a82f277fe18c,43051,1733101597955 2024-12-02T01:06:38,804 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:06:38,804 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:06:38,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:38,805 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:38,805 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:06:38,806 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a82f277fe18c,43051,1733101597955 from backup master directory 2024-12-02T01:06:38,806 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:06:38,817 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a82f277fe18c,43051,1733101597955 2024-12-02T01:06:38,817 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:06:38,817 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:06:38,818 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:06:38,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a82f277fe18c,43051,1733101597955 2024-12-02T01:06:38,821 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-02T01:06:38,823 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-02T01:06:38,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:06:38,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:06:38,885 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase.id with ID: be9f829b-578c-4e6d-a8a7-b94929658759 2024-12-02T01:06:38,926 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:38,959 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:38,959 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:38,977 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:06:38,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:06:38,994 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:06:38,996 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T01:06:39,000 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:06:39,026 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:06:39,026 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:06:39,041 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store 2024-12-02T01:06:39,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:06:39,056 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:06:39,061 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-02T01:06:39,062 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:39,063 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:06:39,063 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:06:39,063 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:06:39,063 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:06:39,063 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:06:39,063 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:06:39,063 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:06:39,065 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/.initializing 2024-12-02T01:06:39,066 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/WALs/a82f277fe18c,43051,1733101597955 2024-12-02T01:06:39,082 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C43051%2C1733101597955, suffix=, logDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/WALs/a82f277fe18c,43051,1733101597955, archiveDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/oldWALs, maxLogs=10 2024-12-02T01:06:39,089 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43051%2C1733101597955.1733101599087 2024-12-02T01:06:39,090 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.CommonFSUtils$DfsBuilderUtility(752): Using builder API via reflection for DFS file creation replicate flag. 2024-12-02T01:06:39,090 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.CommonFSUtils$DfsBuilderUtility(762): Using builder API via reflection for DFS file creation noLocalWrite flag. 2024-12-02T01:06:39,105 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/WALs/a82f277fe18c,43051,1733101597955/a82f277fe18c%2C43051%2C1733101597955.1733101599087 2024-12-02T01:06:39,114 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:06:39,114 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:06:39,114 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:39,117 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,118 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,155 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,175 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T01:06:39,179 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,182 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:39,182 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,186 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T01:06:39,186 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,187 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:06:39,187 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,191 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T01:06:39,191 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,192 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:06:39,193 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,195 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T01:06:39,195 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,196 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:06:39,200 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,201 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,209 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T01:06:39,213 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:06:39,217 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:06:39,218 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=799533, jitterRate=0.01665973663330078}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T01:06:39,222 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:06:39,223 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T01:06:39,247 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6cb1af34, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:06:39,274 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-02T01:06:39,283 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T01:06:39,283 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T01:06:39,285 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T01:06:39,286 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 1 msec 2024-12-02T01:06:39,291 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 4 msec 2024-12-02T01:06:39,291 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T01:06:39,313 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T01:06:39,324 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T01:06:39,371 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-02T01:06:39,374 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T01:06:39,376 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T01:06:39,383 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-02T01:06:39,386 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T01:06:39,390 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T01:06:39,400 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-02T01:06:39,402 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T01:06:39,413 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T01:06:39,424 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T01:06:39,434 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T01:06:39,446 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:06:39,446 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:06:39,447 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:39,447 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:39,447 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=a82f277fe18c,43051,1733101597955, sessionid=0x10194dd549e0000, setting cluster-up flag (Was=false) 2024-12-02T01:06:39,471 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:39,471 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:39,496 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T01:06:39,498 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,43051,1733101597955 2024-12-02T01:06:39,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:39,517 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:39,546 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T01:06:39,548 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,43051,1733101597955 2024-12-02T01:06:39,574 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a82f277fe18c:35243 2024-12-02T01:06:39,575 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1008): ClusterId : be9f829b-578c-4e6d-a8a7-b94929658759 2024-12-02T01:06:39,577 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:06:39,589 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:06:39,590 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:06:39,597 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:06:39,598 DEBUG [RS:0;a82f277fe18c:35243 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1ebcb461, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:06:39,599 DEBUG [RS:0;a82f277fe18c:35243 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3cbfb23e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:06:39,602 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:06:39,602 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:06:39,602 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:06:39,604 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,43051,1733101597955 with isa=a82f277fe18c/172.17.0.2:35243, startcode=1733101598726 2024-12-02T01:06:39,615 DEBUG [RS:0;a82f277fe18c:35243 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:06:39,622 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-02T01:06:39,627 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-02T01:06:39,629 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T01:06:39,634 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a82f277fe18c,43051,1733101597955 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T01:06:39,637 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:06:39,637 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:06:39,637 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:06:39,637 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:06:39,638 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a82f277fe18c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T01:06:39,638 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,638 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:06:39,638 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,641 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733101629641 2024-12-02T01:06:39,642 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:06:39,642 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-02T01:06:39,643 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T01:06:39,644 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T01:06:39,645 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,646 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:06:39,648 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T01:06:39,648 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T01:06:39,648 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T01:06:39,649 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T01:06:39,649 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,650 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T01:06:39,651 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T01:06:39,652 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T01:06:39,653 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T01:06:39,653 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T01:06:39,655 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101599655,5,FailOnTimeoutGroup] 2024-12-02T01:06:39,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:06:39,656 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101599656,5,FailOnTimeoutGroup] 2024-12-02T01:06:39,656 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:06:39,656 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T01:06:39,658 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,659 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-02T01:06:39,659 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,659 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345 2024-12-02T01:06:39,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:06:39,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:06:39,670 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:39,672 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:06:39,674 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:06:39,675 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,675 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:39,676 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:06:39,678 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:06:39,678 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,679 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:39,679 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:06:39,682 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:06:39,682 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:39,683 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34447, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:06:39,684 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:39,685 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740 2024-12-02T01:06:39,687 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740 2024-12-02T01:06:39,690 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43051 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,35243,1733101598726 2024-12-02T01:06:39,691 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:06:39,693 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=43051 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:39,694 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:06:39,699 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:06:39,701 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=813604, jitterRate=0.03455115854740143}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:06:39,704 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:06:39,704 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:06:39,704 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:06:39,704 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:06:39,704 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:06:39,705 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:06:39,706 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:06:39,706 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345 2024-12-02T01:06:39,706 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:06:39,706 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:41885 2024-12-02T01:06:39,706 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:06:39,709 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:06:39,709 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-02T01:06:39,714 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T01:06:39,717 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:06:39,718 DEBUG [RS:0;a82f277fe18c:35243 {}] zookeeper.ZKUtil(111): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,35243,1733101598726 2024-12-02T01:06:39,718 WARN [RS:0;a82f277fe18c:35243 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:06:39,718 INFO [RS:0;a82f277fe18c:35243 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:06:39,718 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726 2024-12-02T01:06:39,720 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,35243,1733101598726] 2024-12-02T01:06:39,722 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T01:06:39,724 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T01:06:39,732 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:06:39,742 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:06:39,752 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:06:39,754 INFO [RS:0;a82f277fe18c:35243 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:06:39,755 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,755 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:06:39,761 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,762 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,763 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,763 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,763 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:06:39,763 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:06:39,763 DEBUG [RS:0;a82f277fe18c:35243 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:06:39,764 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,764 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,764 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,764 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,764 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35243,1733101598726-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:06:39,780 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:06:39,781 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35243,1733101598726-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:39,797 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.Replication(204): a82f277fe18c,35243,1733101598726 started 2024-12-02T01:06:39,797 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,35243,1733101598726, RpcServer on a82f277fe18c/172.17.0.2:35243, sessionid=0x10194dd549e0001 2024-12-02T01:06:39,798 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:06:39,798 DEBUG [RS:0;a82f277fe18c:35243 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,35243,1733101598726 2024-12-02T01:06:39,798 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,35243,1733101598726' 2024-12-02T01:06:39,798 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:06:39,799 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:06:39,800 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:06:39,800 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:06:39,800 DEBUG [RS:0;a82f277fe18c:35243 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,35243,1733101598726 2024-12-02T01:06:39,800 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,35243,1733101598726' 2024-12-02T01:06:39,800 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:06:39,800 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:06:39,801 DEBUG [RS:0;a82f277fe18c:35243 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:06:39,801 INFO [RS:0;a82f277fe18c:35243 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:06:39,801 INFO [RS:0;a82f277fe18c:35243 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:06:39,875 WARN [a82f277fe18c:43051 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-02T01:06:39,912 INFO [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C35243%2C1733101598726, suffix=, logDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726, archiveDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs, maxLogs=32 2024-12-02T01:06:39,915 INFO [RS:0;a82f277fe18c:35243 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101599914 2024-12-02T01:06:39,923 INFO [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101599914 2024-12-02T01:06:39,923 DEBUG [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37523:37523),(127.0.0.1/127.0.0.1:46831:46831)] 2024-12-02T01:06:40,128 DEBUG [a82f277fe18c:43051 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T01:06:40,135 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:40,139 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,35243,1733101598726, state=OPENING 2024-12-02T01:06:40,196 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T01:06:40,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:40,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:40,207 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:06:40,207 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:06:40,211 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=a82f277fe18c,35243,1733101598726}] 2024-12-02T01:06:40,388 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35243,1733101598726 2024-12-02T01:06:40,390 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T01:06:40,392 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51364, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T01:06:40,402 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-02T01:06:40,403 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:06:40,406 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C35243%2C1733101598726.meta, suffix=.meta, logDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726, archiveDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs, maxLogs=32 2024-12-02T01:06:40,408 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.meta.1733101600408.meta 2024-12-02T01:06:40,416 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.meta.1733101600408.meta 2024-12-02T01:06:40,416 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:06:40,417 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:06:40,418 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T01:06:40,465 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T01:06:40,468 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T01:06:40,472 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T01:06:40,472 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:40,472 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-02T01:06:40,472 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-02T01:06:40,475 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:06:40,476 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:06:40,476 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:40,477 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:40,477 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:06:40,478 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:06:40,478 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:40,479 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:40,479 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:06:40,480 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:06:40,481 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:40,481 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:06:40,483 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740 2024-12-02T01:06:40,485 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740 2024-12-02T01:06:40,488 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:06:40,490 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:06:40,492 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=849071, jitterRate=0.0796501487493515}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:06:40,493 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:06:40,499 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733101600384 2024-12-02T01:06:40,507 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T01:06:40,508 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-02T01:06:40,509 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:40,510 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,35243,1733101598726, state=OPEN 2024-12-02T01:06:40,540 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:06:40,540 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:06:40,540 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:06:40,540 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:06:40,548 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T01:06:40,548 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=a82f277fe18c,35243,1733101598726 in 329 msec 2024-12-02T01:06:40,555 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T01:06:40,556 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 835 msec 2024-12-02T01:06:40,564 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 977 msec 2024-12-02T01:06:40,564 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733101600564, completionTime=-1 2024-12-02T01:06:40,564 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T01:06:40,564 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-02T01:06:40,595 DEBUG [hconnection-0x5c508872-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:06:40,597 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51368, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:06:40,605 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-02T01:06:40,605 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733101660605 2024-12-02T01:06:40,605 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733101720605 2024-12-02T01:06:40,605 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 40 msec 2024-12-02T01:06:40,646 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43051,1733101597955-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:40,646 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43051,1733101597955-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:40,646 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43051,1733101597955-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:40,648 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a82f277fe18c:43051, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:40,649 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:40,655 DEBUG [master/a82f277fe18c:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-02T01:06:40,657 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-02T01:06:40,658 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:06:40,663 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-02T01:06:40,666 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:06:40,667 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:40,669 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:06:40,680 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:06:40,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:06:40,684 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 6fb7adfbc00bb9ee301414b06eb6b14a, NAME => 'hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345 2024-12-02T01:06:40,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:06:40,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:06:40,696 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:40,697 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 6fb7adfbc00bb9ee301414b06eb6b14a, disabling compactions & flushes 2024-12-02T01:06:40,697 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:40,697 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:40,697 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. after waiting 0 ms 2024-12-02T01:06:40,697 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:40,697 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:40,697 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 6fb7adfbc00bb9ee301414b06eb6b14a: 2024-12-02T01:06:40,700 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:06:40,707 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733101600701"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101600701"}]},"ts":"1733101600701"} 2024-12-02T01:06:40,729 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:06:40,732 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:06:40,734 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101600732"}]},"ts":"1733101600732"} 2024-12-02T01:06:40,739 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-02T01:06:40,760 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=6fb7adfbc00bb9ee301414b06eb6b14a, ASSIGN}] 2024-12-02T01:06:40,763 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=6fb7adfbc00bb9ee301414b06eb6b14a, ASSIGN 2024-12-02T01:06:40,765 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=6fb7adfbc00bb9ee301414b06eb6b14a, ASSIGN; state=OFFLINE, location=a82f277fe18c,35243,1733101598726; forceNewPlan=false, retain=false 2024-12-02T01:06:40,916 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=6fb7adfbc00bb9ee301414b06eb6b14a, regionState=OPENING, regionLocation=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:40,924 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 6fb7adfbc00bb9ee301414b06eb6b14a, server=a82f277fe18c,35243,1733101598726}] 2024-12-02T01:06:41,081 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35243,1733101598726 2024-12-02T01:06:41,091 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:41,091 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 6fb7adfbc00bb9ee301414b06eb6b14a, NAME => 'hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:06:41,092 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,092 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:41,092 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,093 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,095 INFO [StoreOpener-6fb7adfbc00bb9ee301414b06eb6b14a-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,097 INFO [StoreOpener-6fb7adfbc00bb9ee301414b06eb6b14a-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6fb7adfbc00bb9ee301414b06eb6b14a columnFamilyName info 2024-12-02T01:06:41,097 DEBUG [StoreOpener-6fb7adfbc00bb9ee301414b06eb6b14a-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:41,098 INFO [StoreOpener-6fb7adfbc00bb9ee301414b06eb6b14a-1 {}] regionserver.HStore(327): Store=6fb7adfbc00bb9ee301414b06eb6b14a/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:06:41,100 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,101 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,105 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:06:41,108 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:06:41,109 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 6fb7adfbc00bb9ee301414b06eb6b14a; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=703904, jitterRate=-0.10494031012058258}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:06:41,110 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 6fb7adfbc00bb9ee301414b06eb6b14a: 2024-12-02T01:06:41,112 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a., pid=6, masterSystemTime=1733101601081 2024-12-02T01:06:41,115 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:41,115 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:06:41,116 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=6fb7adfbc00bb9ee301414b06eb6b14a, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:41,125 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T01:06:41,126 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 6fb7adfbc00bb9ee301414b06eb6b14a, server=a82f277fe18c,35243,1733101598726 in 197 msec 2024-12-02T01:06:41,128 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T01:06:41,128 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=6fb7adfbc00bb9ee301414b06eb6b14a, ASSIGN in 365 msec 2024-12-02T01:06:41,130 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:06:41,130 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101601130"}]},"ts":"1733101601130"} 2024-12-02T01:06:41,133 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-02T01:06:41,144 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:06:41,148 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 485 msec 2024-12-02T01:06:41,168 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-02T01:06:41,238 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:06:41,238 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:41,238 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:06:41,270 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-02T01:06:41,292 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:06:41,307 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 37 msec 2024-12-02T01:06:41,316 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-02T01:06:41,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:06:41,356 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 38 msec 2024-12-02T01:06:41,379 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-02T01:06:41,396 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-02T01:06:41,396 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 2.577sec 2024-12-02T01:06:41,399 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T01:06:41,401 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T01:06:41,403 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T01:06:41,404 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T01:06:41,404 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T01:06:41,406 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43051,1733101597955-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:06:41,407 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43051,1733101597955-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T01:06:41,414 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-02T01:06:41,415 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T01:06:41,415 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43051,1733101597955-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:06:41,480 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x262c14ad to 127.0.0.1:52426 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@744a1e4d 2024-12-02T01:06:41,480 WARN [Time-limited test {}] client.ZKConnectionRegistry(90): ZKConnectionRegistry is deprecated. See https://hbase.apache.org/book.html#client.rpcconnectionregistry 2024-12-02T01:06:41,493 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@18645a9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:06:41,495 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-02T01:06:41,495 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-02T01:06:41,522 DEBUG [hconnection-0x61089d20-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:06:41,561 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51374, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:06:41,570 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=a82f277fe18c,43051,1733101597955 2024-12-02T01:06:41,571 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:06:41,576 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-02T01:06:41,581 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T01:06:41,584 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45788, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T01:06:41,589 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T01:06:41,589 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T01:06:41,592 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testSlowSyncLogRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:06:41,594 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling 2024-12-02T01:06:41,596 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:06:41,596 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testSlowSyncLogRolling" procId is: 9 2024-12-02T01:06:41,596 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:41,598 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:06:41,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:06:41,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741837_1013 (size=389) 2024-12-02T01:06:41,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741837_1013 (size=389) 2024-12-02T01:06:41,610 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 43914f779250d48ca76badde3ab4efdf, NAME => 'TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testSlowSyncLogRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345 2024-12-02T01:06:41,618 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741838_1014 (size=72) 2024-12-02T01:06:41,618 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741838_1014 (size=72) 2024-12-02T01:06:41,619 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:41,620 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1681): Closing 43914f779250d48ca76badde3ab4efdf, disabling compactions & flushes 2024-12-02T01:06:41,620 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,620 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,620 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. after waiting 0 ms 2024-12-02T01:06:41,620 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,620 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,620 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1635): Region close journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:06:41,622 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:06:41,622 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.","families":{"info":[{"qualifier":"regioninfo","vlen":71,"tag":[],"timestamp":"1733101601622"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101601622"}]},"ts":"1733101601622"} 2024-12-02T01:06:41,625 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:06:41,626 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:06:41,626 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testSlowSyncLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101601626"}]},"ts":"1733101601626"} 2024-12-02T01:06:41,629 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testSlowSyncLogRolling, state=ENABLING in hbase:meta 2024-12-02T01:06:41,646 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=43914f779250d48ca76badde3ab4efdf, ASSIGN}] 2024-12-02T01:06:41,648 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=43914f779250d48ca76badde3ab4efdf, ASSIGN 2024-12-02T01:06:41,650 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=43914f779250d48ca76badde3ab4efdf, ASSIGN; state=OFFLINE, location=a82f277fe18c,35243,1733101598726; forceNewPlan=false, retain=false 2024-12-02T01:06:41,801 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=43914f779250d48ca76badde3ab4efdf, regionState=OPENING, regionLocation=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:41,810 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 43914f779250d48ca76badde3ab4efdf, server=a82f277fe18c,35243,1733101598726}] 2024-12-02T01:06:41,966 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35243,1733101598726 2024-12-02T01:06:41,976 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,977 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 43914f779250d48ca76badde3ab4efdf, NAME => 'TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:06:41,977 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testSlowSyncLogRolling 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,977 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:06:41,977 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,977 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,980 INFO [StoreOpener-43914f779250d48ca76badde3ab4efdf-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,982 INFO [StoreOpener-43914f779250d48ca76badde3ab4efdf-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 43914f779250d48ca76badde3ab4efdf columnFamilyName info 2024-12-02T01:06:41,982 DEBUG [StoreOpener-43914f779250d48ca76badde3ab4efdf-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:06:41,984 INFO [StoreOpener-43914f779250d48ca76badde3ab4efdf-1 {}] regionserver.HStore(327): Store=43914f779250d48ca76badde3ab4efdf/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:06:41,985 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,986 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,989 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:06:41,993 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:06:41,994 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened 43914f779250d48ca76badde3ab4efdf; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=777740, jitterRate=-0.011052995920181274}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:06:41,994 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:06:41,996 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf., pid=11, masterSystemTime=1733101601966 2024-12-02T01:06:41,998 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,999 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:41,999 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=43914f779250d48ca76badde3ab4efdf, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,35243,1733101598726 2024-12-02T01:06:42,005 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-02T01:06:42,006 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 43914f779250d48ca76badde3ab4efdf, server=a82f277fe18c,35243,1733101598726 in 192 msec 2024-12-02T01:06:42,008 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-02T01:06:42,008 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=43914f779250d48ca76badde3ab4efdf, ASSIGN in 359 msec 2024-12-02T01:06:42,009 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:06:42,009 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testSlowSyncLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101602009"}]},"ts":"1733101602009"} 2024-12-02T01:06:42,011 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testSlowSyncLogRolling, state=ENABLED in hbase:meta 2024-12-02T01:06:42,057 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:06:42,062 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling in 465 msec 2024-12-02T01:06:45,978 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-hbase.properties,hadoop-metrics2.properties 2024-12-02T01:06:46,028 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T01:06:46,029 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-02T01:06:46,030 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testSlowSyncLogRolling' 2024-12-02T01:06:48,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-02T01:06:48,354 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-02T01:06:48,357 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling 2024-12-02T01:06:48,357 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling Metrics about Tables on a single HBase RegionServer 2024-12-02T01:06:48,359 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:06:48,359 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2024-12-02T01:06:48,361 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T01:06:48,362 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-02T01:06:48,362 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-02T01:06:48,362 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-02T01:06:51,608 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43051 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:06:51,609 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testSlowSyncLogRolling, procId: 9 completed 2024-12-02T01:06:51,616 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testSlowSyncLogRolling 2024-12-02T01:06:51,617 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:06:51,618 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101611618 2024-12-02T01:06:51,630 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101599914 with entries=4, filesize=947 B; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101611618 2024-12-02T01:06:51,631 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37523:37523),(127.0.0.1/127.0.0.1:46831:46831)] 2024-12-02T01:06:51,631 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101599914 is not closed yet, will try archiving it next time 2024-12-02T01:06:51,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741833_1009 (size=955) 2024-12-02T01:06:51,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741833_1009 (size=955) 2024-12-02T01:07:03,682 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35243 {}] regionserver.HRegion(8581): Flush requested on 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:07:03,683 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 43914f779250d48ca76badde3ab4efdf 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:07:03,741 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/4ea9f99a0f6743a693b8b0e3c74fba91 is 1080, key is row0001/info:/1733101611637/Put/seqid=0 2024-12-02T01:07:03,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741840_1016 (size=12509) 2024-12-02T01:07:03,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741840_1016 (size=12509) 2024-12-02T01:07:03,753 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/4ea9f99a0f6743a693b8b0e3c74fba91 2024-12-02T01:07:03,793 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/4ea9f99a0f6743a693b8b0e3c74fba91 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91 2024-12-02T01:07:03,803 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91, entries=7, sequenceid=11, filesize=12.2 K 2024-12-02T01:07:03,805 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 43914f779250d48ca76badde3ab4efdf in 123ms, sequenceid=11, compaction requested=false 2024-12-02T01:07:03,806 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:07:07,179 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:07:11,604 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:07:11,607 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52326, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:07:11,698 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101631697 2024-12-02T01:07:11,913 INFO [Time-limited test {}] wal.AbstractFSWAL(1183): Slow sync cost: 211 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:11,915 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101611618 with entries=12, filesize=12.10 KB; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101631697 2024-12-02T01:07:11,915 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37523:37523),(127.0.0.1/127.0.0.1:46831:46831)] 2024-12-02T01:07:11,915 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101611618 is not closed yet, will try archiving it next time 2024-12-02T01:07:11,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741839_1015 (size=12399) 2024-12-02T01:07:11,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741839_1015 (size=12399) 2024-12-02T01:07:12,119 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:14,326 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:16,530 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:18,737 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 202 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:18,738 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35243 {}] regionserver.HRegion(8581): Flush requested on 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:07:18,738 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 43914f779250d48ca76badde3ab4efdf 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:07:18,942 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 202 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:18,952 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/f49837702c9f43a1a2ea26f108f4feff is 1080, key is row0008/info:/1733101625686/Put/seqid=0 2024-12-02T01:07:18,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741842_1018 (size=12509) 2024-12-02T01:07:18,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741842_1018 (size=12509) 2024-12-02T01:07:18,961 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=21 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/f49837702c9f43a1a2ea26f108f4feff 2024-12-02T01:07:18,972 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/f49837702c9f43a1a2ea26f108f4feff as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/f49837702c9f43a1a2ea26f108f4feff 2024-12-02T01:07:18,983 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/f49837702c9f43a1a2ea26f108f4feff, entries=7, sequenceid=21, filesize=12.2 K 2024-12-02T01:07:19,185 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:19,185 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 43914f779250d48ca76badde3ab4efdf in 447ms, sequenceid=21, compaction requested=false 2024-12-02T01:07:19,186 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:07:19,186 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=24.4 K, sizeToCheck=16.0 K 2024-12-02T01:07:19,186 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:07:19,188 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91 because midkey is the same as first or last row 2024-12-02T01:07:20,943 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:21,481 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T01:07:21,481 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T01:07:23,149 WARN [sync.1 {}] wal.AbstractFSWAL(1346): Requesting log roll because we exceeded slow sync threshold; count=7, threshold=5, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:23,153 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C35243%2C1733101598726:(num 1733101631697) roll requested 2024-12-02T01:07:23,153 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 205 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:23,154 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101643153 2024-12-02T01:07:23,368 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(1183): Slow sync cost: 212 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:23,569 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:23,570 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101631697 with entries=8, filesize=7.55 KB; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101643153 2024-12-02T01:07:23,570 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:07:23,570 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101631697 is not closed yet, will try archiving it next time 2024-12-02T01:07:23,572 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101611618 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101611618 2024-12-02T01:07:23,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741841_1017 (size=7739) 2024-12-02T01:07:23,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741841_1017 (size=7739) 2024-12-02T01:07:25,354 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:26,978 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 43914f779250d48ca76badde3ab4efdf, had cached 0 bytes from a total of 25018 2024-12-02T01:07:27,560 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 202 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:29,765 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:31,970 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 202 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:33,975 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T01:07:33,976 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101653975 2024-12-02T01:07:37,179 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:07:38,990 INFO [Time-limited test {}] wal.AbstractFSWAL(1183): Slow sync cost: 5009 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:38,990 WARN [Time-limited test {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5009 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:38,990 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C35243%2C1733101598726:(num 1733101653975) roll requested 2024-12-02T01:07:40,702 DEBUG [master/a82f277fe18c:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 6fb7adfbc00bb9ee301414b06eb6b14a changed from -1.0 to 0.0, refreshing cache 2024-12-02T01:07:43,991 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:43,991 WARN [sync.2 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:07:43,992 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101643153 with entries=4, filesize=4.63 KB; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101653975 2024-12-02T01:07:43,992 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37523:37523),(127.0.0.1/127.0.0.1:46831:46831)] 2024-12-02T01:07:43,993 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101643153 is not closed yet, will try archiving it next time 2024-12-02T01:07:43,994 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101663993 2024-12-02T01:07:44,103 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741843_1019 (size=4753) 2024-12-02T01:07:44,103 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741843_1019 (size=4753) 2024-12-02T01:07:48,998 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5002 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:48,998 WARN [sync.3 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5002 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:48,999 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35243 {}] regionserver.HRegion(8581): Flush requested on 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:07:49,000 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 43914f779250d48ca76badde3ab4efdf 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:07:49,109 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(1183): Slow sync cost: 5010 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:49,110 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5010 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:51,000 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T01:07:54,003 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5001 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:54,003 WARN [sync.4 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5001 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:54,010 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/9e1141206d954ccdb3997c193a1a3705 is 1080, key is row0015/info:/1733101640741/Put/seqid=0 2024-12-02T01:07:54,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741846_1022 (size=12509) 2024-12-02T01:07:54,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741846_1022 (size=12509) 2024-12-02T01:07:54,017 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=31 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/9e1141206d954ccdb3997c193a1a3705 2024-12-02T01:07:54,026 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/9e1141206d954ccdb3997c193a1a3705 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/9e1141206d954ccdb3997c193a1a3705 2024-12-02T01:07:54,033 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/9e1141206d954ccdb3997c193a1a3705, entries=7, sequenceid=31, filesize=12.2 K 2024-12-02T01:07:54,110 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:54,110 WARN [sync.0 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:54,111 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101653975 with entries=2, filesize=1.52 KB; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101663993 2024-12-02T01:07:54,111 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37523:37523),(127.0.0.1/127.0.0.1:46831:46831)] 2024-12-02T01:07:54,112 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101653975 is not closed yet, will try archiving it next time 2024-12-02T01:07:54,112 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C35243%2C1733101598726:(num 1733101663993) roll requested 2024-12-02T01:07:54,112 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101674112 2024-12-02T01:07:54,115 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741844_1020 (size=1569) 2024-12-02T01:07:54,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741844_1020 (size=1569) 2024-12-02T01:07:59,113 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:59,113 WARN [sync.1 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:59,113 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 43914f779250d48ca76badde3ab4efdf in 10114ms, sequenceid=31, compaction requested=true 2024-12-02T01:07:59,113 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:07:59,113 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=36.6 K, sizeToCheck=16.0 K 2024-12-02T01:07:59,113 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:07:59,113 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91 because midkey is the same as first or last row 2024-12-02T01:07:59,115 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 43914f779250d48ca76badde3ab4efdf:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:07:59,115 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:07:59,116 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:07:59,120 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 37527 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:07:59,122 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(1183): Slow sync cost: 5006 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:59,122 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.HStore(1540): 43914f779250d48ca76badde3ab4efdf/info is initiating minor compaction (all files) 2024-12-02T01:07:59,122 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5006 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:07:59,123 INFO [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 43914f779250d48ca76badde3ab4efdf/info in TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:07:59,123 INFO [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/f49837702c9f43a1a2ea26f108f4feff, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/9e1141206d954ccdb3997c193a1a3705] into tmpdir=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp, totalSize=36.6 K 2024-12-02T01:07:59,125 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] compactions.Compactor(224): Compacting 4ea9f99a0f6743a693b8b0e3c74fba91, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733101611637 2024-12-02T01:07:59,126 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] compactions.Compactor(224): Compacting f49837702c9f43a1a2ea26f108f4feff, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=21, earliestPutTs=1733101625686 2024-12-02T01:07:59,127 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] compactions.Compactor(224): Compacting 9e1141206d954ccdb3997c193a1a3705, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=31, earliestPutTs=1733101640741 2024-12-02T01:07:59,158 INFO [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 43914f779250d48ca76badde3ab4efdf#info#compaction#3 average throughput is 10.77 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:07:59,159 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/a50067b3212c428bbbc824fe59874b8c is 1080, key is row0001/info:/1733101611637/Put/seqid=0 2024-12-02T01:07:59,171 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741848_1024 (size=27710) 2024-12-02T01:07:59,171 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741848_1024 (size=27710) 2024-12-02T01:07:59,182 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/a50067b3212c428bbbc824fe59874b8c as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/a50067b3212c428bbbc824fe59874b8c 2024-12-02T01:08:04,123 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:08:04,123 WARN [sync.2 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK], DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK]] 2024-12-02T01:08:04,124 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101663993 with entries=1, filesize=430 B; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101674112 2024-12-02T01:08:04,125 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:08:04,125 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101663993 is not closed yet, will try archiving it next time 2024-12-02T01:08:04,126 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101631697 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101631697 2024-12-02T01:08:04,126 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C35243%2C1733101598726:(num 1733101684126) roll requested 2024-12-02T01:08:04,127 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101684126 2024-12-02T01:08:04,132 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101643153 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101643153 2024-12-02T01:08:04,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741845_1021 (size=438) 2024-12-02T01:08:04,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741845_1021 (size=438) 2024-12-02T01:08:04,134 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101653975 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101653975 2024-12-02T01:08:04,136 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101663993 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101663993 2024-12-02T01:08:07,180 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:08:09,127 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5001 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:08:09,127 WARN [sync.3 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5001 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:08:09,129 INFO [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 43914f779250d48ca76badde3ab4efdf/info of 43914f779250d48ca76badde3ab4efdf into a50067b3212c428bbbc824fe59874b8c(size=27.1 K), total size for store is 27.1 K. This selection was in queue for 0sec, and took 10sec to execute. 2024-12-02T01:08:09,129 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:08:09,129 INFO [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf., storeName=43914f779250d48ca76badde3ab4efdf/info, priority=13, startTime=1733101679115; duration=10sec 2024-12-02T01:08:09,130 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=27.1 K, sizeToCheck=16.0 K 2024-12-02T01:08:09,130 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:08:09,130 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/a50067b3212c428bbbc824fe59874b8c because midkey is the same as first or last row 2024-12-02T01:08:09,130 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:08:09,131 DEBUG [RS:0;a82f277fe18c:35243-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 43914f779250d48ca76badde3ab4efdf:info 2024-12-02T01:08:09,139 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:08:09,139 WARN [sync.4 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:38321,DS-3f50df00-a2d7-4a74-ae06-82ad1c242c95,DISK], DatanodeInfoWithStorage[127.0.0.1:44461,DS-d0537616-f695-40c0-b0ec-af4dc2f1af95,DISK]] 2024-12-02T01:08:09,139 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101674112 with entries=1, filesize=531 B; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101684126 2024-12-02T01:08:09,140 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:08:09,140 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101674112 is not closed yet, will try archiving it next time 2024-12-02T01:08:09,140 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101689140 2024-12-02T01:08:09,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741847_1023 (size=539) 2024-12-02T01:08:09,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741847_1023 (size=539) 2024-12-02T01:08:09,177 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101674112 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101674112 2024-12-02T01:08:09,183 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101684126 with entries=1, filesize=1.22 KB; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101689140 2024-12-02T01:08:09,183 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:08:09,183 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101684126 is not closed yet, will try archiving it next time 2024-12-02T01:08:09,183 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C35243%2C1733101598726:(num 1733101689140) roll requested 2024-12-02T01:08:09,183 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35243%2C1733101598726.1733101689183 2024-12-02T01:08:09,185 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741849_1025 (size=1258) 2024-12-02T01:08:09,185 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741849_1025 (size=1258) 2024-12-02T01:08:09,193 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101689140 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101689183 2024-12-02T01:08:09,193 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46831:46831),(127.0.0.1/127.0.0.1:37523:37523)] 2024-12-02T01:08:09,193 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101689140 is not closed yet, will try archiving it next time 2024-12-02T01:08:09,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741850_1026 (size=93) 2024-12-02T01:08:09,195 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741850_1026 (size=93) 2024-12-02T01:08:09,196 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726/a82f277fe18c%2C35243%2C1733101598726.1733101689140 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs/a82f277fe18c%2C35243%2C1733101598726.1733101689140 2024-12-02T01:08:11,979 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 43914f779250d48ca76badde3ab4efdf, had cached 0 bytes from a total of 27710 2024-12-02T01:08:21,204 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35243 {}] regionserver.HRegion(8581): Flush requested on 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:08:21,205 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 43914f779250d48ca76badde3ab4efdf 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:08:21,215 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/cf475d6a1fa94c35baf869e3e538a083 is 1080, key is row0022/info:/1733101689142/Put/seqid=0 2024-12-02T01:08:21,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741852_1028 (size=12509) 2024-12-02T01:08:21,222 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741852_1028 (size=12509) 2024-12-02T01:08:21,223 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=42 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/cf475d6a1fa94c35baf869e3e538a083 2024-12-02T01:08:21,233 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/cf475d6a1fa94c35baf869e3e538a083 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/cf475d6a1fa94c35baf869e3e538a083 2024-12-02T01:08:21,240 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/cf475d6a1fa94c35baf869e3e538a083, entries=7, sequenceid=42, filesize=12.2 K 2024-12-02T01:08:21,241 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 43914f779250d48ca76badde3ab4efdf in 37ms, sequenceid=42, compaction requested=false 2024-12-02T01:08:21,241 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:08:21,242 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=39.3 K, sizeToCheck=16.0 K 2024-12-02T01:08:21,242 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:08:21,242 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/a50067b3212c428bbbc824fe59874b8c because midkey is the same as first or last row 2024-12-02T01:08:29,219 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-02T01:08:29,220 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-02T01:08:29,221 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x262c14ad to 127.0.0.1:52426 2024-12-02T01:08:29,221 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:08:29,223 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T01:08:29,223 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1741889779, stopped=false 2024-12-02T01:08:29,224 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=a82f277fe18c,43051,1733101597955 2024-12-02T01:08:29,268 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:08:29,268 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:08:29,268 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:29,268 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:29,268 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-02T01:08:29,269 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:08:29,270 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,35243,1733101598726' ***** 2024-12-02T01:08:29,270 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:08:29,270 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:08:29,271 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:08:29,271 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:08:29,272 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:08:29,272 INFO [RS:0;a82f277fe18c:35243 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:08:29,272 INFO [RS:0;a82f277fe18c:35243 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:08:29,273 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(3579): Received CLOSE for 43914f779250d48ca76badde3ab4efdf 2024-12-02T01:08:29,274 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(3579): Received CLOSE for 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:08:29,275 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,35243,1733101598726 2024-12-02T01:08:29,275 DEBUG [RS:0;a82f277fe18c:35243 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:08:29,275 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:08:29,276 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:08:29,276 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:08:29,276 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 43914f779250d48ca76badde3ab4efdf, disabling compactions & flushes 2024-12-02T01:08:29,276 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:08:29,276 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:08:29,276 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:08:29,276 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. after waiting 0 ms 2024-12-02T01:08:29,277 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:08:29,277 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-02T01:08:29,277 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 43914f779250d48ca76badde3ab4efdf 1/1 column families, dataSize=3.15 KB heapSize=3.63 KB 2024-12-02T01:08:29,277 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1603): Online Regions={43914f779250d48ca76badde3ab4efdf=TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf., 1588230740=hbase:meta,,1.1588230740, 6fb7adfbc00bb9ee301414b06eb6b14a=hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a.} 2024-12-02T01:08:29,277 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:08:29,277 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:08:29,278 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:08:29,278 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:08:29,278 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:08:29,278 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.81 KB heapSize=5.32 KB 2024-12-02T01:08:29,278 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 43914f779250d48ca76badde3ab4efdf, 6fb7adfbc00bb9ee301414b06eb6b14a 2024-12-02T01:08:29,282 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/c5f959b3bb7640668c35355287e8edea is 1080, key is row0029/info:/1733101703208/Put/seqid=0 2024-12-02T01:08:29,287 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741853_1029 (size=8193) 2024-12-02T01:08:29,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741853_1029 (size=8193) 2024-12-02T01:08:29,288 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=3.15 KB at sequenceid=48 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/c5f959b3bb7640668c35355287e8edea 2024-12-02T01:08:29,297 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/.tmp/info/c5f959b3bb7640668c35355287e8edea as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/c5f959b3bb7640668c35355287e8edea 2024-12-02T01:08:29,299 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/.tmp/info/d31a3538aef44a8e8570aa5d807f6ed2 is 195, key is TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf./info:regioninfo/1733101601999/Put/seqid=0 2024-12-02T01:08:29,305 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741854_1030 (size=8172) 2024-12-02T01:08:29,305 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/c5f959b3bb7640668c35355287e8edea, entries=3, sequenceid=48, filesize=8.0 K 2024-12-02T01:08:29,305 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741854_1030 (size=8172) 2024-12-02T01:08:29,306 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.59 KB at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/.tmp/info/d31a3538aef44a8e8570aa5d807f6ed2 2024-12-02T01:08:29,306 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~3.15 KB/3228, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 43914f779250d48ca76badde3ab4efdf in 29ms, sequenceid=48, compaction requested=true 2024-12-02T01:08:29,307 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/f49837702c9f43a1a2ea26f108f4feff, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/9e1141206d954ccdb3997c193a1a3705] to archive 2024-12-02T01:08:29,310 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T01:08:29,313 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/archive/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/4ea9f99a0f6743a693b8b0e3c74fba91 2024-12-02T01:08:29,314 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/f49837702c9f43a1a2ea26f108f4feff to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/archive/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/f49837702c9f43a1a2ea26f108f4feff 2024-12-02T01:08:29,316 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/9e1141206d954ccdb3997c193a1a3705 to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/archive/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/info/9e1141206d954ccdb3997c193a1a3705 2024-12-02T01:08:29,326 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/.tmp/table/d2b43c3e2bd6496a8ec8b627827f8aae is 73, key is TestLogRolling-testSlowSyncLogRolling/table:state/1733101602009/Put/seqid=0 2024-12-02T01:08:29,331 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/default/TestLogRolling-testSlowSyncLogRolling/43914f779250d48ca76badde3ab4efdf/recovered.edits/51.seqid, newMaxSeqId=51, maxSeqId=1 2024-12-02T01:08:29,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741855_1031 (size=5452) 2024-12-02T01:08:29,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741855_1031 (size=5452) 2024-12-02T01:08:29,333 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:08:29,333 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 43914f779250d48ca76badde3ab4efdf: 2024-12-02T01:08:29,333 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testSlowSyncLogRolling,,1733101601589.43914f779250d48ca76badde3ab4efdf. 2024-12-02T01:08:29,334 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 6fb7adfbc00bb9ee301414b06eb6b14a, disabling compactions & flushes 2024-12-02T01:08:29,334 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:08:29,334 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:08:29,334 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. after waiting 0 ms 2024-12-02T01:08:29,334 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:08:29,334 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 6fb7adfbc00bb9ee301414b06eb6b14a 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-02T01:08:29,351 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/.tmp/info/fae8882b62eb4a5d98194772db2523e0 is 45, key is default/info:d/1733101601280/Put/seqid=0 2024-12-02T01:08:29,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741856_1032 (size=5037) 2024-12-02T01:08:29,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741856_1032 (size=5037) 2024-12-02T01:08:29,357 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/.tmp/info/fae8882b62eb4a5d98194772db2523e0 2024-12-02T01:08:29,364 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/.tmp/info/fae8882b62eb4a5d98194772db2523e0 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/info/fae8882b62eb4a5d98194772db2523e0 2024-12-02T01:08:29,371 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/info/fae8882b62eb4a5d98194772db2523e0, entries=2, sequenceid=6, filesize=4.9 K 2024-12-02T01:08:29,372 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 6fb7adfbc00bb9ee301414b06eb6b14a in 38ms, sequenceid=6, compaction requested=false 2024-12-02T01:08:29,377 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/namespace/6fb7adfbc00bb9ee301414b06eb6b14a/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T01:08:29,378 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:08:29,378 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 6fb7adfbc00bb9ee301414b06eb6b14a: 2024-12-02T01:08:29,379 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733101600658.6fb7adfbc00bb9ee301414b06eb6b14a. 2024-12-02T01:08:29,479 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1629): Waiting on 1588230740 2024-12-02T01:08:29,679 DEBUG [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1629): Waiting on 1588230740 2024-12-02T01:08:29,735 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=232 B at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/.tmp/table/d2b43c3e2bd6496a8ec8b627827f8aae 2024-12-02T01:08:29,752 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/.tmp/info/d31a3538aef44a8e8570aa5d807f6ed2 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/info/d31a3538aef44a8e8570aa5d807f6ed2 2024-12-02T01:08:29,760 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/info/d31a3538aef44a8e8570aa5d807f6ed2, entries=20, sequenceid=14, filesize=8.0 K 2024-12-02T01:08:29,762 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/.tmp/table/d2b43c3e2bd6496a8ec8b627827f8aae as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/table/d2b43c3e2bd6496a8ec8b627827f8aae 2024-12-02T01:08:29,769 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/table/d2b43c3e2bd6496a8ec8b627827f8aae, entries=4, sequenceid=14, filesize=5.3 K 2024-12-02T01:08:29,771 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~2.81 KB/2882, heapSize ~5.04 KB/5160, currentSize=0 B/0 for 1588230740 in 492ms, sequenceid=14, compaction requested=false 2024-12-02T01:08:29,772 INFO [regionserver/a82f277fe18c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:08:29,776 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/data/hbase/meta/1588230740/recovered.edits/17.seqid, newMaxSeqId=17, maxSeqId=1 2024-12-02T01:08:29,776 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T01:08:29,777 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:08:29,777 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:08:29,777 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T01:08:29,809 INFO [regionserver/a82f277fe18c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T01:08:29,810 INFO [regionserver/a82f277fe18c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T01:08:29,879 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,35243,1733101598726; all regions closed. 2024-12-02T01:08:29,880 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726 2024-12-02T01:08:29,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741834_1010 (size=4330) 2024-12-02T01:08:29,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741834_1010 (size=4330) 2024-12-02T01:08:29,885 DEBUG [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs 2024-12-02T01:08:29,885 INFO [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C35243%2C1733101598726.meta:.meta(num 1733101600408) 2024-12-02T01:08:29,885 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/WALs/a82f277fe18c,35243,1733101598726 2024-12-02T01:08:29,888 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741851_1027 (size=13066) 2024-12-02T01:08:29,888 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741851_1027 (size=13066) 2024-12-02T01:08:30,295 DEBUG [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(1071): Moved 3 WAL file(s) to /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/oldWALs 2024-12-02T01:08:30,295 INFO [RS:0;a82f277fe18c:35243 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C35243%2C1733101598726:(num 1733101689183) 2024-12-02T01:08:30,295 DEBUG [RS:0;a82f277fe18c:35243 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:08:30,295 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:08:30,295 INFO [RS:0;a82f277fe18c:35243 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-02T01:08:30,296 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:08:30,296 INFO [RS:0;a82f277fe18c:35243 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:35243 2024-12-02T01:08:30,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,35243,1733101598726 2024-12-02T01:08:30,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:08:30,343 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,35243,1733101598726] 2024-12-02T01:08:30,343 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,35243,1733101598726; numProcessing=1 2024-12-02T01:08:30,350 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,35243,1733101598726 already deleted, retry=false 2024-12-02T01:08:30,350 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,35243,1733101598726 expired; onlineServers=0 2024-12-02T01:08:30,350 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,43051,1733101597955' ***** 2024-12-02T01:08:30,350 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T01:08:30,351 DEBUG [M:0;a82f277fe18c:43051 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@a99cf3d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:08:30,351 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,43051,1733101597955 2024-12-02T01:08:30,351 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,43051,1733101597955; all regions closed. 2024-12-02T01:08:30,351 DEBUG [M:0;a82f277fe18c:43051 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:08:30,351 DEBUG [M:0;a82f277fe18c:43051 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T01:08:30,351 DEBUG [M:0;a82f277fe18c:43051 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T01:08:30,351 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T01:08:30,351 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101599656 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101599656,5,FailOnTimeoutGroup] 2024-12-02T01:08:30,351 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101599655 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101599655,5,FailOnTimeoutGroup] 2024-12-02T01:08:30,351 INFO [M:0;a82f277fe18c:43051 {}] hbase.ChoreService(370): Chore service for: master/a82f277fe18c:0 had [] on shutdown 2024-12-02T01:08:30,352 DEBUG [M:0;a82f277fe18c:43051 {}] master.HMaster(1733): Stopping service threads 2024-12-02T01:08:30,352 INFO [M:0;a82f277fe18c:43051 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T01:08:30,352 INFO [M:0;a82f277fe18c:43051 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T01:08:30,352 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T01:08:30,359 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T01:08:30,359 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:30,359 DEBUG [M:0;a82f277fe18c:43051 {}] zookeeper.ZKUtil(347): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T01:08:30,359 WARN [M:0;a82f277fe18c:43051 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T01:08:30,359 INFO [M:0;a82f277fe18c:43051 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-02T01:08:30,359 INFO [M:0;a82f277fe18c:43051 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T01:08:30,359 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:08:30,359 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:30,359 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:08:30,359 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:30,360 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:08:30,360 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:30,360 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=40.21 KB heapSize=50.14 KB 2024-12-02T01:08:30,378 DEBUG [M:0;a82f277fe18c:43051 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/7fdfdd7f5a0649cc8611e3f045d67af4 is 82, key is hbase:meta,,1/info:regioninfo/1733101600508/Put/seqid=0 2024-12-02T01:08:30,383 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741857_1033 (size=5672) 2024-12-02T01:08:30,383 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741857_1033 (size=5672) 2024-12-02T01:08:30,384 INFO [M:0;a82f277fe18c:43051 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/7fdfdd7f5a0649cc8611e3f045d67af4 2024-12-02T01:08:30,405 DEBUG [M:0;a82f277fe18c:43051 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6466a668e9204bce888bb4d318db0d19 is 766, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733101602060/Put/seqid=0 2024-12-02T01:08:30,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741858_1034 (size=6426) 2024-12-02T01:08:30,410 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741858_1034 (size=6426) 2024-12-02T01:08:30,411 INFO [M:0;a82f277fe18c:43051 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=39.61 KB at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6466a668e9204bce888bb4d318db0d19 2024-12-02T01:08:30,417 INFO [M:0;a82f277fe18c:43051 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 6466a668e9204bce888bb4d318db0d19 2024-12-02T01:08:30,432 DEBUG [M:0;a82f277fe18c:43051 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/fc68f782272f47e0884e90e75d624f93 is 69, key is a82f277fe18c,35243,1733101598726/rs:state/1733101599695/Put/seqid=0 2024-12-02T01:08:30,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741859_1035 (size=5156) 2024-12-02T01:08:30,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741859_1035 (size=5156) 2024-12-02T01:08:30,441 INFO [M:0;a82f277fe18c:43051 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/fc68f782272f47e0884e90e75d624f93 2024-12-02T01:08:30,442 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:08:30,442 INFO [RS:0;a82f277fe18c:35243 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,35243,1733101598726; zookeeper connection closed. 2024-12-02T01:08:30,443 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35243-0x10194dd549e0001, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:08:30,443 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@546ca626 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@546ca626 2024-12-02T01:08:30,443 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T01:08:30,460 DEBUG [M:0;a82f277fe18c:43051 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/4dfaf1bf0a24484691141718015c157f is 52, key is load_balancer_on/state:d/1733101601574/Put/seqid=0 2024-12-02T01:08:30,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741860_1036 (size=5056) 2024-12-02T01:08:30,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741860_1036 (size=5056) 2024-12-02T01:08:30,466 INFO [M:0;a82f277fe18c:43051 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/4dfaf1bf0a24484691141718015c157f 2024-12-02T01:08:30,472 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/7fdfdd7f5a0649cc8611e3f045d67af4 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/7fdfdd7f5a0649cc8611e3f045d67af4 2024-12-02T01:08:30,479 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/7fdfdd7f5a0649cc8611e3f045d67af4, entries=8, sequenceid=104, filesize=5.5 K 2024-12-02T01:08:30,480 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6466a668e9204bce888bb4d318db0d19 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/6466a668e9204bce888bb4d318db0d19 2024-12-02T01:08:30,485 INFO [M:0;a82f277fe18c:43051 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 6466a668e9204bce888bb4d318db0d19 2024-12-02T01:08:30,485 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/6466a668e9204bce888bb4d318db0d19, entries=11, sequenceid=104, filesize=6.3 K 2024-12-02T01:08:30,487 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/fc68f782272f47e0884e90e75d624f93 as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/fc68f782272f47e0884e90e75d624f93 2024-12-02T01:08:30,492 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/fc68f782272f47e0884e90e75d624f93, entries=1, sequenceid=104, filesize=5.0 K 2024-12-02T01:08:30,493 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/4dfaf1bf0a24484691141718015c157f as hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/4dfaf1bf0a24484691141718015c157f 2024-12-02T01:08:30,499 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/4dfaf1bf0a24484691141718015c157f, entries=1, sequenceid=104, filesize=4.9 K 2024-12-02T01:08:30,500 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(3040): Finished flush of dataSize ~40.21 KB/41173, heapSize ~50.08 KB/51280, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 140ms, sequenceid=104, compaction requested=false 2024-12-02T01:08:30,501 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:30,501 DEBUG [M:0;a82f277fe18c:43051 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:08:30,502 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/MasterData/WALs/a82f277fe18c,43051,1733101597955 2024-12-02T01:08:30,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:44461 is added to blk_1073741830_1006 (size=48474) 2024-12-02T01:08:30,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38321 is added to blk_1073741830_1006 (size=48474) 2024-12-02T01:08:30,504 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:08:30,504 INFO [M:0;a82f277fe18c:43051 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-02T01:08:30,504 INFO [M:0;a82f277fe18c:43051 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:43051 2024-12-02T01:08:30,508 DEBUG [M:0;a82f277fe18c:43051 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/a82f277fe18c,43051,1733101597955 already deleted, retry=false 2024-12-02T01:08:30,617 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:08:30,618 INFO [M:0;a82f277fe18c:43051 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,43051,1733101597955; zookeeper connection closed. 2024-12-02T01:08:30,618 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:43051-0x10194dd549e0000, quorum=127.0.0.1:52426, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:08:30,660 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@63d8281c{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:30,665 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@55b489f0{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:08:30,665 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:08:30,666 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@49fab02b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:08:30,666 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2170f3b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir/,STOPPED} 2024-12-02T01:08:30,670 WARN [BP-1841578127-172.17.0.2-1733101594540 heartbeating to localhost/127.0.0.1:41885 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:08:30,670 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:08:30,670 WARN [BP-1841578127-172.17.0.2-1733101594540 heartbeating to localhost/127.0.0.1:41885 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1841578127-172.17.0.2-1733101594540 (Datanode Uuid 46dad73d-aeaa-4d89-aa02-b0f83b51536f) service to localhost/127.0.0.1:41885 2024-12-02T01:08:30,670 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:08:30,671 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data3/current/BP-1841578127-172.17.0.2-1733101594540 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:30,671 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data4/current/BP-1841578127-172.17.0.2-1733101594540 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:30,672 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:08:30,674 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@9f8c8e3{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:30,674 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@49793231{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:08:30,674 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:08:30,674 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4c2208{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:08:30,674 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4239ce1e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir/,STOPPED} 2024-12-02T01:08:30,676 WARN [BP-1841578127-172.17.0.2-1733101594540 heartbeating to localhost/127.0.0.1:41885 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:08:30,676 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:08:30,676 WARN [BP-1841578127-172.17.0.2-1733101594540 heartbeating to localhost/127.0.0.1:41885 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1841578127-172.17.0.2-1733101594540 (Datanode Uuid 348ded74-6a6b-4069-9a99-01aa7523fcc1) service to localhost/127.0.0.1:41885 2024-12-02T01:08:30,676 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:08:30,676 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data1/current/BP-1841578127-172.17.0.2-1733101594540 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:30,676 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/cluster_5b49ecbc-4499-3e9e-33d6-4678538faa68/dfs/data/data2/current/BP-1841578127-172.17.0.2-1733101594540 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:30,677 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:08:30,703 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4cd532bc{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:08:30,703 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@619b4309{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:08:30,703 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:08:30,704 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7c7fe1c5{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:08:30,704 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@37896107{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir/,STOPPED} 2024-12-02T01:08:30,711 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-02T01:08:30,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-02T01:08:30,749 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testSlowSyncLogRolling Thread=65 (was 12) Potentially hanging thread: nioEventLoopGroup-3-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: nioEventLoopGroup-5-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Monitor thread for TaskMonitor java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: master/a82f277fe18c:0:becomeActiveMaster-MemStoreChunkPool Statistics java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-4-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SSL Certificates Store Monitor java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:41885 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: LeaseRenewer:jenkins@localhost:41885 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async-Client-Retry-Timer-pool-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Time-limited test.named-queue-events-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41885 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-1-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Idle-Rpc-Conn-Sweeper-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: RS-EventLoopGroup-1-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:41885 from jenkins.hfs.0 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41885 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:41885 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-4-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-5-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HBase-Metrics2-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-1-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RpcClient-timer-pool-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Potentially hanging thread: nioEventLoopGroup-5-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: regionserver/a82f277fe18c:0.procedureResultReporter java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Potentially hanging thread: ForkJoinPool-2-worker-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: org.apache.hadoop.hdfs.PeerCache@32ddba60 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-4-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SessionTracker java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Potentially hanging thread: nioEventLoopGroup-2-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: ForkJoinPool-2-worker-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: Timer for 'HBase' metrics system java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:41885 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.0@localhost:41885 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-1 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-2 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: master/a82f277fe18c:0:becomeActiveMaster-MemStoreChunkPool Statistics java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SnapshotHandlerChoreCleaner java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) - Thread LEAK? -, OpenFileDescriptor=403 (was 286) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=53 (was 206), ProcessCount=11 (was 11), AvailableMemoryMB=5826 (was 6409) 2024-12-02T01:08:30,754 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnDatanodeDeath Thread=66, OpenFileDescriptor=403, MaxFileDescriptor=1048576, SystemLoadAverage=53, ProcessCount=11, AvailableMemoryMB=5826 2024-12-02T01:08:30,754 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.log.dir so I do NOT create it in target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/7d5ae82c-447e-a0ff-8015-554ad20bcd44/hadoop.tmp.dir so I do NOT create it in target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6, deleteOnExit=true 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/test.cache.data in system properties and HBase conf 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir in system properties and HBase conf 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T01:08:30,755 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-02T01:08:30,755 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/nfs.dump.dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:08:30,756 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T01:08:30,757 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T01:08:30,767 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:08:31,030 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:31,035 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:08:31,036 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:08:31,036 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:08:31,036 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:08:31,036 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:31,037 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4888cf22{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:08:31,037 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@469fe845{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:08:31,129 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4d8cb8b4{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-40661-hadoop-hdfs-3_4_1-tests_jar-_-any-4893349979971833646/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:08:31,129 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1b1e7792{HTTP/1.1, (http/1.1)}{localhost:40661} 2024-12-02T01:08:31,129 INFO [Time-limited test {}] server.Server(415): Started @118258ms 2024-12-02T01:08:31,140 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:08:31,353 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:31,356 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:08:31,357 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:08:31,357 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:08:31,357 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:08:31,358 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3de9ccbc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:08:31,358 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@65050874{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:08:31,448 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5746e761{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-43811-hadoop-hdfs-3_4_1-tests_jar-_-any-1790282141666979497/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:31,449 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2ba2f018{HTTP/1.1, (http/1.1)}{localhost:43811} 2024-12-02T01:08:31,449 INFO [Time-limited test {}] server.Server(415): Started @118577ms 2024-12-02T01:08:31,450 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:08:31,477 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:31,481 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:08:31,482 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:08:31,483 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:08:31,483 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:08:31,483 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@35577aad{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:08:31,483 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21e823e3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:08:31,575 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1d128d9e{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-45507-hadoop-hdfs-3_4_1-tests_jar-_-any-16298709560937126418/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:31,575 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@201a6835{HTTP/1.1, (http/1.1)}{localhost:45507} 2024-12-02T01:08:31,575 INFO [Time-limited test {}] server.Server(415): Started @118704ms 2024-12-02T01:08:31,577 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:08:32,428 WARN [Thread-465 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data1/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:32,428 WARN [Thread-466 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data2/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:32,446 WARN [Thread-429 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:08:32,448 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc853217056f934ef with lease ID 0x9558ea20ec860509: Processing first storage report for DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7 from datanode DatanodeRegistration(127.0.0.1:43415, datanodeUuid=73426485-8606-4598-83ad-d124d678981d, infoPort=46351, infoSecurePort=0, ipcPort=42195, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:32,448 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc853217056f934ef with lease ID 0x9558ea20ec860509: from storage DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7 node DatanodeRegistration(127.0.0.1:43415, datanodeUuid=73426485-8606-4598-83ad-d124d678981d, infoPort=46351, infoSecurePort=0, ipcPort=42195, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:32,448 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc853217056f934ef with lease ID 0x9558ea20ec860509: Processing first storage report for DS-b18dd70f-6caf-4b6c-92f7-eed780b3078e from datanode DatanodeRegistration(127.0.0.1:43415, datanodeUuid=73426485-8606-4598-83ad-d124d678981d, infoPort=46351, infoSecurePort=0, ipcPort=42195, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:32,448 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc853217056f934ef with lease ID 0x9558ea20ec860509: from storage DS-b18dd70f-6caf-4b6c-92f7-eed780b3078e node DatanodeRegistration(127.0.0.1:43415, datanodeUuid=73426485-8606-4598-83ad-d124d678981d, infoPort=46351, infoSecurePort=0, ipcPort=42195, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:32,564 WARN [Thread-476 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data3/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:32,565 WARN [Thread-477 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data4/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:32,580 WARN [Thread-452 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:08:32,582 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdfb2abc5601e5c8e with lease ID 0x9558ea20ec86050a: Processing first storage report for DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470 from datanode DatanodeRegistration(127.0.0.1:46875, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=35591, infoSecurePort=0, ipcPort=45839, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:32,583 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdfb2abc5601e5c8e with lease ID 0x9558ea20ec86050a: from storage DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470 node DatanodeRegistration(127.0.0.1:46875, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=35591, infoSecurePort=0, ipcPort=45839, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:32,583 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xdfb2abc5601e5c8e with lease ID 0x9558ea20ec86050a: Processing first storage report for DS-8594e2e6-8250-4e68-a8d1-67d17010e217 from datanode DatanodeRegistration(127.0.0.1:46875, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=35591, infoSecurePort=0, ipcPort=45839, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:32,583 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xdfb2abc5601e5c8e with lease ID 0x9558ea20ec86050a: from storage DS-8594e2e6-8250-4e68-a8d1-67d17010e217 node DatanodeRegistration(127.0.0.1:46875, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=35591, infoSecurePort=0, ipcPort=45839, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:32,616 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3 2024-12-02T01:08:32,623 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/zookeeper_0, clientPort=54150, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T01:08:32,625 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=54150 2024-12-02T01:08:32,625 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,627 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:08:32,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:08:32,637 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4 with version=8 2024-12-02T01:08:32,637 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase-staging 2024-12-02T01:08:32,639 INFO [Time-limited test {}] client.ConnectionUtils(129): master/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:08:32,639 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:32,640 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:32,640 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:08:32,640 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:32,640 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:08:32,640 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:08:32,640 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:08:32,641 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:39475 2024-12-02T01:08:32,641 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,643 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,645 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:39475 connecting to ZooKeeper ensemble=127.0.0.1:54150 2024-12-02T01:08:32,707 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:394750x0, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:08:32,708 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:39475-0x10194df17600000 connected 2024-12-02T01:08:32,769 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:08:32,771 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:08:32,773 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:08:32,774 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39475 2024-12-02T01:08:32,774 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39475 2024-12-02T01:08:32,775 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39475 2024-12-02T01:08:32,776 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39475 2024-12-02T01:08:32,777 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39475 2024-12-02T01:08:32,777 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4, hbase.cluster.distributed=false 2024-12-02T01:08:32,797 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:08:32,798 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:08:32,799 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:43621 2024-12-02T01:08:32,799 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:08:32,800 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:08:32,800 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,802 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,805 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:43621 connecting to ZooKeeper ensemble=127.0.0.1:54150 2024-12-02T01:08:32,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:436210x0, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:08:32,809 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:436210x0, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:08:32,809 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43621-0x10194df17600001 connected 2024-12-02T01:08:32,810 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:08:32,811 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:08:32,811 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43621 2024-12-02T01:08:32,814 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43621 2024-12-02T01:08:32,815 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43621 2024-12-02T01:08:32,815 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43621 2024-12-02T01:08:32,815 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43621 2024-12-02T01:08:32,816 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/a82f277fe18c,39475,1733101712639 2024-12-02T01:08:32,825 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:08:32,825 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:08:32,826 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a82f277fe18c,39475,1733101712639 2024-12-02T01:08:32,827 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a82f277fe18c:39475 2024-12-02T01:08:32,834 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:08:32,834 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:08:32,834 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:32,834 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:32,834 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:08:32,835 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a82f277fe18c,39475,1733101712639 from backup master directory 2024-12-02T01:08:32,842 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:08:32,842 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a82f277fe18c,39475,1733101712639 2024-12-02T01:08:32,842 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:08:32,842 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:08:32,842 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:08:32,842 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a82f277fe18c,39475,1733101712639 2024-12-02T01:08:32,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:08:32,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:08:32,856 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/hbase.id with ID: eb12433f-4e2d-49e6-94e0-185fb135153c 2024-12-02T01:08:32,868 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:32,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:32,875 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:32,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:08:32,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:08:32,883 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:08:32,884 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T01:08:32,884 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:08:32,892 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:08:32,892 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:08:32,893 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store 2024-12-02T01:08:32,899 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:08:32,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:08:32,900 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:32,901 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:08:32,901 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:32,901 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:32,901 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:08:32,901 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:32,901 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:08:32,901 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:08:32,902 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/.initializing 2024-12-02T01:08:32,902 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639 2024-12-02T01:08:32,905 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C39475%2C1733101712639, suffix=, logDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639, archiveDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/oldWALs, maxLogs=10 2024-12-02T01:08:32,906 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C39475%2C1733101712639.1733101712905 2024-12-02T01:08:32,911 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 2024-12-02T01:08:32,911 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46351:46351),(127.0.0.1/127.0.0.1:35591:35591)] 2024-12-02T01:08:32,911 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:08:32,912 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:32,912 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,912 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,913 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,915 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T01:08:32,915 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:32,916 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:32,916 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,918 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T01:08:32,918 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:32,919 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:08:32,919 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,920 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T01:08:32,921 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:32,921 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:08:32,921 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,924 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T01:08:32,924 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:32,925 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:08:32,926 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,926 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,928 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T01:08:32,929 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:08:32,931 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:08:32,932 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=700681, jitterRate=-0.10903814435005188}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T01:08:32,932 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:08:32,933 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T01:08:32,937 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5605e5ed, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:08:32,938 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-02T01:08:32,938 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T01:08:32,938 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T01:08:32,938 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T01:08:32,939 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T01:08:32,939 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-02T01:08:32,939 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T01:08:32,941 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T01:08:32,942 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T01:08:32,950 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-02T01:08:32,950 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T01:08:32,951 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T01:08:32,959 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-02T01:08:32,959 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T01:08:32,960 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T01:08:32,967 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-02T01:08:32,968 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T01:08:32,975 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T01:08:32,977 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T01:08:32,984 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T01:08:32,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:08:32,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:08:32,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:32,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:32,993 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=a82f277fe18c,39475,1733101712639, sessionid=0x10194df17600000, setting cluster-up flag (Was=false) 2024-12-02T01:08:33,009 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:33,009 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:33,034 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T01:08:33,038 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,39475,1733101712639 2024-12-02T01:08:33,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:33,060 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:33,084 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T01:08:33,085 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,39475,1733101712639 2024-12-02T01:08:33,088 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-02T01:08:33,088 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-02T01:08:33,088 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T01:08:33,088 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a82f277fe18c,39475,1733101712639 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a82f277fe18c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:08:33,089 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733101743090 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T01:08:33,090 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T01:08:33,091 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,091 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T01:08:33,091 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:08:33,091 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T01:08:33,091 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-02T01:08:33,091 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T01:08:33,091 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T01:08:33,091 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T01:08:33,092 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101713092,5,FailOnTimeoutGroup] 2024-12-02T01:08:33,092 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101713092,5,FailOnTimeoutGroup] 2024-12-02T01:08:33,092 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,092 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T01:08:33,092 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,092 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,092 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,092 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:08:33,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:08:33,100 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:08:33,101 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-02T01:08:33,101 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4 2024-12-02T01:08:33,110 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:08:33,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:08:33,111 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:33,113 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:08:33,114 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:08:33,114 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,115 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:33,115 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:08:33,117 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:08:33,117 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,117 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:33,118 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:08:33,119 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:08:33,119 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,120 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:33,121 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/meta/1588230740 2024-12-02T01:08:33,121 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/meta/1588230740 2024-12-02T01:08:33,123 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:08:33,124 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:08:33,126 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:08:33,127 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=881377, jitterRate=0.12072926759719849}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:08:33,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:08:33,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:08:33,128 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:08:33,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:08:33,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:08:33,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:08:33,128 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:08:33,128 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:08:33,129 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:08:33,129 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-02T01:08:33,129 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T01:08:33,130 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T01:08:33,131 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T01:08:33,133 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a82f277fe18c:43621 2024-12-02T01:08:33,134 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1008): ClusterId : eb12433f-4e2d-49e6-94e0-185fb135153c 2024-12-02T01:08:33,134 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:08:33,143 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:08:33,143 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:08:33,151 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:08:33,151 DEBUG [RS:0;a82f277fe18c:43621 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@a2dc24d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:08:33,152 DEBUG [RS:0;a82f277fe18c:43621 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6bcf6386, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:08:33,152 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:08:33,152 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:08:33,152 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:08:33,153 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,39475,1733101712639 with isa=a82f277fe18c/172.17.0.2:43621, startcode=1733101712797 2024-12-02T01:08:33,153 DEBUG [RS:0;a82f277fe18c:43621 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:08:33,155 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44075, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:08:33,156 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39475 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,156 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39475 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,158 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4 2024-12-02T01:08:33,158 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:45673 2024-12-02T01:08:33,158 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:08:33,167 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:08:33,168 DEBUG [RS:0;a82f277fe18c:43621 {}] zookeeper.ZKUtil(111): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,168 WARN [RS:0;a82f277fe18c:43621 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:08:33,168 INFO [RS:0;a82f277fe18c:43621 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:08:33,168 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,168 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,43621,1733101712797] 2024-12-02T01:08:33,171 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:08:33,172 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:08:33,174 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:08:33,174 INFO [RS:0;a82f277fe18c:43621 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:08:33,174 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,176 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:08:33,177 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,177 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,177 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,177 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,177 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:08:33,178 DEBUG [RS:0;a82f277fe18c:43621 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:08:33,178 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,179 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,179 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,179 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,179 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43621,1733101712797-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:08:33,191 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:08:33,191 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,43621,1733101712797-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,204 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.Replication(204): a82f277fe18c,43621,1733101712797 started 2024-12-02T01:08:33,204 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,43621,1733101712797, RpcServer on a82f277fe18c/172.17.0.2:43621, sessionid=0x10194df17600001 2024-12-02T01:08:33,204 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:08:33,204 DEBUG [RS:0;a82f277fe18c:43621 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,204 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,43621,1733101712797' 2024-12-02T01:08:33,204 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:08:33,205 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:08:33,205 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:08:33,206 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:08:33,206 DEBUG [RS:0;a82f277fe18c:43621 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,206 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,43621,1733101712797' 2024-12-02T01:08:33,206 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:08:33,206 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:08:33,206 DEBUG [RS:0;a82f277fe18c:43621 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:08:33,206 INFO [RS:0;a82f277fe18c:43621 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:08:33,206 INFO [RS:0;a82f277fe18c:43621 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:08:33,282 WARN [a82f277fe18c:39475 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-02T01:08:33,313 INFO [RS:0;a82f277fe18c:43621 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C43621%2C1733101712797, suffix=, logDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797, archiveDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs, maxLogs=32 2024-12-02T01:08:33,318 INFO [RS:0;a82f277fe18c:43621 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.1733101713317 2024-12-02T01:08:33,325 INFO [RS:0;a82f277fe18c:43621 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 2024-12-02T01:08:33,325 DEBUG [RS:0;a82f277fe18c:43621 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46351:46351),(127.0.0.1/127.0.0.1:35591:35591)] 2024-12-02T01:08:33,532 DEBUG [a82f277fe18c:39475 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T01:08:33,533 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,537 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,43621,1733101712797, state=OPENING 2024-12-02T01:08:33,592 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T01:08:33,601 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:33,601 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:33,602 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=a82f277fe18c,43621,1733101712797}] 2024-12-02T01:08:33,602 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:08:33,602 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:08:33,759 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,759 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T01:08:33,765 INFO [RS-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36376, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T01:08:33,772 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-02T01:08:33,772 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:08:33,777 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C43621%2C1733101712797.meta, suffix=.meta, logDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797, archiveDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs, maxLogs=32 2024-12-02T01:08:33,780 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta 2024-12-02T01:08:33,786 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta 2024-12-02T01:08:33,786 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35591:35591),(127.0.0.1/127.0.0.1:46351:46351)] 2024-12-02T01:08:33,786 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:08:33,786 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T01:08:33,787 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T01:08:33,787 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T01:08:33,787 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T01:08:33,787 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:33,787 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-02T01:08:33,787 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-02T01:08:33,789 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:08:33,790 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:08:33,791 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,791 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:33,791 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:08:33,792 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:08:33,792 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,793 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:33,793 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:08:33,794 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:08:33,794 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,794 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:08:33,795 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/meta/1588230740 2024-12-02T01:08:33,796 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/meta/1588230740 2024-12-02T01:08:33,798 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:08:33,799 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:08:33,799 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=697346, jitterRate=-0.1132795512676239}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:08:33,800 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:08:33,801 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733101713759 2024-12-02T01:08:33,803 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T01:08:33,803 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-02T01:08:33,804 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:33,805 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,43621,1733101712797, state=OPEN 2024-12-02T01:08:33,832 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:08:33,832 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:08:33,832 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:08:33,832 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:08:33,834 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T01:08:33,834 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=a82f277fe18c,43621,1733101712797 in 230 msec 2024-12-02T01:08:33,836 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T01:08:33,836 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 705 msec 2024-12-02T01:08:33,838 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 750 msec 2024-12-02T01:08:33,838 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733101713838, completionTime=-1 2024-12-02T01:08:33,839 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T01:08:33,839 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-02T01:08:33,839 DEBUG [hconnection-0x1ffa02cf-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:08:33,840 INFO [RS-EventLoopGroup-6-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36384, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:08:33,842 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-02T01:08:33,842 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733101773842 2024-12-02T01:08:33,842 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733101833842 2024-12-02T01:08:33,842 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 3 msec 2024-12-02T01:08:33,867 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39475,1733101712639-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,867 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39475,1733101712639-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,867 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39475,1733101712639-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,868 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a82f277fe18c:39475, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,868 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:33,868 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-02T01:08:33,868 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:08:33,869 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-02T01:08:33,870 DEBUG [master/a82f277fe18c:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-02T01:08:33,871 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:08:33,871 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:33,872 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:08:33,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:08:33,880 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:08:33,882 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 1e4c9691efacb5eafb00181ffc92f4e6, NAME => 'hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4 2024-12-02T01:08:33,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:08:33,891 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:08:33,891 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:33,891 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 1e4c9691efacb5eafb00181ffc92f4e6, disabling compactions & flushes 2024-12-02T01:08:33,891 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:33,891 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:33,891 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. after waiting 0 ms 2024-12-02T01:08:33,891 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:33,892 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:33,892 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 1e4c9691efacb5eafb00181ffc92f4e6: 2024-12-02T01:08:33,893 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:08:33,893 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733101713893"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101713893"}]},"ts":"1733101713893"} 2024-12-02T01:08:33,895 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:08:33,897 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:08:33,897 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101713897"}]},"ts":"1733101713897"} 2024-12-02T01:08:33,899 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-02T01:08:33,917 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=1e4c9691efacb5eafb00181ffc92f4e6, ASSIGN}] 2024-12-02T01:08:33,919 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=1e4c9691efacb5eafb00181ffc92f4e6, ASSIGN 2024-12-02T01:08:33,920 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=1e4c9691efacb5eafb00181ffc92f4e6, ASSIGN; state=OFFLINE, location=a82f277fe18c,43621,1733101712797; forceNewPlan=false, retain=false 2024-12-02T01:08:34,071 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=1e4c9691efacb5eafb00181ffc92f4e6, regionState=OPENING, regionLocation=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:34,076 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 1e4c9691efacb5eafb00181ffc92f4e6, server=a82f277fe18c,43621,1733101712797}] 2024-12-02T01:08:34,232 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,43621,1733101712797 2024-12-02T01:08:34,238 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:34,238 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 1e4c9691efacb5eafb00181ffc92f4e6, NAME => 'hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:08:34,239 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,239 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:34,239 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,239 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,242 INFO [StoreOpener-1e4c9691efacb5eafb00181ffc92f4e6-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,244 INFO [StoreOpener-1e4c9691efacb5eafb00181ffc92f4e6-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1e4c9691efacb5eafb00181ffc92f4e6 columnFamilyName info 2024-12-02T01:08:34,244 DEBUG [StoreOpener-1e4c9691efacb5eafb00181ffc92f4e6-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:34,245 INFO [StoreOpener-1e4c9691efacb5eafb00181ffc92f4e6-1 {}] regionserver.HStore(327): Store=1e4c9691efacb5eafb00181ffc92f4e6/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:08:34,247 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,248 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,251 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:08:34,254 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:08:34,255 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 1e4c9691efacb5eafb00181ffc92f4e6; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=804166, jitterRate=0.02255004644393921}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:08:34,256 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 1e4c9691efacb5eafb00181ffc92f4e6: 2024-12-02T01:08:34,257 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6., pid=6, masterSystemTime=1733101714232 2024-12-02T01:08:34,259 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:34,259 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:08:34,260 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=1e4c9691efacb5eafb00181ffc92f4e6, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:34,265 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T01:08:34,266 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 1e4c9691efacb5eafb00181ffc92f4e6, server=a82f277fe18c,43621,1733101712797 in 186 msec 2024-12-02T01:08:34,267 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T01:08:34,267 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=1e4c9691efacb5eafb00181ffc92f4e6, ASSIGN in 348 msec 2024-12-02T01:08:34,268 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:08:34,268 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101714268"}]},"ts":"1733101714268"} 2024-12-02T01:08:34,270 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-02T01:08:34,310 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-02T01:08:34,311 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:08:34,314 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 443 msec 2024-12-02T01:08:34,317 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:08:34,317 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:34,317 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:08:34,326 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-02T01:08:34,335 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:08:34,339 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:08:34,342 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:08:34,353 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 27 msec 2024-12-02T01:08:34,358 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-02T01:08:34,375 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:08:34,386 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 28 msec 2024-12-02T01:08:34,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-02T01:08:34,425 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-02T01:08:34,425 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.583sec 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39475,1733101712639-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:08:34,426 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39475,1733101712639-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T01:08:34,428 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-02T01:08:34,428 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T01:08:34,428 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39475,1733101712639-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,519 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x3d4e4d20 to 127.0.0.1:54150 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@102c8f 2024-12-02T01:08:34,526 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@83bb3e2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:08:34,528 DEBUG [hconnection-0x6eba50b1-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:08:34,529 INFO [RS-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36398, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:08:34,531 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=a82f277fe18c,39475,1733101712639 2024-12-02T01:08:34,532 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:34,534 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-02T01:08:34,547 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:08:34,548 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:08:34,549 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:40483 2024-12-02T01:08:34,549 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:08:34,550 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:08:34,550 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:34,552 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:08:34,554 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:40483 connecting to ZooKeeper ensemble=127.0.0.1:54150 2024-12-02T01:08:34,559 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:404830x0, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:08:34,559 DEBUG [Time-limited test {}] zookeeper.ZKUtil(111): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:08:34,559 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:40483-0x10194df17600003 connected 2024-12-02T01:08:34,560 DEBUG [Time-limited test {}] zookeeper.ZKUtil(111): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/running 2024-12-02T01:08:34,561 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:08:34,561 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40483 2024-12-02T01:08:34,561 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40483 2024-12-02T01:08:34,562 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40483 2024-12-02T01:08:34,562 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40483 2024-12-02T01:08:34,562 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40483 2024-12-02T01:08:34,563 DEBUG [pool-282-thread-1 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: INIT 2024-12-02T01:08:34,580 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;a82f277fe18c:40483 2024-12-02T01:08:34,581 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1008): ClusterId : eb12433f-4e2d-49e6-94e0-185fb135153c 2024-12-02T01:08:34,581 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:08:34,592 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:08:34,593 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:08:34,601 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:08:34,602 DEBUG [RS:1;a82f277fe18c:40483 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@582e693c, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:08:34,602 DEBUG [RS:1;a82f277fe18c:40483 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2327ebd2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:08:34,603 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:08:34,603 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:08:34,603 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:08:34,603 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,39475,1733101712639 with isa=a82f277fe18c/172.17.0.2:40483, startcode=1733101714547 2024-12-02T01:08:34,604 DEBUG [RS:1;a82f277fe18c:40483 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:08:34,606 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33097, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:08:34,606 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39475 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,40483,1733101714547 2024-12-02T01:08:34,606 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39475 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,40483,1733101714547 2024-12-02T01:08:34,608 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4 2024-12-02T01:08:34,608 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:45673 2024-12-02T01:08:34,608 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:08:34,617 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:08:34,618 DEBUG [RS:1;a82f277fe18c:40483 {}] zookeeper.ZKUtil(111): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,40483,1733101714547 2024-12-02T01:08:34,618 WARN [RS:1;a82f277fe18c:40483 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:08:34,618 INFO [RS:1;a82f277fe18c:40483 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:08:34,618 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,40483,1733101714547] 2024-12-02T01:08:34,618 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,40483,1733101714547 2024-12-02T01:08:34,624 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:08:34,624 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:08:34,627 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:08:34,627 INFO [RS:1;a82f277fe18c:40483 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:08:34,627 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,627 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:08:34,629 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,629 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,630 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,630 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:08:34,630 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:08:34,630 DEBUG [RS:1;a82f277fe18c:40483 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:08:34,630 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,630 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,630 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,631 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,631 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40483,1733101714547-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:08:34,671 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:08:34,671 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40483,1733101714547-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:08:34,683 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.Replication(204): a82f277fe18c,40483,1733101714547 started 2024-12-02T01:08:34,683 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,40483,1733101714547, RpcServer on a82f277fe18c/172.17.0.2:40483, sessionid=0x10194df17600003 2024-12-02T01:08:34,683 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:08:34,683 INFO [Time-limited test {}] hbase.HBaseTestingUtility(3355): Started new server=Thread[RS:1;a82f277fe18c:40483,5,FailOnTimeoutGroup] 2024-12-02T01:08:34,683 DEBUG [RS:1;a82f277fe18c:40483 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,40483,1733101714547 2024-12-02T01:08:34,683 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,40483,1733101714547' 2024-12-02T01:08:34,683 INFO [Time-limited test {}] wal.TestLogRolling(191): Replication=2 2024-12-02T01:08:34,683 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:08:34,683 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:08:34,684 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,40483,1733101714547 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,40483,1733101714547' 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:08:34,684 DEBUG [RS:1;a82f277fe18c:40483 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:08:34,684 INFO [RS:1;a82f277fe18c:40483 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:08:34,684 INFO [RS:1;a82f277fe18c:40483 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:08:34,686 INFO [RS-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:52204, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T01:08:34,687 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T01:08:34,687 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T01:08:34,687 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testLogRollOnDatanodeDeath', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:08:34,689 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T01:08:34,690 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:08:34,690 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:34,690 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRollOnDatanodeDeath" procId is: 9 2024-12-02T01:08:34,691 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:08:34,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:08:34,699 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741837_1013 (size=393) 2024-12-02T01:08:34,699 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741837_1013 (size=393) 2024-12-02T01:08:34,701 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 61b9a0fcb731850a98b7b912948ae7ef, NAME => 'TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRollOnDatanodeDeath', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4 2024-12-02T01:08:34,708 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43415 is added to blk_1073741838_1014 (size=76) 2024-12-02T01:08:34,708 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46875 is added to blk_1073741838_1014 (size=76) 2024-12-02T01:08:34,709 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:34,709 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1681): Closing 61b9a0fcb731850a98b7b912948ae7ef, disabling compactions & flushes 2024-12-02T01:08:34,709 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:34,709 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:34,709 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. after waiting 0 ms 2024-12-02T01:08:34,709 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:34,709 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:34,709 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1635): Region close journal for 61b9a0fcb731850a98b7b912948ae7ef: 2024-12-02T01:08:34,711 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:08:34,711 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef.","families":{"info":[{"qualifier":"regioninfo","vlen":75,"tag":[],"timestamp":"1733101714711"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101714711"}]},"ts":"1733101714711"} 2024-12-02T01:08:34,713 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:08:34,714 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:08:34,714 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnDatanodeDeath","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101714714"}]},"ts":"1733101714714"} 2024-12-02T01:08:34,716 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnDatanodeDeath, state=ENABLING in hbase:meta 2024-12-02T01:08:34,734 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=61b9a0fcb731850a98b7b912948ae7ef, ASSIGN}] 2024-12-02T01:08:34,736 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=61b9a0fcb731850a98b7b912948ae7ef, ASSIGN 2024-12-02T01:08:34,736 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=61b9a0fcb731850a98b7b912948ae7ef, ASSIGN; state=OFFLINE, location=a82f277fe18c,43621,1733101712797; forceNewPlan=false, retain=false 2024-12-02T01:08:34,787 INFO [RS:1;a82f277fe18c:40483 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C40483%2C1733101714547, suffix=, logDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,40483,1733101714547, archiveDir=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs, maxLogs=32 2024-12-02T01:08:34,787 INFO [RS:1;a82f277fe18c:40483 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C40483%2C1733101714547.1733101714787 2024-12-02T01:08:34,793 INFO [RS:1;a82f277fe18c:40483 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,40483,1733101714547/a82f277fe18c%2C40483%2C1733101714547.1733101714787 2024-12-02T01:08:34,794 DEBUG [RS:1;a82f277fe18c:40483 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35591:35591),(127.0.0.1/127.0.0.1:46351:46351)] 2024-12-02T01:08:34,852 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:08:34,859 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:08:34,877 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:08:34,888 INFO [a82f277fe18c:39475 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-02T01:08:34,888 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=61b9a0fcb731850a98b7b912948ae7ef, regionState=OPENING, regionLocation=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:34,890 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 61b9a0fcb731850a98b7b912948ae7ef, server=a82f277fe18c,43621,1733101712797}] 2024-12-02T01:08:35,044 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to a82f277fe18c,43621,1733101712797 2024-12-02T01:08:35,053 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:35,053 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 61b9a0fcb731850a98b7b912948ae7ef, NAME => 'TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:08:35,054 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRollOnDatanodeDeath 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,054 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:08:35,055 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,055 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,056 INFO [StoreOpener-61b9a0fcb731850a98b7b912948ae7ef-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,058 INFO [StoreOpener-61b9a0fcb731850a98b7b912948ae7ef-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 61b9a0fcb731850a98b7b912948ae7ef columnFamilyName info 2024-12-02T01:08:35,058 DEBUG [StoreOpener-61b9a0fcb731850a98b7b912948ae7ef-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:08:35,058 INFO [StoreOpener-61b9a0fcb731850a98b7b912948ae7ef-1 {}] regionserver.HStore(327): Store=61b9a0fcb731850a98b7b912948ae7ef/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:08:35,059 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,059 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,061 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:08:35,063 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:08:35,064 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened 61b9a0fcb731850a98b7b912948ae7ef; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=874527, jitterRate=0.11201916635036469}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:08:35,064 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 61b9a0fcb731850a98b7b912948ae7ef: 2024-12-02T01:08:35,065 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef., pid=11, masterSystemTime=1733101715043 2024-12-02T01:08:35,067 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:35,067 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:35,068 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=61b9a0fcb731850a98b7b912948ae7ef, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,43621,1733101712797 2024-12-02T01:08:35,074 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-02T01:08:35,075 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 61b9a0fcb731850a98b7b912948ae7ef, server=a82f277fe18c,43621,1733101712797 in 181 msec 2024-12-02T01:08:35,076 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-02T01:08:35,076 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=61b9a0fcb731850a98b7b912948ae7ef, ASSIGN in 340 msec 2024-12-02T01:08:35,077 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:08:35,077 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnDatanodeDeath","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101715077"}]},"ts":"1733101715077"} 2024-12-02T01:08:35,079 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnDatanodeDeath, state=ENABLED in hbase:meta 2024-12-02T01:08:35,119 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:08:35,122 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath in 432 msec 2024-12-02T01:08:38,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T01:08:38,353 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath Metrics about Tables on a single HBase RegionServer 2024-12-02T01:08:38,355 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling 2024-12-02T01:08:39,172 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T01:08:39,173 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-02T01:08:39,174 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRollOnDatanodeDeath' 2024-12-02T01:08:39,744 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:08:39,751 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:08:39,768 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:08:44,694 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39475 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:08:44,696 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testLogRollOnDatanodeDeath, procId: 9 completed 2024-12-02T01:08:44,701 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testLogRollOnDatanodeDeath 2024-12-02T01:08:44,701 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:08:44,716 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:44,718 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:08:44,720 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:08:44,720 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:08:44,720 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:08:44,721 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2b1e9118{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:08:44,721 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@37cda5fa{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:08:44,811 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5447727{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-36193-hadoop-hdfs-3_4_1-tests_jar-_-any-5956498152520612611/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:44,811 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@d776958{HTTP/1.1, (http/1.1)}{localhost:36193} 2024-12-02T01:08:44,811 INFO [Time-limited test {}] server.Server(415): Started @131940ms 2024-12-02T01:08:44,812 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:08:44,836 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:44,840 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:08:44,841 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:08:44,841 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:08:44,841 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:08:44,841 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5c3fb4a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:08:44,841 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@32501436{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:08:44,933 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4eefe283{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-38719-hadoop-hdfs-3_4_1-tests_jar-_-any-12017922951922125597/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:44,934 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1db6f92{HTTP/1.1, (http/1.1)}{localhost:38719} 2024-12-02T01:08:44,934 INFO [Time-limited test {}] server.Server(415): Started @132062ms 2024-12-02T01:08:44,935 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:08:44,960 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:08:44,963 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:08:44,964 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:08:44,964 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:08:44,964 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:08:44,964 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@291e2cdd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:08:44,965 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@71afbf1a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:08:45,054 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3bc72c07{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-46877-hadoop-hdfs-3_4_1-tests_jar-_-any-6929183063650934423/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:45,054 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@64f48490{HTTP/1.1, (http/1.1)}{localhost:46877} 2024-12-02T01:08:45,054 INFO [Time-limited test {}] server.Server(415): Started @132182ms 2024-12-02T01:08:45,055 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:08:45,966 WARN [Thread-671 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:45,966 WARN [Thread-672 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:45,986 WARN [Thread-613 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:08:45,988 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x667276764ee89743 with lease ID 0x9558ea20ec86050b: Processing first storage report for DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4 from datanode DatanodeRegistration(127.0.0.1:46881, datanodeUuid=738f7a67-e050-43f5-8a33-713780210261, infoPort=35799, infoSecurePort=0, ipcPort=43955, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:45,989 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x667276764ee89743 with lease ID 0x9558ea20ec86050b: from storage DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4 node DatanodeRegistration(127.0.0.1:46881, datanodeUuid=738f7a67-e050-43f5-8a33-713780210261, infoPort=35799, infoSecurePort=0, ipcPort=43955, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:45,989 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x667276764ee89743 with lease ID 0x9558ea20ec86050b: Processing first storage report for DS-6b5ed385-afb6-4ce9-bb6b-741d4589be75 from datanode DatanodeRegistration(127.0.0.1:46881, datanodeUuid=738f7a67-e050-43f5-8a33-713780210261, infoPort=35799, infoSecurePort=0, ipcPort=43955, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:45,989 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x667276764ee89743 with lease ID 0x9558ea20ec86050b: from storage DS-6b5ed385-afb6-4ce9-bb6b-741d4589be75 node DatanodeRegistration(127.0.0.1:46881, datanodeUuid=738f7a67-e050-43f5-8a33-713780210261, infoPort=35799, infoSecurePort=0, ipcPort=43955, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:46,163 WARN [Thread-684 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data8/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:46,163 WARN [Thread-683 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data7/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:46,181 WARN [Thread-635 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:08:46,183 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5b80eb0c0691714e with lease ID 0x9558ea20ec86050c: Processing first storage report for DS-912c962e-d06f-4e15-b497-4bed30d45b30 from datanode DatanodeRegistration(127.0.0.1:35113, datanodeUuid=d372c827-d061-4af2-a36b-359ce3b195ae, infoPort=45841, infoSecurePort=0, ipcPort=45135, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:46,183 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5b80eb0c0691714e with lease ID 0x9558ea20ec86050c: from storage DS-912c962e-d06f-4e15-b497-4bed30d45b30 node DatanodeRegistration(127.0.0.1:35113, datanodeUuid=d372c827-d061-4af2-a36b-359ce3b195ae, infoPort=45841, infoSecurePort=0, ipcPort=45135, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:46,183 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x5b80eb0c0691714e with lease ID 0x9558ea20ec86050c: Processing first storage report for DS-f8ec7a5f-f45c-4496-acc8-b266481bdc24 from datanode DatanodeRegistration(127.0.0.1:35113, datanodeUuid=d372c827-d061-4af2-a36b-359ce3b195ae, infoPort=45841, infoSecurePort=0, ipcPort=45135, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:46,183 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x5b80eb0c0691714e with lease ID 0x9558ea20ec86050c: from storage DS-f8ec7a5f-f45c-4496-acc8-b266481bdc24 node DatanodeRegistration(127.0.0.1:35113, datanodeUuid=d372c827-d061-4af2-a36b-359ce3b195ae, infoPort=45841, infoSecurePort=0, ipcPort=45135, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:46,288 WARN [Thread-694 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data9/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:46,289 WARN [Thread-695 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data10/current/BP-1905452902-172.17.0.2-1733101710778/current, will proceed with Du for space computation calculation, 2024-12-02T01:08:46,305 WARN [Thread-657 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:08:46,307 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xba1d399a051fb96f with lease ID 0x9558ea20ec86050d: Processing first storage report for DS-29547388-b272-43d0-a834-3d972ecfa228 from datanode DatanodeRegistration(127.0.0.1:36073, datanodeUuid=66565eb9-6cb2-49a9-abc0-6f89b5337d8c, infoPort=37987, infoSecurePort=0, ipcPort=39525, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:46,307 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xba1d399a051fb96f with lease ID 0x9558ea20ec86050d: from storage DS-29547388-b272-43d0-a834-3d972ecfa228 node DatanodeRegistration(127.0.0.1:36073, datanodeUuid=66565eb9-6cb2-49a9-abc0-6f89b5337d8c, infoPort=37987, infoSecurePort=0, ipcPort=39525, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:46,308 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xba1d399a051fb96f with lease ID 0x9558ea20ec86050d: Processing first storage report for DS-db091874-4857-4497-ba08-1c90caa221ab from datanode DatanodeRegistration(127.0.0.1:36073, datanodeUuid=66565eb9-6cb2-49a9-abc0-6f89b5337d8c, infoPort=37987, infoSecurePort=0, ipcPort=39525, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778) 2024-12-02T01:08:46,308 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xba1d399a051fb96f with lease ID 0x9558ea20ec86050d: from storage DS-db091874-4857-4497-ba08-1c90caa221ab node DatanodeRegistration(127.0.0.1:36073, datanodeUuid=66565eb9-6cb2-49a9-abc0-6f89b5337d8c, infoPort=37987, infoSecurePort=0, ipcPort=39525, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:08:46,392 WARN [ResponseProcessor for block BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006 java.io.IOException: Bad response ERROR for BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006 from datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,392 WARN [ResponseProcessor for block BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009 java.io.IOException: Bad response ERROR for BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009 from datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,391 WARN [ResponseProcessor for block BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,392 WARN [ResponseProcessor for block BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,399 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:08:46,400 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta block BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK], DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:08:46,400 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:08:46,400 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,40483,1733101714547/a82f277fe18c%2C40483%2C1733101714547.1733101714787 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK], DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:08:46,400 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1d128d9e{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:46,400 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@201a6835{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:08:46,401 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:08:46,401 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21e823e3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:08:46,401 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@35577aad{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:08:46,401 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1984233940_22 at /127.0.0.1:47914 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015] {}] datanode.DataXceiver(331): 127.0.0.1:43415:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47914 dst: /127.0.0.1:43415 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,401 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:47866 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:43415:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47866 dst: /127.0.0.1:43415 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,400 WARN [PacketResponder: BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:46875] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,400 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1984233940_22 at /127.0.0.1:36960 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015] {}] datanode.DataXceiver(331): 127.0.0.1:46875:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36960 dst: /127.0.0.1:46875 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,400 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:36906 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:46875:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36906 dst: /127.0.0.1:46875 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,400 WARN [PacketResponder: BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:46875] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Broken pipe at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,402 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:47852 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:43415:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47852 dst: /127.0.0.1:43415 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,402 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:47838 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:43415:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47838 dst: /127.0.0.1:43415 java.net.SocketException: Connection reset at sun.nio.ch.SocketChannelImpl.throwConnectionReset(SocketChannelImpl.java:394) ~[?:?] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:426) ~[?:?] at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,402 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:36900 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:46875:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36900 dst: /127.0.0.1:46875 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,402 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:36862 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:46875:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36862 dst: /127.0.0.1:46875 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:46,404 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:08:46,404 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:08:46,404 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1905452902-172.17.0.2-1733101710778 (Datanode Uuid 5379d92f-966e-4337-93c6-da8a06d97702) service to localhost/127.0.0.1:45673 2024-12-02T01:08:46,404 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:08:46,404 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data3/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:46,405 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data4/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:46,405 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:08:46,406 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta block BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741834_1010 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,406 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741830_1006 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,406 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741833_1009 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,406 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,40483,1733101714547/a82f277fe18c%2C40483%2C1733101714547.1733101714787 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741839_1015 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741839_1015 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,407 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5746e761{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:46,408 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2ba2f018{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:08:46,408 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:08:46,408 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@65050874{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:08:46,408 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3de9ccbc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:08:46,409 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:08:46,409 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:08:46,409 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1905452902-172.17.0.2-1733101710778 (Datanode Uuid 73426485-8606-4598-83ad-d124d678981d) service to localhost/127.0.0.1:45673 2024-12-02T01:08:46,409 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:08:46,409 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data1/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:46,410 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data2/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:46,410 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:08:46,413 WARN [RS:0;a82f277fe18c:43621.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=4, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,414 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C43621%2C1733101712797:(num 1733101713317) roll requested 2024-12-02T01:08:46,414 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.1733101726414 2024-12-02T01:08:46,414 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43621 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=4, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,417 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43621 {}] ipc.CallRunner(138): callId: 9 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:36398 deadline: 1733101736413, exception=org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=4, requesting roll of WAL 2024-12-02T01:08:46,420 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=4, requesting roll of WAL 2024-12-02T01:08:46,420 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 with entries=4, filesize=959 B; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101726414 2024-12-02T01:08:46,421 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37987:37987),(127.0.0.1/127.0.0.1:45841:45841)] 2024-12-02T01:08:46,421 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:08:46,421 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,421 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:46,422 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(47): Initialize RecoverLeaseFSUtils 2024-12-02T01:08:46,422 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(59): set recoverLeaseMethod to org.apache.hadoop.fs.LeaseRecoverable.recoverLease() 2024-12-02T01:08:46,422 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 2024-12-02T01:08:46,425 WARN [IPC Server handler 4 on default port 45673 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 has not been closed. Lease recovery is in progress. RecoveryId = 1021 for block blk_1073741833_1009 2024-12-02T01:08:46,426 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 after 4ms 2024-12-02T01:08:50,428 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 after 4006ms 2024-12-02T01:08:58,513 INFO [Time-limited test {}] wal.TestLogRolling(243): log.getCurrentFileName(): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101726414 2024-12-02T01:08:58,514 WARN [ResponseProcessor for block BP-1905452902-172.17.0.2-1733101710778:blk_1073741840_1020 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1905452902-172.17.0.2-1733101710778:blk_1073741840_1020 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:58,515 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101726414 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741840_1020 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741840_1020 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK], DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]) is bad. 2024-12-02T01:08:58,516 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:52894 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741840_1020] {}] datanode.DataXceiver(331): 127.0.0.1:36073:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52894 dst: /127.0.0.1:36073 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,517 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50542 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741840_1020] {}] datanode.DataXceiver(331): 127.0.0.1:35113:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50542 dst: /127.0.0.1:35113 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,570 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3bc72c07{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:08:58,571 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@64f48490{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:08:58,571 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:08:58,572 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@71afbf1a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:08:58,572 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@291e2cdd{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:08:58,574 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:08:58,574 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:08:58,574 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1905452902-172.17.0.2-1733101710778 (Datanode Uuid 66565eb9-6cb2-49a9-abc0-6f89b5337d8c) service to localhost/127.0.0.1:45673 2024-12-02T01:08:58,574 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:08:58,575 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data9/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:58,575 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data10/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:08:58,575 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:08:58,578 WARN [sync.1 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]] 2024-12-02T01:08:58,578 WARN [sync.1 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]] 2024-12-02T01:08:58,579 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C43621%2C1733101712797:(num 1733101726414) roll requested 2024-12-02T01:08:58,579 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.1733101738579 2024-12-02T01:08:58,584 WARN [Thread-714 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741841_1023 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:36073 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:58,584 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50124 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741841_1023] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741841_1023 to mirror 127.0.0.1:36073 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,585 WARN [Thread-714 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741841_1023 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]) is bad. 2024-12-02T01:08:58,585 WARN [Thread-714 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741841_1023 2024-12-02T01:08:58,585 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50124 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741841_1023] {}] datanode.BlockReceiver(316): Block 1073741841 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:08:58,585 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50124 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741841_1023] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50124 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,588 WARN [Thread-714 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK] 2024-12-02T01:08:58,591 WARN [Thread-714 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741842_1024 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:43415 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:58,591 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50604 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741842_1024] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data7, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data8]'}, localName='127.0.0.1:35113', datanodeUuid='d372c827-d061-4af2-a36b-359ce3b195ae', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741842_1024 to mirror 127.0.0.1:43415 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,591 WARN [Thread-714 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741842_1024 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]) is bad. 2024-12-02T01:08:58,591 WARN [Thread-714 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741842_1024 2024-12-02T01:08:58,591 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50604 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741842_1024] {}] datanode.BlockReceiver(316): Block 1073741842 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:08:58,591 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50604 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741842_1024] {}] datanode.DataXceiver(331): 127.0.0.1:35113:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50604 dst: /127.0.0.1:35113 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,592 WARN [Thread-714 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK] 2024-12-02T01:08:58,594 WARN [Thread-714 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741843_1025 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:46875 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:08:58,594 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50132 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741843_1025] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741843_1025 to mirror 127.0.0.1:46875 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,594 WARN [Thread-714 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741843_1025 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:08:58,594 WARN [Thread-714 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741843_1025 2024-12-02T01:08:58,594 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50132 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741843_1025] {}] datanode.BlockReceiver(316): Block 1073741843 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:08:58,594 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50132 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741843_1025] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50132 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:08:58,595 WARN [Thread-714 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] 2024-12-02T01:08:58,599 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101726414 with entries=2, filesize=2.37 KB; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 2024-12-02T01:08:58,600 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35799:35799),(127.0.0.1/127.0.0.1:45841:45841)] 2024-12-02T01:08:58,600 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:08:58,600 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101726414 is not closed yet, will try archiving it next time 2024-12-02T01:08:58,601 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35113 is added to blk_1073741840_1022 (size=2431) 2024-12-02T01:08:59,004 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741840_1022 (size=2431) 2024-12-02T01:09:02,584 WARN [ResponseProcessor for block BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026 java.io.IOException: Bad response ERROR for BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026 from datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,585 WARN [DataStreamer for file /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 block BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:02,585 WARN [PacketResponder: BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:35113] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,587 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50136 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50136 dst: /127.0.0.1:46881 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,587 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50620 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1026] {}] datanode.DataXceiver(331): 127.0.0.1:35113:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50620 dst: /127.0.0.1:35113 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,615 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:09:02,616 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4eefe283{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:02,617 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1db6f92{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:09:02,617 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:09:02,617 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@32501436{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:09:02,618 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5c3fb4a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:09:02,620 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:09:02,620 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:09:02,620 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1905452902-172.17.0.2-1733101710778 (Datanode Uuid d372c827-d061-4af2-a36b-359ce3b195ae) service to localhost/127.0.0.1:45673 2024-12-02T01:09:02,620 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:09:02,621 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data7/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:02,621 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data8/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:02,622 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:09:02,624 WARN [sync.4 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]] 2024-12-02T01:09:02,624 WARN [sync.4 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]] 2024-12-02T01:09:02,624 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C43621%2C1733101712797:(num 1733101738579) roll requested 2024-12-02T01:09:02,625 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.1733101742625 2024-12-02T01:09:02,628 WARN [Thread-731 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741845_1028 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,629 WARN [Thread-731 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741845_1028 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:02,629 WARN [Thread-731 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741845_1028 2024-12-02T01:09:02,630 WARN [Thread-731 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:02,631 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43621 {}] regionserver.HRegion(8581): Flush requested on 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:09:02,631 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 61b9a0fcb731850a98b7b912948ae7ef 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:09:02,631 WARN [Thread-731 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741846_1029 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,631 WARN [Thread-731 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741846_1029 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK], DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]) is bad. 2024-12-02T01:09:02,632 WARN [Thread-731 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741846_1029 2024-12-02T01:09:02,632 WARN [Thread-731 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK] 2024-12-02T01:09:02,635 WARN [Thread-731 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741847_1030 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:46875 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,635 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50160 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741847_1030] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741847_1030 to mirror 127.0.0.1:46875 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,635 WARN [Thread-731 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741847_1030 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:09:02,635 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50160 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741847_1030] {}] datanode.BlockReceiver(316): Block 1073741847 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:09:02,635 WARN [Thread-731 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741847_1030 2024-12-02T01:09:02,635 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50160 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741847_1030] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50160 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,637 WARN [Thread-731 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] 2024-12-02T01:09:02,639 WARN [Thread-731 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741848_1031 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,639 WARN [Thread-731 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741848_1031 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]) is bad. 2024-12-02T01:09:02,639 WARN [Thread-731 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741848_1031 2024-12-02T01:09:02,640 WARN [Thread-731 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK] 2024-12-02T01:09:02,641 WARN [IPC Server handler 1 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T01:09:02,641 WARN [IPC Server handler 1 on default port 45673 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T01:09:02,641 WARN [IPC Server handler 1 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T01:09:02,645 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 with entries=12, filesize=12.96 KB; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742625 2024-12-02T01:09:02,645 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35799:35799)] 2024-12-02T01:09:02,645 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,645 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,645 WARN [sync.1 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]] 2024-12-02T01:09:02,646 WARN [sync.1 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]] 2024-12-02T01:09:02,646 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C43621%2C1733101712797:(num 1733101742625) roll requested 2024-12-02T01:09:02,646 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.1733101742646 2024-12-02T01:09:02,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741844_1027 (size=13275) 2024-12-02T01:09:02,648 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741850_1033 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,649 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741850_1033 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:02,649 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741850_1033 2024-12-02T01:09:02,649 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/.tmp/info/5771d431432048f08fbb91cf5d72b42e is 1080, key is row0002/info:/1733101738577/Put/seqid=0 2024-12-02T01:09:02,649 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:02,651 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741851_1034 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,651 WARN [Thread-734 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741852_1035 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,651 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741851_1034 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]) is bad. 2024-12-02T01:09:02,651 WARN [Thread-734 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741852_1035 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK], DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]) is bad. 2024-12-02T01:09:02,651 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741851_1034 2024-12-02T01:09:02,651 WARN [Thread-734 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741852_1035 2024-12-02T01:09:02,651 WARN [Thread-734 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK] 2024-12-02T01:09:02,651 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK] 2024-12-02T01:09:02,652 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741854_1037 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,653 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741854_1037 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:09:02,653 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741854_1037 2024-12-02T01:09:02,653 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] 2024-12-02T01:09:02,653 WARN [Thread-734 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741853_1036 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:36073 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,653 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50170 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741853_1036] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741853_1036 to mirror 127.0.0.1:36073 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,653 WARN [Thread-734 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741853_1036 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]) is bad. 2024-12-02T01:09:02,653 WARN [Thread-734 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741853_1036 2024-12-02T01:09:02,653 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50170 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741853_1036] {}] datanode.BlockReceiver(316): Block 1073741853 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T01:09:02,654 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50170 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741853_1036] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50170 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,654 WARN [Thread-734 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK] 2024-12-02T01:09:02,655 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741855_1038 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:43415 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,655 WARN [Thread-734 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741856_1039 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,655 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50176 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741855_1038] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741855_1038 to mirror 127.0.0.1:43415 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,655 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741855_1038 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]) is bad. 2024-12-02T01:09:02,655 WARN [Thread-734 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741856_1039 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:02,655 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741855_1038 2024-12-02T01:09:02,655 WARN [Thread-734 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741856_1039 2024-12-02T01:09:02,655 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50176 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741855_1038] {}] datanode.BlockReceiver(316): Block 1073741855 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:09:02,655 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50176 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741855_1038] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50176 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,656 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK] 2024-12-02T01:09:02,656 WARN [Thread-734 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:02,656 WARN [IPC Server handler 2 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T01:09:02,657 WARN [IPC Server handler 2 on default port 45673 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T01:09:02,657 WARN [IPC Server handler 2 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T01:09:02,658 WARN [Thread-734 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741857_1040 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:46875 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:02,658 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50186 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741857_1040] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741857_1040 to mirror 127.0.0.1:46875 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,658 WARN [Thread-734 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741857_1040 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:09:02,658 WARN [Thread-734 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741857_1040 2024-12-02T01:09:02,658 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50186 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741857_1040] {}] datanode.BlockReceiver(316): Block 1073741857 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T01:09:02,658 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:50186 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741857_1040] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50186 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:02,659 WARN [Thread-734 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] 2024-12-02T01:09:02,660 WARN [IPC Server handler 3 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T01:09:02,660 WARN [IPC Server handler 3 on default port 45673 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T01:09:02,660 WARN [IPC Server handler 3 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T01:09:02,662 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742625 with entries=1, filesize=1.22 KB; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742646 2024-12-02T01:09:02,662 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35799:35799)] 2024-12-02T01:09:02,662 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,662 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,662 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742625 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741859_1042 (size=10347) 2024-12-02T01:09:02,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741849_1032 (size=1261) 2024-12-02T01:09:02,663 DEBUG [Close-WAL-Writer-2 {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,663 DEBUG [Close-WAL-Writer-2 {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 is not closed yet, will try archiving it next time 2024-12-02T01:09:02,849 WARN [sync.3 {}] wal.FSHLog(760): Too many consecutive RollWriter requests, it's a sign of the total number of live datanodes is lower than the tolerable replicas. 2024-12-02T01:09:03,050 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 is not closed yet, will try archiving it next time 2024-12-02T01:09:03,064 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:03,064 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=12 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/.tmp/info/5771d431432048f08fbb91cf5d72b42e 2024-12-02T01:09:03,067 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:09:03,067 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:09:03,068 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:09:03,068 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:09:03,069 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40516419{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:09:03,069 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@69d8bad2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:09:03,072 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/.tmp/info/5771d431432048f08fbb91cf5d72b42e as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/info/5771d431432048f08fbb91cf5d72b42e 2024-12-02T01:09:03,080 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/info/5771d431432048f08fbb91cf5d72b42e, entries=5, sequenceid=12, filesize=10.1 K 2024-12-02T01:09:03,081 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=8.40 KB/8606 for 61b9a0fcb731850a98b7b912948ae7ef in 449ms, sequenceid=12, compaction requested=false 2024-12-02T01:09:03,081 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 61b9a0fcb731850a98b7b912948ae7ef: 2024-12-02T01:09:03,091 WARN [master/a82f277fe18c:0:becomeActiveMaster.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=96, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,091 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C39475%2C1733101712639:(num 1733101712905) roll requested 2024-12-02T01:09:03,092 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C39475%2C1733101712639.1733101743091 2024-12-02T01:09:03,092 ERROR [ProcExecTimeout {}] region.RegionProcedureStore(422): Failed to delete pids=[4, 7, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,092 ERROR [ProcExecTimeout {}] procedure2.TimeoutExecutorThread(124): Ignoring pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner exception: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL java.io.UncheckedIOException: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL at org.apache.hadoop.hbase.procedure2.store.region.RegionProcedureStore.delete(RegionProcedureStore.java:423) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner.periodicExecute(CompletedProcedureCleaner.java:135) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.executeInMemoryChore(TimeoutExecutorThread.java:122) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.execDelayedProcedure(TimeoutExecutorThread.java:101) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.run(TimeoutExecutorThread.java:68) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] Caused by: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,095 WARN [Thread-757 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741860_1043 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,095 WARN [Thread-757 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741860_1043 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]) is bad. 2024-12-02T01:09:03,095 WARN [Thread-757 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741860_1043 2024-12-02T01:09:03,095 WARN [Thread-757 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK] 2024-12-02T01:09:03,096 WARN [Thread-757 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741861_1044 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,097 WARN [Thread-757 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741861_1044 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:03,097 WARN [Thread-757 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741861_1044 2024-12-02T01:09:03,097 WARN [Thread-757 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:03,099 WARN [Thread-757 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741862_1045 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:46875 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,099 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:50212 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741862_1045] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741862_1045 to mirror 127.0.0.1:46875 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:03,099 WARN [Thread-757 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741862_1045 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]) is bad. 2024-12-02T01:09:03,099 WARN [Thread-757 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741862_1045 2024-12-02T01:09:03,099 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:50212 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741862_1045] {}] datanode.BlockReceiver(316): Block 1073741862 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:09:03,100 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:50212 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741862_1045] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50212 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:03,100 WARN [Thread-757 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:46875,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK] 2024-12-02T01:09:03,101 WARN [Thread-757 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741863_1046 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,102 WARN [Thread-757 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741863_1046 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK]) is bad. 2024-12-02T01:09:03,102 WARN [Thread-757 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741863_1046 2024-12-02T01:09:03,102 WARN [Thread-757 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:36073,DS-29547388-b272-43d0-a834-3d972ecfa228,DISK] 2024-12-02T01:09:03,103 WARN [IPC Server handler 2 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-02T01:09:03,103 WARN [IPC Server handler 2 on default port 45673 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-02T01:09:03,103 WARN [IPC Server handler 2 on default port 45673 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-02T01:09:03,106 WARN [master:store-WAL-Roller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL 2024-12-02T01:09:03,106 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 with entries=93, filesize=46.04 KB; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101743091 2024-12-02T01:09:03,106 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35799:35799)] 2024-12-02T01:09:03,106 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 is not closed yet, will try archiving it next time 2024-12-02T01:09:03,106 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,106 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:03,106 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 2024-12-02T01:09:03,107 WARN [IPC Server handler 4 on default port 45673 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 has not been closed. Lease recovery is in progress. RecoveryId = 1048 for block blk_1073741830_1006 2024-12-02T01:09:03,107 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 after 0ms 2024-12-02T01:09:03,162 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5dbaaf37{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/java.io.tmpdir/jetty-localhost-45323-hadoop-hdfs-3_4_1-tests_jar-_-any-6934400140615037428/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:03,162 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@d79cb{HTTP/1.1, (http/1.1)}{localhost:45323} 2024-12-02T01:09:03,162 INFO [Time-limited test {}] server.Server(415): Started @150290ms 2024-12-02T01:09:03,163 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:09:03,542 WARN [Thread-768 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:09:03,548 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6298214f48c036d7 with lease ID 0x9558ea20ec86050e: from storage DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470 node DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 8, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:03,548 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6298214f48c036d7 with lease ID 0x9558ea20ec86050e: from storage DS-8594e2e6-8250-4e68-a8d1-67d17010e217 node DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:04,994 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@74a09e82[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:46881, datanodeUuid=738f7a67-e050-43f5-8a33-713780210261, infoPort=35799, infoSecurePort=0, ipcPort=43955, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741859_1042 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:04,994 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@677edaca[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:46881, datanodeUuid=738f7a67-e050-43f5-8a33-713780210261, infoPort=35799, infoSecurePort=0, ipcPort=43955, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741844_1027 to 127.0.0.1:36073 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:05,152 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:09:05,155 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49172, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:09:05,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741849_1032 (size=1261) 2024-12-02T01:09:06,603 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:09:06,607 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49186, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:09:07,110 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101712905 after 4003ms 2024-12-02T01:09:13,546 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@31d274f3[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741837_1013 to 127.0.0.1:36073 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:13,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:09:13,562 WARN [org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1@67052e87 {}] datanode.BlockRecoveryWorker$RecoveryTaskContiguous(164): Failed to recover block (block=BP-1905452902-172.17.0.2-1733101710778:blk_1073741833_1009, datanode=DatanodeInfoWithStorage[127.0.0.1:43415,null,null]) java.net.ConnectException: Call From a82f277fe18c/172.17.0.2 to localhost:42195 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:961) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:876) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1588) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy55.initReplicaRecovery(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.lambda$initReplicaRecovery$0(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.initReplicaRecovery(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.callInitReplicaRecovery(BlockRecoveryWorker.java:561) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.access$400(BlockRecoveryWorker.java:57) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$RecoveryTaskContiguous.recover(BlockRecoveryWorker.java:135) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1.run(BlockRecoveryWorker.java:602) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:668) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:789) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:364) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getConnection(Client.java:1649) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1473) ~[hadoop-common-3.4.1.jar:?] ... 12 more 2024-12-02T01:09:13,571 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741833_1021 (size=959) 2024-12-02T01:09:14,468 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101726414 to hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs/a82f277fe18c%2C43621%2C1733101712797.1733101726414 2024-12-02T01:09:14,548 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@56a8abb1[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741831_1007 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:14,548 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@31d274f3[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741829_1005 to 127.0.0.1:36073 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:14,633 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T01:09:14,633 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T01:09:16,548 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@31d274f3[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741832_1008 to 127.0.0.1:36073 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:16,553 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:09:17,547 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@56a8abb1[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741826_1002 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:17,548 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741833_1021 (size=959) 2024-12-02T01:09:19,549 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@56a8abb1[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741825_1001 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:19,549 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@31d274f3[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741827_1003 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:20,054 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 61b9a0fcb731850a98b7b912948ae7ef, had cached 0 bytes from a total of 10347 2024-12-02T01:09:20,548 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@31d274f3[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741838_1014 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:20,548 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@56a8abb1[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741836_1012 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:22,239 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.1733101762239 2024-12-02T01:09:22,252 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742646 with entries=2, filesize=1.57 KB; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101762239 2024-12-02T01:09:22,252 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43975:43975),(127.0.0.1/127.0.0.1:35799:35799)] 2024-12-02T01:09:22,252 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742646 is not closed yet, will try archiving it next time 2024-12-02T01:09:22,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741858_1041 (size=1618) 2024-12-02T01:09:22,255 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43621 {}] regionserver.HRegion(8581): Flush requested on 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:09:22,255 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 61b9a0fcb731850a98b7b912948ae7ef 1/1 column families, dataSize=9.46 KB heapSize=10.38 KB 2024-12-02T01:09:22,256 INFO [sync.2 {}] wal.FSHLog(777): LowReplication-Roller was enabled. 2024-12-02T01:09:22,260 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/.tmp/info/203fb8e0222a425e83cac9ac4a57ce1f is 1080, key is row0007/info:/1733101742632/Put/seqid=0 2024-12-02T01:09:22,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741866_1050 (size=13583) 2024-12-02T01:09:22,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741866_1050 (size=13583) 2024-12-02T01:09:22,267 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=9.46 KB at sequenceid=24 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/.tmp/info/203fb8e0222a425e83cac9ac4a57ce1f 2024-12-02T01:09:22,271 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-02T01:09:22,271 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-02T01:09:22,271 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x3d4e4d20 to 127.0.0.1:54150 2024-12-02T01:09:22,272 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:22,272 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T01:09:22,272 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=359808442, stopped=false 2024-12-02T01:09:22,272 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=a82f277fe18c,39475,1733101712639 2024-12-02T01:09:22,275 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/.tmp/info/203fb8e0222a425e83cac9ac4a57ce1f as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/info/203fb8e0222a425e83cac9ac4a57ce1f 2024-12-02T01:09:22,280 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/info/203fb8e0222a425e83cac9ac4a57ce1f, entries=8, sequenceid=24, filesize=13.3 K 2024-12-02T01:09:22,281 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~9.46 KB/9682, heapSize ~10.36 KB/10608, currentSize=9.46 KB/9684 for 61b9a0fcb731850a98b7b912948ae7ef in 26ms, sequenceid=24, compaction requested=false 2024-12-02T01:09:22,282 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 61b9a0fcb731850a98b7b912948ae7ef: 2024-12-02T01:09:22,282 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=23.4 K, sizeToCheck=16.0 K 2024-12-02T01:09:22,282 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:09:22,282 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/default/TestLogRolling-testLogRollOnDatanodeDeath/61b9a0fcb731850a98b7b912948ae7ef/info/203fb8e0222a425e83cac9ac4a57ce1f because midkey is the same as first or last row 2024-12-02T01:09:22,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:09:22,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:09:22,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:09:22,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:22,290 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-02T01:09:22,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:22,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:22,290 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:22,290 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,43621,1733101712797' ***** 2024-12-02T01:09:22,290 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:09:22,291 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,40483,1733101714547' ***** 2024-12-02T01:09:22,291 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:09:22,291 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:09:22,291 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:09:22,291 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:09:22,291 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:09:22,291 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:09:22,291 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:09:22,291 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:09:22,291 INFO [RS:0;a82f277fe18c:43621 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:09:22,291 INFO [RS:0;a82f277fe18c:43621 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:09:22,291 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(3579): Received CLOSE for 1e4c9691efacb5eafb00181ffc92f4e6 2024-12-02T01:09:22,291 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(3579): Received CLOSE for 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:09:22,291 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,43621,1733101712797 2024-12-02T01:09:22,292 DEBUG [RS:0;a82f277fe18c:43621 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 1e4c9691efacb5eafb00181ffc92f4e6, disabling compactions & flushes 2024-12-02T01:09:22,292 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:09:22,292 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:09:22,292 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:09:22,292 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:09:22,292 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:09:22,292 INFO [RS:1;a82f277fe18c:40483 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. after waiting 0 ms 2024-12-02T01:09:22,292 INFO [RS:1;a82f277fe18c:40483 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:09:22,292 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,40483,1733101714547 2024-12-02T01:09:22,292 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 1e4c9691efacb5eafb00181ffc92f4e6 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-02T01:09:22,292 DEBUG [RS:1;a82f277fe18c:40483 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:22,292 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-02T01:09:22,292 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,40483,1733101714547; all regions closed. 2024-12-02T01:09:22,292 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1603): Online Regions={1588230740=hbase:meta,,1.1588230740, 1e4c9691efacb5eafb00181ffc92f4e6=hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6., 61b9a0fcb731850a98b7b912948ae7ef=TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef.} 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:09:22,292 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:09:22,292 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 1e4c9691efacb5eafb00181ffc92f4e6, 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:09:22,292 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:09:22,293 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:09:22,293 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,40483,1733101714547 2024-12-02T01:09:22,293 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.87 KB heapSize=5.40 KB 2024-12-02T01:09:22,293 WARN [WAL-Shutdown-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,293 ERROR [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1664): Shutdown / close of WAL failed: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... 2024-12-02T01:09:22,293 WARN [RS_OPEN_META-regionserver/a82f277fe18c:0-0.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=15, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,293 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C43621%2C1733101712797.meta:.meta(num 1733101713779) roll requested 2024-12-02T01:09:22,293 DEBUG [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1665): Shutdown / close exception details: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,293 DEBUG [RS:1;a82f277fe18c:40483 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:22,293 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:09:22,294 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:09:22,294 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C43621%2C1733101712797.meta.1733101762293.meta 2024-12-02T01:09:22,294 INFO [RS:1;a82f277fe18c:40483 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-02T01:09:22,294 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:09:22,294 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:09:22,294 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:09:22,294 ERROR [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2808): ***** ABORTING region server a82f277fe18c,43621,1733101712797: Unrecoverable exception while closing hbase:meta,,1.1588230740 ***** org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,294 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:09:22,294 ERROR [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2815): RegionServer abort: loaded coprocessors are: [org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint] 2024-12-02T01:09:22,295 INFO [RS:1;a82f277fe18c:40483 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:40483 2024-12-02T01:09:22,298 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for java.lang:type=Memory 2024-12-02T01:09:22,298 WARN [Thread-811 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741867_1051 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:35113 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,298 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:38856 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741867_1051] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data3, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data4]'}, localName='127.0.0.1:42011', datanodeUuid='5379d92f-966e-4337-93c6-da8a06d97702', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741867_1051 to mirror 127.0.0.1:35113 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:22,298 WARN [Thread-811 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741867_1051 in pipeline [DatanodeInfoWithStorage[127.0.0.1:42011,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK], DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:22,299 WARN [Thread-811 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741867_1051 2024-12-02T01:09:22,299 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:38856 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741867_1051] {}] datanode.BlockReceiver(316): Block 1073741867 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-02T01:09:22,299 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1412922333_22 at /127.0.0.1:38856 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741867_1051] {}] datanode.DataXceiver(331): 127.0.0.1:42011:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:38856 dst: /127.0.0.1:42011 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:22,299 WARN [Thread-811 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:22,300 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=IPC 2024-12-02T01:09:22,300 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=Replication 2024-12-02T01:09:22,300 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=Server 2024-12-02T01:09:22,300 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2819): Dump of metrics as JSON on abort: { "beans": [ { "name": "java.lang:type=Memory", "modelerType": "sun.management.MemoryImpl", "Verbose": false, "ObjectPendingFinalizationCount": 0, "HeapMemoryUsage": { "committed": 1048576000, "init": 1048576000, "max": 2306867200, "used": 437745664 }, "NonHeapMemoryUsage": { "committed": 161873920, "init": 7667712, "max": -1, "used": 160075928 }, "ObjectName": "java.lang:type=Memory" } ], "beans": [], "beans": [], "beans": [] } 2024-12-02T01:09:22,300 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,40483,1733101714547 2024-12-02T01:09:22,300 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:09:22,303 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL 2024-12-02T01:09:22,303 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta with entries=11, filesize=3.63 KB; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101762293.meta 2024-12-02T01:09:22,304 WARN [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39475 {}] master.MasterRpcServices(626): a82f277fe18c,43621,1733101712797 reported a fatal error: ***** ABORTING region server a82f277fe18c,43621,1733101712797: Unrecoverable exception while closing hbase:meta,,1.1588230740 ***** Cause: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) at java.base/java.lang.Thread.run(Thread.java:840) Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) 2024-12-02T01:09:22,305 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43975:43975),(127.0.0.1/127.0.0.1:35799:35799)] 2024-12-02T01:09:22,305 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta is not closed yet, will try archiving it next time 2024-12-02T01:09:22,305 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,305 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43415,DS-4de4faff-fe41-4d98-86c7-d5a91e6eb2f7,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:22,305 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta 2024-12-02T01:09:22,306 WARN [IPC Server handler 1 on default port 45673 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta has not been closed. Lease recovery is in progress. RecoveryId = 1053 for block blk_1073741834_1010 2024-12-02T01:09:22,306 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta after 1ms 2024-12-02T01:09:22,309 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,40483,1733101714547] 2024-12-02T01:09:22,309 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,40483,1733101714547; numProcessing=1 2024-12-02T01:09:22,314 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/.tmp/info/89d22c31120b40d6a13bbc8cb9230eec is 45, key is default/info:d/1733101714333/Put/seqid=0 2024-12-02T01:09:22,317 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,40483,1733101714547 already deleted, retry=false 2024-12-02T01:09:22,317 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,40483,1733101714547 expired; onlineServers=1 2024-12-02T01:09:22,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741869_1054 (size=5037) 2024-12-02T01:09:22,320 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741869_1054 (size=5037) 2024-12-02T01:09:22,320 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/.tmp/info/89d22c31120b40d6a13bbc8cb9230eec 2024-12-02T01:09:22,327 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/.tmp/info/89d22c31120b40d6a13bbc8cb9230eec as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/info/89d22c31120b40d6a13bbc8cb9230eec 2024-12-02T01:09:22,336 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/info/89d22c31120b40d6a13bbc8cb9230eec, entries=2, sequenceid=6, filesize=4.9 K 2024-12-02T01:09:22,337 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 1e4c9691efacb5eafb00181ffc92f4e6 in 45ms, sequenceid=6, compaction requested=false 2024-12-02T01:09:22,341 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/data/hbase/namespace/1e4c9691efacb5eafb00181ffc92f4e6/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T01:09:22,342 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:09:22,342 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 1e4c9691efacb5eafb00181ffc92f4e6: 2024-12-02T01:09:22,342 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733101713868.1e4c9691efacb5eafb00181ffc92f4e6. 2024-12-02T01:09:22,342 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 61b9a0fcb731850a98b7b912948ae7ef, disabling compactions & flushes 2024-12-02T01:09:22,342 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,342 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,342 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. after waiting 0 ms 2024-12-02T01:09:22,342 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,343 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 61b9a0fcb731850a98b7b912948ae7ef: 2024-12-02T01:09:22,343 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:09:22,409 INFO [RS:1;a82f277fe18c:40483 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,40483,1733101714547; zookeeper connection closed. 2024-12-02T01:09:22,409 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:40483-0x10194df17600003, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:09:22,409 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@7ba191b3 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@7ba191b3 2024-12-02T01:09:22,493 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:09:22,493 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(3579): Received CLOSE for 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:09:22,493 DEBUG [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 61b9a0fcb731850a98b7b912948ae7ef 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 61b9a0fcb731850a98b7b912948ae7ef, disabling compactions & flushes 2024-12-02T01:09:22,493 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:09:22,493 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. after waiting 0 ms 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 61b9a0fcb731850a98b7b912948ae7ef: 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing hbase:meta,,1.1588230740 2024-12-02T01:09:22,493 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing TestLogRolling-testLogRollOnDatanodeDeath,,1733101714687.61b9a0fcb731850a98b7b912948ae7ef. 2024-12-02T01:09:22,634 INFO [regionserver/a82f277fe18c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:09:22,656 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101713317 to hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs/a82f277fe18c%2C43621%2C1733101712797.1733101713317 2024-12-02T01:09:22,658 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101738579 to hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs/a82f277fe18c%2C43621%2C1733101712797.1733101738579 2024-12-02T01:09:22,659 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742625 to hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs/a82f277fe18c%2C43621%2C1733101712797.1733101742625 2024-12-02T01:09:22,660 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.1733101742646 to hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/oldWALs/a82f277fe18c%2C43621%2C1733101712797.1733101742646 2024-12-02T01:09:22,693 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1624): We were exiting though online regions are not empty, because some regions failed closing 2024-12-02T01:09:22,693 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,43621,1733101712797; all regions closed. 2024-12-02T01:09:22,694 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797 2024-12-02T01:09:22,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741868_1052 (size=93) 2024-12-02T01:09:22,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741868_1052 (size=93) 2024-12-02T01:09:23,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741858_1041 (size=1618) 2024-12-02T01:09:23,182 INFO [regionserver/a82f277fe18c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:09:23,184 INFO [regionserver/a82f277fe18c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T01:09:23,184 INFO [regionserver/a82f277fe18c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T01:09:23,582 WARN [org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1@67052e87 {}] datanode.BlockRecoveryWorker$RecoveryTaskContiguous(164): Failed to recover block (block=BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1006, datanode=DatanodeInfoWithStorage[127.0.0.1:43415,null,null]) java.net.ConnectException: Call From a82f277fe18c/172.17.0.2 to localhost:42195 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:961) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:876) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1588) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy55.initReplicaRecovery(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.lambda$initReplicaRecovery$0(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.initReplicaRecovery(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.callInitReplicaRecovery(BlockRecoveryWorker.java:561) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.access$400(BlockRecoveryWorker.java:57) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$RecoveryTaskContiguous.recover(BlockRecoveryWorker.java:135) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1.run(BlockRecoveryWorker.java:602) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:668) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:789) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:364) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getConnection(Client.java:1649) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1473) ~[hadoop-common-3.4.1.jar:?] ... 12 more 2024-12-02T01:09:25,549 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@56a8abb1[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:42011, datanodeUuid=5379d92f-966e-4337-93c6-da8a06d97702, infoPort=43975, infoSecurePort=0, ipcPort=45971, storageInfo=lv=-57;cid=testClusterID;nsid=1162842704;c=1733101710778):Failed to transfer BP-1905452902-172.17.0.2-1733101710778:blk_1073741830_1048 to 127.0.0.1:35113 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:26,307 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta after 4002ms 2024-12-02T01:09:27,346 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,360 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,362 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,362 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,368 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,368 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,699 ERROR [WAL-Shutdown-0 {}] wal.FSHLog(508): We have waited 5 seconds but the close of writer(s) doesn't complete.Please check the status of underlying filesystem or increase the wait time by the config "hbase.wal.fshlog.wait.on.shutdown.seconds" 2024-12-02T01:09:27,700 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797 2024-12-02T01:09:27,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741865_1049 (size=13280) 2024-12-02T01:09:27,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741865_1049 (size=13280) 2024-12-02T01:09:27,706 DEBUG [RS:0;a82f277fe18c:43621 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:27,706 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:09:27,707 INFO [RS:0;a82f277fe18c:43621 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-02T01:09:27,707 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:09:27,707 INFO [RS:0;a82f277fe18c:43621 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:43621 2024-12-02T01:09:27,752 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,43621,1733101712797 2024-12-02T01:09:27,752 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:09:27,797 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,43621,1733101712797] 2024-12-02T01:09:27,798 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,43621,1733101712797; numProcessing=2 2024-12-02T01:09:27,809 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,43621,1733101712797 already deleted, retry=false 2024-12-02T01:09:27,809 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,43621,1733101712797 expired; onlineServers=0 2024-12-02T01:09:27,809 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,39475,1733101712639' ***** 2024-12-02T01:09:27,809 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T01:09:27,809 DEBUG [M:0;a82f277fe18c:39475 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@23369301, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:09:27,809 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,39475,1733101712639 2024-12-02T01:09:27,809 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,39475,1733101712639; all regions closed. 2024-12-02T01:09:27,809 DEBUG [M:0;a82f277fe18c:39475 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:09:27,810 DEBUG [M:0;a82f277fe18c:39475 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T01:09:27,810 DEBUG [M:0;a82f277fe18c:39475 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T01:09:27,810 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T01:09:27,810 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101713092 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101713092,5,FailOnTimeoutGroup] 2024-12-02T01:09:27,810 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101713092 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101713092,5,FailOnTimeoutGroup] 2024-12-02T01:09:27,810 INFO [M:0;a82f277fe18c:39475 {}] hbase.ChoreService(370): Chore service for: master/a82f277fe18c:0 had [] on shutdown 2024-12-02T01:09:27,810 DEBUG [M:0;a82f277fe18c:39475 {}] master.HMaster(1733): Stopping service threads 2024-12-02T01:09:27,810 INFO [M:0;a82f277fe18c:39475 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T01:09:27,810 INFO [M:0;a82f277fe18c:39475 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T01:09:27,811 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T01:09:27,831 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T01:09:27,831 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:27,832 DEBUG [M:0;a82f277fe18c:39475 {}] zookeeper.ZKUtil(347): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T01:09:27,832 WARN [M:0;a82f277fe18c:39475 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T01:09:27,832 INFO [M:0;a82f277fe18c:39475 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-02T01:09:27,832 INFO [M:0;a82f277fe18c:39475 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T01:09:27,832 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:09:27,832 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:09:27,832 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:27,833 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:27,833 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:09:27,833 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:27,833 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=40.08 KB heapSize=49.29 KB 2024-12-02T01:09:27,834 WARN [sync.4 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]] 2024-12-02T01:09:27,835 WARN [sync.4 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]] 2024-12-02T01:09:27,835 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C39475%2C1733101712639:(num 1733101743091) roll requested 2024-12-02T01:09:27,835 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C39475%2C1733101712639.1733101767835 2024-12-02T01:09:27,844 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101743091 with entries=1, filesize=349 B; new WAL /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101767835 2024-12-02T01:09:27,844 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43975:43975),(127.0.0.1/127.0.0.1:35799:35799)] 2024-12-02T01:09:27,844 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(751): hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639/a82f277fe18c%2C39475%2C1733101712639.1733101743091 is not closed yet, will try archiving it next time 2024-12-02T01:09:27,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741864_1047 (size=357) 2024-12-02T01:09:27,853 DEBUG [M:0;a82f277fe18c:39475 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/d60db9dcc9544c388f168c9f2ccc2ff2 is 82, key is hbase:meta,,1/info:regioninfo/1733101713804/Put/seqid=0 2024-12-02T01:09:27,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741871_1056 (size=5672) 2024-12-02T01:09:27,859 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741871_1056 (size=5672) 2024-12-02T01:09:27,859 INFO [M:0;a82f277fe18c:39475 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/d60db9dcc9544c388f168c9f2ccc2ff2 2024-12-02T01:09:27,870 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:09:27,874 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,885 DEBUG [M:0;a82f277fe18c:39475 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/54bdecf0fdb34f5f80e17a5d02196f21 is 774, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733101715121/Put/seqid=0 2024-12-02T01:09:27,887 WARN [Thread-843 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741872_1057 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:27,888 WARN [Thread-843 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741872_1057 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:27,888 WARN [Thread-843 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741872_1057 2024-12-02T01:09:27,888 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,888 WARN [Thread-843 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:27,888 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,889 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,892 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,893 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:27,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741873_1058 (size=7465) 2024-12-02T01:09:27,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741873_1058 (size=7465) 2024-12-02T01:09:27,898 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:09:27,898 INFO [RS:0;a82f277fe18c:43621 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,43621,1733101712797; zookeeper connection closed. 2024-12-02T01:09:27,898 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43621-0x10194df17600001, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:09:27,898 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@425032fd {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@425032fd 2024-12-02T01:09:27,898 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 2 regionserver(s) complete 2024-12-02T01:09:28,297 INFO [M:0;a82f277fe18c:39475 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=39.41 KB at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/54bdecf0fdb34f5f80e17a5d02196f21 2024-12-02T01:09:28,308 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:28,324 DEBUG [M:0;a82f277fe18c:39475 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/555ecd6dacaf4ae587831eee1c19887e is 69, key is a82f277fe18c,40483,1733101714547/rs:state/1733101714606/Put/seqid=0 2024-12-02T01:09:28,327 WARN [Thread-851 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741874_1059 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:35113 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:28,327 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:49262 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741874_1059] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6]'}, localName='127.0.0.1:46881', datanodeUuid='738f7a67-e050-43f5-8a33-713780210261', xmitsInProgress=0}:Exception transferring block BP-1905452902-172.17.0.2-1733101710778:blk_1073741874_1059 to mirror 127.0.0.1:35113 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:28,327 WARN [Thread-851 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741874_1059 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46881,DS-4ee5e85d-c974-4da1-8ea3-cbcd9ef76bc4,DISK], DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:28,327 WARN [Thread-851 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741874_1059 2024-12-02T01:09:28,327 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:49262 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741874_1059] {}] datanode.BlockReceiver(316): Block 1073741874 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-02T01:09:28,327 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_187892269_22 at /127.0.0.1:49262 [Receiving block BP-1905452902-172.17.0.2-1733101710778:blk_1073741874_1059] {}] datanode.DataXceiver(331): 127.0.0.1:46881:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49262 dst: /127.0.0.1:46881 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:28,327 WARN [Thread-851 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:28,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741875_1060 (size=5224) 2024-12-02T01:09:28,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741875_1060 (size=5224) 2024-12-02T01:09:28,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:09:28,735 INFO [M:0;a82f277fe18c:39475 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=130 B at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/555ecd6dacaf4ae587831eee1c19887e 2024-12-02T01:09:28,762 DEBUG [M:0;a82f277fe18c:39475 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/2fdd6d29bd204649809c4d1564df5385 is 52, key is load_balancer_on/state:d/1733101714533/Put/seqid=0 2024-12-02T01:09:28,763 WARN [Thread-858 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741876_1061 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:28,764 WARN [Thread-858 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1905452902-172.17.0.2-1733101710778:blk_1073741876_1061 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK], DatanodeInfoWithStorage[127.0.0.1:42011,DS-4cb95f8e-ca80-4c1e-94a4-ef3c4ceac470,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK]) is bad. 2024-12-02T01:09:28,764 WARN [Thread-858 {}] hdfs.DataStreamer(1850): Abandoning BP-1905452902-172.17.0.2-1733101710778:blk_1073741876_1061 2024-12-02T01:09:28,764 WARN [Thread-858 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:35113,DS-912c962e-d06f-4e15-b497-4bed30d45b30,DISK] 2024-12-02T01:09:28,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741877_1062 (size=5056) 2024-12-02T01:09:28,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741877_1062 (size=5056) 2024-12-02T01:09:28,769 INFO [M:0;a82f277fe18c:39475 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/2fdd6d29bd204649809c4d1564df5385 2024-12-02T01:09:28,775 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/d60db9dcc9544c388f168c9f2ccc2ff2 as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/d60db9dcc9544c388f168c9f2ccc2ff2 2024-12-02T01:09:28,781 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/d60db9dcc9544c388f168c9f2ccc2ff2, entries=8, sequenceid=97, filesize=5.5 K 2024-12-02T01:09:28,782 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/54bdecf0fdb34f5f80e17a5d02196f21 as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/54bdecf0fdb34f5f80e17a5d02196f21 2024-12-02T01:09:28,788 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/54bdecf0fdb34f5f80e17a5d02196f21, entries=11, sequenceid=97, filesize=7.3 K 2024-12-02T01:09:28,789 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/555ecd6dacaf4ae587831eee1c19887e as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/555ecd6dacaf4ae587831eee1c19887e 2024-12-02T01:09:28,795 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/555ecd6dacaf4ae587831eee1c19887e, entries=2, sequenceid=97, filesize=5.1 K 2024-12-02T01:09:28,796 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/2fdd6d29bd204649809c4d1564df5385 as hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/2fdd6d29bd204649809c4d1564df5385 2024-12-02T01:09:28,801 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/2fdd6d29bd204649809c4d1564df5385, entries=1, sequenceid=97, filesize=4.9 K 2024-12-02T01:09:28,802 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(3040): Finished flush of dataSize ~40.08 KB/41039, heapSize ~49.23 KB/50408, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 969ms, sequenceid=97, compaction requested=false 2024-12-02T01:09:28,804 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:28,804 DEBUG [M:0;a82f277fe18c:39475 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:09:28,804 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/MasterData/WALs/a82f277fe18c,39475,1733101712639 2024-12-02T01:09:28,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46881 is added to blk_1073741870_1055 (size=493) 2024-12-02T01:09:28,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:42011 is added to blk_1073741870_1055 (size=493) 2024-12-02T01:09:28,807 INFO [M:0;a82f277fe18c:39475 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-02T01:09:28,807 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:09:28,807 INFO [M:0;a82f277fe18c:39475 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:39475 2024-12-02T01:09:28,817 DEBUG [M:0;a82f277fe18c:39475 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/a82f277fe18c,39475,1733101712639 already deleted, retry=false 2024-12-02T01:09:28,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:09:28,926 INFO [M:0;a82f277fe18c:39475 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,39475,1733101712639; zookeeper connection closed. 2024-12-02T01:09:28,926 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39475-0x10194df17600000, quorum=127.0.0.1:54150, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:09:28,932 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5dbaaf37{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:28,933 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@d79cb{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:09:28,933 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:09:28,934 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@69d8bad2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:09:28,934 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40516419{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:09:28,936 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:09:28,936 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1905452902-172.17.0.2-1733101710778 (Datanode Uuid 5379d92f-966e-4337-93c6-da8a06d97702) service to localhost/127.0.0.1:45673 2024-12-02T01:09:28,936 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:09:28,937 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:09:28,936 WARN [org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1@4995af6b {}] datanode.BlockRecoveryWorker$RecoveryTaskContiguous(164): Failed to recover block (block=BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010, datanode=DatanodeInfoWithStorage[127.0.0.1:43415,null,null]) java.io.InterruptedIOException: DestHost:destPort localhost:42195 , LocalHost:localPort a82f277fe18c/172.17.0.2:0. Failed on local exception: java.io.InterruptedIOException: Interrupted: action=RetryAction(action=RETRY, delayMillis=1000, reason=retries get failed due to exceeded maximum allowed retries number: 10), retry policy=RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS) at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:961) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:936) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1588) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy55.initReplicaRecovery(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.lambda$initReplicaRecovery$0(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.initReplicaRecovery(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.callInitReplicaRecovery(BlockRecoveryWorker.java:561) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.access$400(BlockRecoveryWorker.java:57) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$RecoveryTaskContiguous.recover(BlockRecoveryWorker.java:135) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1.run(BlockRecoveryWorker.java:602) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.InterruptedIOException: Interrupted: action=RetryAction(action=RETRY, delayMillis=1000, reason=retries get failed due to exceeded maximum allowed retries number: 10), retry policy=RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS) at org.apache.hadoop.ipc.Client$Connection.handleConnectionFailure(Client.java:963) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:691) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:789) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:364) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getConnection(Client.java:1649) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1473) ~[hadoop-common-3.4.1.jar:?] ... 12 more Caused by: java.lang.InterruptedException: sleep interrupted at java.lang.Thread.sleep(Native Method) ~[?:?] at org.apache.hadoop.ipc.Client$Connection.handleConnectionFailure(Client.java:961) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:691) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:789) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:364) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getConnection(Client.java:1649) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1473) ~[hadoop-common-3.4.1.jar:?] ... 12 more 2024-12-02T01:09:28,937 WARN [org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1@4995af6b {}] datanode.BlockRecoveryWorker$1(605): recover Block: RecoveringBlock{BP-1905452902-172.17.0.2-1733101710778:blk_1073741834_1010; getBlockSize()=85; corrupt=false; offset=-1; locs=[DatanodeInfoWithStorage[127.0.0.1:42011,null,null], DatanodeInfoWithStorage[127.0.0.1:43415,null,null]]; cachedLocs=[]} FAILED: java.io.IOException: No block pool offer service for bpid=BP-1905452902-172.17.0.2-1733101710778 2024-12-02T01:09:28,937 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data3/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:28,938 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data4/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:28,938 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:09:28,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5447727{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:28,942 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@d776958{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:09:28,942 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:09:28,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@37cda5fa{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:09:28,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2b1e9118{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:09:28,943 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:09:28,943 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:09:28,943 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:09:28,943 WARN [BP-1905452902-172.17.0.2-1733101710778 heartbeating to localhost/127.0.0.1:45673 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1905452902-172.17.0.2-1733101710778 (Datanode Uuid 738f7a67-e050-43f5-8a33-713780210261) service to localhost/127.0.0.1:45673 2024-12-02T01:09:28,944 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data5/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:28,944 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/cluster_e1483482-af87-82e3-6d8d-b3fcf55a8df6/dfs/data/data6/current/BP-1905452902-172.17.0.2-1733101710778 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:28,944 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:09:28,951 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4d8cb8b4{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:09:28,951 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1b1e7792{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:09:28,951 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:09:28,951 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@469fe845{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:09:28,952 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4888cf22{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir/,STOPPED} 2024-12-02T01:09:28,963 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-02T01:09:28,989 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-02T01:09:28,994 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnDatanodeDeath Thread=89 (was 66) Potentially hanging thread: RS-EventLoopGroup-5-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-17-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-3 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45673 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-17-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-5-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-17-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Timer for 'DataNode' metrics system java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: RS-EventLoopGroup-6-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Close-WAL-Writer-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:175) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$$Lambda$797/0x00007fe4b0b87418.run(Unknown Source) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:45673 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-16-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-10-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-6-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.1@localhost:45673 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:45673 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Abort regionserver monitor java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: LeaseRenewer:jenkins@localhost:45673 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45673 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-10-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.2@localhost:45673 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-7-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-11-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-4 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-10-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-16-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45673 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-11-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:45673 from jenkins.hfs.1 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-11-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-7-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-7-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-16-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-5-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-6-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=424 (was 403) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=80 (was 53) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=5259 (was 5826) 2024-12-02T01:09:28,999 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnPipelineRestart Thread=89, OpenFileDescriptor=424, MaxFileDescriptor=1048576, SystemLoadAverage=80, ProcessCount=11, AvailableMemoryMB=5259 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.log.dir so I do NOT create it in target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/8a614a82-91d3-7f06-0dc6-f598fdc84ea3/hadoop.tmp.dir so I do NOT create it in target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad, deleteOnExit=true 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/test.cache.data in system properties and HBase conf 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir in system properties and HBase conf 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T01:09:29,000 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-02T01:09:29,000 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/nfs.dump.dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T01:09:29,001 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T01:09:29,011 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:09:29,309 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:29,323 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:29,327 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:09:29,328 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:09:29,328 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:09:29,328 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:09:29,329 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:29,329 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@40383e9{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:09:29,330 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@49fc0599{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:09:29,416 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@28247350{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-32889-hadoop-hdfs-3_4_1-tests_jar-_-any-9836336446843995337/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:09:29,417 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5b812b08{HTTP/1.1, (http/1.1)}{localhost:32889} 2024-12-02T01:09:29,417 INFO [Time-limited test {}] server.Server(415): Started @176545ms 2024-12-02T01:09:29,427 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:09:29,656 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:29,659 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:09:29,660 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:09:29,660 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:09:29,660 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:09:29,661 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@30668f10{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:09:29,661 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@52941675{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:09:29,758 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1ffe3d71{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-40561-hadoop-hdfs-3_4_1-tests_jar-_-any-15471404000920968877/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:29,758 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@32216334{HTTP/1.1, (http/1.1)}{localhost:40561} 2024-12-02T01:09:29,758 INFO [Time-limited test {}] server.Server(415): Started @176886ms 2024-12-02T01:09:29,759 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:09:29,784 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:29,787 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:09:29,787 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:09:29,788 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:09:29,788 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:09:29,788 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@52c4b57c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:09:29,788 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7837e91d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:09:29,877 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7a20cb9f{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-36601-hadoop-hdfs-3_4_1-tests_jar-_-any-9296340144284941771/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:29,877 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@75bcdbb0{HTTP/1.1, (http/1.1)}{localhost:36601} 2024-12-02T01:09:29,877 INFO [Time-limited test {}] server.Server(415): Started @177006ms 2024-12-02T01:09:29,878 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:09:30,310 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:30,555 WARN [Thread-953 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data2/current/BP-1350002905-172.17.0.2-1733101769022/current, will proceed with Du for space computation calculation, 2024-12-02T01:09:30,555 WARN [Thread-952 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data1/current/BP-1350002905-172.17.0.2-1733101769022/current, will proceed with Du for space computation calculation, 2024-12-02T01:09:30,581 WARN [Thread-916 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:09:30,583 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe4e21d1b64390cf7 with lease ID 0x72717cc0d7589cad: Processing first storage report for DS-6bb55151-2f94-47f3-815c-1718d5945e72 from datanode DatanodeRegistration(127.0.0.1:38409, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=36819, infoSecurePort=0, ipcPort=45233, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022) 2024-12-02T01:09:30,583 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe4e21d1b64390cf7 with lease ID 0x72717cc0d7589cad: from storage DS-6bb55151-2f94-47f3-815c-1718d5945e72 node DatanodeRegistration(127.0.0.1:38409, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=36819, infoSecurePort=0, ipcPort=45233, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:30,583 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xe4e21d1b64390cf7 with lease ID 0x72717cc0d7589cad: Processing first storage report for DS-2d512eb9-320e-4e9b-9be6-9e98cf10c5e5 from datanode DatanodeRegistration(127.0.0.1:38409, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=36819, infoSecurePort=0, ipcPort=45233, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022) 2024-12-02T01:09:30,583 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe4e21d1b64390cf7 with lease ID 0x72717cc0d7589cad: from storage DS-2d512eb9-320e-4e9b-9be6-9e98cf10c5e5 node DatanodeRegistration(127.0.0.1:38409, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=36819, infoSecurePort=0, ipcPort=45233, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:30,701 WARN [Thread-963 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data3/current/BP-1350002905-172.17.0.2-1733101769022/current, will proceed with Du for space computation calculation, 2024-12-02T01:09:30,701 WARN [Thread-964 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data4/current/BP-1350002905-172.17.0.2-1733101769022/current, will proceed with Du for space computation calculation, 2024-12-02T01:09:30,721 WARN [Thread-939 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:09:30,723 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb092781912267aa1 with lease ID 0x72717cc0d7589cae: Processing first storage report for DS-d00c427d-07db-48c7-b2d9-c26fb965db5e from datanode DatanodeRegistration(127.0.0.1:37433, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=38169, infoSecurePort=0, ipcPort=45437, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022) 2024-12-02T01:09:30,723 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb092781912267aa1 with lease ID 0x72717cc0d7589cae: from storage DS-d00c427d-07db-48c7-b2d9-c26fb965db5e node DatanodeRegistration(127.0.0.1:37433, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=38169, infoSecurePort=0, ipcPort=45437, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:30,723 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb092781912267aa1 with lease ID 0x72717cc0d7589cae: Processing first storage report for DS-37271db0-c377-4193-bb8c-fd8ded95f57d from datanode DatanodeRegistration(127.0.0.1:37433, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=38169, infoSecurePort=0, ipcPort=45437, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022) 2024-12-02T01:09:30,723 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb092781912267aa1 with lease ID 0x72717cc0d7589cae: from storage DS-37271db0-c377-4193-bb8c-fd8ded95f57d node DatanodeRegistration(127.0.0.1:37433, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=38169, infoSecurePort=0, ipcPort=45437, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:30,810 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2 2024-12-02T01:09:30,815 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/zookeeper_0, clientPort=61902, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T01:09:30,816 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=61902 2024-12-02T01:09:30,816 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:30,817 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:30,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:09:30,826 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:09:30,827 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c with version=8 2024-12-02T01:09:30,827 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase-staging 2024-12-02T01:09:30,829 INFO [Time-limited test {}] client.ConnectionUtils(129): master/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:09:30,829 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:09:30,829 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:09:30,829 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:09:30,829 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:09:30,829 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:09:30,829 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:09:30,830 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:09:30,830 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:38809 2024-12-02T01:09:30,831 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:30,832 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:30,834 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:38809 connecting to ZooKeeper ensemble=127.0.0.1:61902 2024-12-02T01:09:30,894 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:388090x0, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:09:30,895 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:38809-0x10194dffab30000 connected 2024-12-02T01:09:30,959 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:09:30,961 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:09:30,961 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:09:30,962 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38809 2024-12-02T01:09:30,963 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38809 2024-12-02T01:09:30,963 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38809 2024-12-02T01:09:30,964 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38809 2024-12-02T01:09:30,964 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38809 2024-12-02T01:09:30,965 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c, hbase.cluster.distributed=false 2024-12-02T01:09:30,984 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:09:30,984 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:09:30,985 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:38681 2024-12-02T01:09:30,985 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:09:30,986 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:09:30,986 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:30,988 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:30,990 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:38681 connecting to ZooKeeper ensemble=127.0.0.1:61902 2024-12-02T01:09:31,000 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:386810x0, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:09:31,001 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:09:31,001 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:38681-0x10194dffab30001 connected 2024-12-02T01:09:31,002 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:09:31,002 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:09:31,002 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38681 2024-12-02T01:09:31,003 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38681 2024-12-02T01:09:31,003 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38681 2024-12-02T01:09:31,003 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38681 2024-12-02T01:09:31,003 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38681 2024-12-02T01:09:31,004 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,014 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:09:31,014 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:09:31,015 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,016 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a82f277fe18c:38809 2024-12-02T01:09:31,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:09:31,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:09:31,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,023 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,023 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:09:31,024 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a82f277fe18c,38809,1733101770828 from backup master directory 2024-12-02T01:09:31,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:09:31,034 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:09:31,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:09:31,034 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:09:31,034 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,044 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:09:31,045 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:09:31,046 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/hbase.id with ID: 3684e04b-6021-4dcb-b4a1-38a6d79013d4 2024-12-02T01:09:31,061 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:31,073 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,073 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:09:31,080 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:09:31,081 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:09:31,082 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T01:09:31,082 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:09:31,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:09:31,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:09:31,093 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store 2024-12-02T01:09:31,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:09:31,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:09:31,100 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:31,100 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:09:31,100 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:31,100 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:31,100 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:09:31,100 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:31,100 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:09:31,100 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:09:31,101 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/.initializing 2024-12-02T01:09:31,101 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,103 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C38809%2C1733101770828, suffix=, logDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828, archiveDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/oldWALs, maxLogs=10 2024-12-02T01:09:31,104 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38809%2C1733101770828.1733101771104 2024-12-02T01:09:31,108 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 2024-12-02T01:09:31,109 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38169:38169),(127.0.0.1/127.0.0.1:36819:36819)] 2024-12-02T01:09:31,109 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:09:31,109 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:31,109 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,109 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,110 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,112 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T01:09:31,112 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,112 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,113 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,114 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T01:09:31,114 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,114 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:09:31,114 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,116 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T01:09:31,116 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,116 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:09:31,116 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,118 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T01:09:31,118 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,118 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:09:31,119 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,120 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,122 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T01:09:31,123 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:09:31,125 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:09:31,126 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=818481, jitterRate=0.04075288772583008}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T01:09:31,126 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:09:31,127 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T01:09:31,130 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@688803b1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:09:31,131 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-02T01:09:31,131 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T01:09:31,131 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T01:09:31,131 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T01:09:31,132 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T01:09:31,132 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-02T01:09:31,132 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T01:09:31,134 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T01:09:31,135 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T01:09:31,142 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-02T01:09:31,142 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T01:09:31,143 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T01:09:31,150 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-02T01:09:31,151 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T01:09:31,152 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T01:09:31,159 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-02T01:09:31,159 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T01:09:31,167 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T01:09:31,169 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T01:09:31,175 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T01:09:31,184 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:09:31,184 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:09:31,184 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,184 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,184 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=a82f277fe18c,38809,1733101770828, sessionid=0x10194dffab30000, setting cluster-up flag (Was=false) 2024-12-02T01:09:31,200 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,200 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,225 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T01:09:31,226 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,242 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,242 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,267 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T01:09:31,269 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,38809,1733101770828 2024-12-02T01:09:31,273 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-02T01:09:31,274 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-02T01:09:31,274 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T01:09:31,274 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a82f277fe18c,38809,1733101770828 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a82f277fe18c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,275 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:09:31,276 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733101801277 2024-12-02T01:09:31,278 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T01:09:31,278 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T01:09:31,278 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T01:09:31,278 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T01:09:31,278 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T01:09:31,278 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T01:09:31,278 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:09:31,279 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-02T01:09:31,279 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,280 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T01:09:31,280 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T01:09:31,280 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T01:09:31,281 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T01:09:31,281 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T01:09:31,281 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,281 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101771281,5,FailOnTimeoutGroup] 2024-12-02T01:09:31,282 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101771281,5,FailOnTimeoutGroup] 2024-12-02T01:09:31,282 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,281 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:09:31,282 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T01:09:31,282 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,282 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:09:31,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:09:31,292 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-02T01:09:31,292 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c 2024-12-02T01:09:31,300 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:09:31,300 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:09:31,301 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:31,302 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:09:31,303 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:09:31,303 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,303 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,304 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:09:31,305 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:09:31,305 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,305 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,305 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:09:31,306 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:09:31,306 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,307 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,308 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740 2024-12-02T01:09:31,308 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740 2024-12-02T01:09:31,309 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:09:31,310 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:09:31,310 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:31,312 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:09:31,312 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=815153, jitterRate=0.03652188181877136}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:09:31,312 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:09:31,313 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:09:31,313 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:09:31,313 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:09:31,313 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:09:31,313 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:09:31,313 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:09:31,313 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:09:31,314 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:09:31,314 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-02T01:09:31,314 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T01:09:31,315 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a82f277fe18c:38681 2024-12-02T01:09:31,315 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T01:09:31,316 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T01:09:31,316 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1008): ClusterId : 3684e04b-6021-4dcb-b4a1-38a6d79013d4 2024-12-02T01:09:31,316 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:09:31,326 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:09:31,326 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:09:31,335 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:09:31,335 DEBUG [RS:0;a82f277fe18c:38681 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7a9e8abd, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:09:31,336 DEBUG [RS:0;a82f277fe18c:38681 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5200e076, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:09:31,336 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:09:31,336 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:09:31,336 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:09:31,336 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,38809,1733101770828 with isa=a82f277fe18c/172.17.0.2:38681, startcode=1733101770983 2024-12-02T01:09:31,337 DEBUG [RS:0;a82f277fe18c:38681 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:09:31,340 INFO [RS-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37663, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:09:31,341 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38809 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,341 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38809 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,343 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c 2024-12-02T01:09:31,343 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:33641 2024-12-02T01:09:31,343 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:09:31,350 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:09:31,351 DEBUG [RS:0;a82f277fe18c:38681 {}] zookeeper.ZKUtil(111): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,351 WARN [RS:0;a82f277fe18c:38681 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:09:31,351 INFO [RS:0;a82f277fe18c:38681 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:09:31,351 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,351 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,38681,1733101770983] 2024-12-02T01:09:31,354 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:09:31,354 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:09:31,355 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:09:31,355 INFO [RS:0;a82f277fe18c:38681 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:09:31,356 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,356 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:09:31,356 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:09:31,357 DEBUG [RS:0;a82f277fe18c:38681 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:09:31,358 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,358 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,358 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,358 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,358 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38681,1733101770983-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:09:31,370 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:09:31,370 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38681,1733101770983-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:31,381 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.Replication(204): a82f277fe18c,38681,1733101770983 started 2024-12-02T01:09:31,382 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,38681,1733101770983, RpcServer on a82f277fe18c/172.17.0.2:38681, sessionid=0x10194dffab30001 2024-12-02T01:09:31,382 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:09:31,382 DEBUG [RS:0;a82f277fe18c:38681 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,382 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,38681,1733101770983' 2024-12-02T01:09:31,382 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:09:31,382 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:09:31,383 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:09:31,383 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:09:31,383 DEBUG [RS:0;a82f277fe18c:38681 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,383 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,38681,1733101770983' 2024-12-02T01:09:31,383 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:09:31,383 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:09:31,384 DEBUG [RS:0;a82f277fe18c:38681 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:09:31,384 INFO [RS:0;a82f277fe18c:38681 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:09:31,384 INFO [RS:0;a82f277fe18c:38681 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:09:31,466 WARN [a82f277fe18c:38809 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-02T01:09:31,488 INFO [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C38681%2C1733101770983, suffix=, logDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983, archiveDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/oldWALs, maxLogs=32 2024-12-02T01:09:31,490 INFO [RS:0;a82f277fe18c:38681 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:09:31,498 INFO [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:09:31,498 DEBUG [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36819:36819),(127.0.0.1/127.0.0.1:38169:38169)] 2024-12-02T01:09:31,717 DEBUG [a82f277fe18c:38809 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T01:09:31,718 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,721 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,38681,1733101770983, state=OPENING 2024-12-02T01:09:31,764 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T01:09:31,776 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,776 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:31,776 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=a82f277fe18c,38681,1733101770983}] 2024-12-02T01:09:31,776 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:09:31,776 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:09:31,930 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,930 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T01:09:31,932 INFO [RS-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58498, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T01:09:31,937 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-02T01:09:31,937 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:09:31,940 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C38681%2C1733101770983.meta, suffix=.meta, logDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983, archiveDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/oldWALs, maxLogs=32 2024-12-02T01:09:31,941 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta 2024-12-02T01:09:31,947 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta 2024-12-02T01:09:31,948 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36819:36819),(127.0.0.1/127.0.0.1:38169:38169)] 2024-12-02T01:09:31,948 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:09:31,948 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T01:09:31,948 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T01:09:31,948 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T01:09:31,948 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T01:09:31,949 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:31,949 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-02T01:09:31,949 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-02T01:09:31,950 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:09:31,951 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:09:31,951 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,952 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,952 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:09:31,953 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:09:31,953 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,954 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,954 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:09:31,955 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:09:31,955 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:31,955 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:09:31,956 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740 2024-12-02T01:09:31,957 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740 2024-12-02T01:09:31,959 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:09:31,960 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:09:31,961 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=776042, jitterRate=-0.013212651014328003}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:09:31,961 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:09:31,962 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733101771929 2024-12-02T01:09:31,964 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T01:09:31,964 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-02T01:09:31,964 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:31,965 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,38681,1733101770983, state=OPEN 2024-12-02T01:09:31,995 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:09:31,995 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:09:31,995 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:09:31,995 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:09:31,997 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T01:09:31,997 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=a82f277fe18c,38681,1733101770983 in 219 msec 2024-12-02T01:09:31,999 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T01:09:31,999 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 683 msec 2024-12-02T01:09:32,001 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 728 msec 2024-12-02T01:09:32,001 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733101772001, completionTime=-1 2024-12-02T01:09:32,001 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T01:09:32,002 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-02T01:09:32,002 DEBUG [hconnection-0x220b5530-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:09:32,003 INFO [RS-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58502, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:09:32,008 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-02T01:09:32,009 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733101832009 2024-12-02T01:09:32,009 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733101892009 2024-12-02T01:09:32,009 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 7 msec 2024-12-02T01:09:32,031 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38809,1733101770828-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:32,031 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38809,1733101770828-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:32,031 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38809,1733101770828-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:32,032 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a82f277fe18c:38809, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:32,032 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:32,032 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-02T01:09:32,032 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:09:32,033 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-02T01:09:32,034 DEBUG [master/a82f277fe18c:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-02T01:09:32,035 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:09:32,035 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:32,037 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:09:32,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:09:32,046 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:09:32,048 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => bb4bec346b76ea94b266353bd821d234, NAME => 'hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c 2024-12-02T01:09:32,054 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:09:32,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:09:32,055 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:32,055 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing bb4bec346b76ea94b266353bd821d234, disabling compactions & flushes 2024-12-02T01:09:32,055 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,055 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,055 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. after waiting 0 ms 2024-12-02T01:09:32,055 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,056 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,056 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for bb4bec346b76ea94b266353bd821d234: 2024-12-02T01:09:32,057 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:09:32,057 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733101772057"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101772057"}]},"ts":"1733101772057"} 2024-12-02T01:09:32,060 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:09:32,061 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:09:32,061 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101772061"}]},"ts":"1733101772061"} 2024-12-02T01:09:32,063 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-02T01:09:32,081 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=bb4bec346b76ea94b266353bd821d234, ASSIGN}] 2024-12-02T01:09:32,084 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=bb4bec346b76ea94b266353bd821d234, ASSIGN 2024-12-02T01:09:32,086 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=bb4bec346b76ea94b266353bd821d234, ASSIGN; state=OFFLINE, location=a82f277fe18c,38681,1733101770983; forceNewPlan=false, retain=false 2024-12-02T01:09:32,236 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=bb4bec346b76ea94b266353bd821d234, regionState=OPENING, regionLocation=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:32,239 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure bb4bec346b76ea94b266353bd821d234, server=a82f277fe18c,38681,1733101770983}] 2024-12-02T01:09:32,311 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:32,393 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,38681,1733101770983 2024-12-02T01:09:32,401 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,401 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => bb4bec346b76ea94b266353bd821d234, NAME => 'hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:09:32,402 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,402 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:32,402 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,402 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,406 INFO [StoreOpener-bb4bec346b76ea94b266353bd821d234-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,408 INFO [StoreOpener-bb4bec346b76ea94b266353bd821d234-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region bb4bec346b76ea94b266353bd821d234 columnFamilyName info 2024-12-02T01:09:32,409 DEBUG [StoreOpener-bb4bec346b76ea94b266353bd821d234-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:32,409 INFO [StoreOpener-bb4bec346b76ea94b266353bd821d234-1 {}] regionserver.HStore(327): Store=bb4bec346b76ea94b266353bd821d234/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:09:32,410 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,411 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,414 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for bb4bec346b76ea94b266353bd821d234 2024-12-02T01:09:32,417 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:09:32,418 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened bb4bec346b76ea94b266353bd821d234; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=785502, jitterRate=-0.0011830031871795654}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:09:32,419 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for bb4bec346b76ea94b266353bd821d234: 2024-12-02T01:09:32,419 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234., pid=6, masterSystemTime=1733101772392 2024-12-02T01:09:32,421 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,421 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:09:32,422 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=bb4bec346b76ea94b266353bd821d234, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:32,426 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T01:09:32,426 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure bb4bec346b76ea94b266353bd821d234, server=a82f277fe18c,38681,1733101770983 in 184 msec 2024-12-02T01:09:32,428 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T01:09:32,428 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=bb4bec346b76ea94b266353bd821d234, ASSIGN in 345 msec 2024-12-02T01:09:32,428 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:09:32,429 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101772428"}]},"ts":"1733101772428"} 2024-12-02T01:09:32,430 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-02T01:09:32,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-02T01:09:32,440 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:09:32,442 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 408 msec 2024-12-02T01:09:32,448 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:32,448 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:09:32,448 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:09:32,452 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-02T01:09:32,467 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:09:32,479 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 26 msec 2024-12-02T01:09:32,486 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-02T01:09:32,506 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:09:32,517 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 31 msec 2024-12-02T01:09:32,542 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-02T01:09:32,559 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-02T01:09:32,559 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.525sec 2024-12-02T01:09:32,559 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T01:09:32,559 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T01:09:32,559 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T01:09:32,559 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T01:09:32,560 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T01:09:32,560 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38809,1733101770828-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:09:32,560 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38809,1733101770828-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T01:09:32,563 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-02T01:09:32,563 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T01:09:32,563 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,38809,1733101770828-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:09:32,609 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x56d05aee to 127.0.0.1:61902 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@1776decf 2024-12-02T01:09:32,624 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4a65b582, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:09:32,627 DEBUG [hconnection-0x6bc49aa7-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:09:32,629 INFO [RS-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:58512, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:09:32,632 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=a82f277fe18c,38809,1733101770828 2024-12-02T01:09:32,632 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:09:32,638 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-02T01:09:32,639 INFO [Time-limited test {}] wal.TestLogRolling(297): Starting testLogRollOnPipelineRestart 2024-12-02T01:09:32,639 INFO [Time-limited test {}] wal.TestLogRolling(300): Replication=2 2024-12-02T01:09:32,640 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T01:09:32,643 INFO [RS-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33720, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T01:09:32,645 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T01:09:32,645 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T01:09:32,646 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testLogRollOnPipelineRestart', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:09:32,648 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T01:09:32,649 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:09:32,650 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:32,650 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRollOnPipelineRestart" procId is: 9 2024-12-02T01:09:32,651 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:09:32,651 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:09:32,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741837_1013 (size=395) 2024-12-02T01:09:32,660 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741837_1013 (size=395) 2024-12-02T01:09:32,662 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => baede2678bc7a8d2454cccc8b7cb1738, NAME => 'TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRollOnPipelineRestart', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c 2024-12-02T01:09:32,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:38409 is added to blk_1073741838_1014 (size=78) 2024-12-02T01:09:32,668 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37433 is added to blk_1073741838_1014 (size=78) 2024-12-02T01:09:32,669 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:32,669 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1681): Closing baede2678bc7a8d2454cccc8b7cb1738, disabling compactions & flushes 2024-12-02T01:09:32,669 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:32,669 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:32,669 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. after waiting 0 ms 2024-12-02T01:09:32,669 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:32,669 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:32,669 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1635): Region close journal for baede2678bc7a8d2454cccc8b7cb1738: 2024-12-02T01:09:32,670 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:09:32,671 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738.","families":{"info":[{"qualifier":"regioninfo","vlen":77,"tag":[],"timestamp":"1733101772671"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101772671"}]},"ts":"1733101772671"} 2024-12-02T01:09:32,672 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:09:32,673 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:09:32,673 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnPipelineRestart","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101772673"}]},"ts":"1733101772673"} 2024-12-02T01:09:32,674 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnPipelineRestart, state=ENABLING in hbase:meta 2024-12-02T01:09:32,692 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=baede2678bc7a8d2454cccc8b7cb1738, ASSIGN}] 2024-12-02T01:09:32,693 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=baede2678bc7a8d2454cccc8b7cb1738, ASSIGN 2024-12-02T01:09:32,694 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=baede2678bc7a8d2454cccc8b7cb1738, ASSIGN; state=OFFLINE, location=a82f277fe18c,38681,1733101770983; forceNewPlan=false, retain=false 2024-12-02T01:09:32,845 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=baede2678bc7a8d2454cccc8b7cb1738, regionState=OPENING, regionLocation=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:32,847 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure baede2678bc7a8d2454cccc8b7cb1738, server=a82f277fe18c,38681,1733101770983}] 2024-12-02T01:09:33,000 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to a82f277fe18c,38681,1733101770983 2024-12-02T01:09:33,009 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:33,009 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => baede2678bc7a8d2454cccc8b7cb1738, NAME => 'TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:09:33,010 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRollOnPipelineRestart baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,010 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:09:33,011 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,011 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,013 INFO [StoreOpener-baede2678bc7a8d2454cccc8b7cb1738-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,014 INFO [StoreOpener-baede2678bc7a8d2454cccc8b7cb1738-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region baede2678bc7a8d2454cccc8b7cb1738 columnFamilyName info 2024-12-02T01:09:33,014 DEBUG [StoreOpener-baede2678bc7a8d2454cccc8b7cb1738-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:09:33,015 INFO [StoreOpener-baede2678bc7a8d2454cccc8b7cb1738-1 {}] regionserver.HStore(327): Store=baede2678bc7a8d2454cccc8b7cb1738/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:09:33,016 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,016 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,019 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:09:33,021 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:09:33,022 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened baede2678bc7a8d2454cccc8b7cb1738; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=859778, jitterRate=0.09326478838920593}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:09:33,023 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for baede2678bc7a8d2454cccc8b7cb1738: 2024-12-02T01:09:33,023 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738., pid=11, masterSystemTime=1733101773000 2024-12-02T01:09:33,025 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:33,025 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:33,026 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=baede2678bc7a8d2454cccc8b7cb1738, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,38681,1733101770983 2024-12-02T01:09:33,029 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-02T01:09:33,029 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure baede2678bc7a8d2454cccc8b7cb1738, server=a82f277fe18c,38681,1733101770983 in 180 msec 2024-12-02T01:09:33,032 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-02T01:09:33,032 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=baede2678bc7a8d2454cccc8b7cb1738, ASSIGN in 337 msec 2024-12-02T01:09:33,032 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:09:33,033 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnPipelineRestart","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101773032"}]},"ts":"1733101773032"} 2024-12-02T01:09:33,034 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnPipelineRestart, state=ENABLED in hbase:meta 2024-12-02T01:09:33,044 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:09:33,046 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart in 398 msec 2024-12-02T01:09:33,312 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:33,860 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:09:33,876 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,876 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,876 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,876 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,877 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,877 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,880 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:33,883 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:09:34,314 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:35,315 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:36,316 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:37,317 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:37,354 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-02T01:09:37,355 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRollOnPipelineRestart' 2024-12-02T01:09:38,319 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:38,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:09:38,353 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2024-12-02T01:09:38,355 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T01:09:38,355 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart Metrics about Tables on a single HBase RegionServer 2024-12-02T01:09:39,320 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:40,321 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:41,322 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:42,323 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:42,654 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38809 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:09:42,655 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testLogRollOnPipelineRestart, procId: 9 completed 2024-12-02T01:09:42,661 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T01:09:42,661 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:09:43,325 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:44,327 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:44,670 INFO [Time-limited test {}] wal.TestLogRolling(337): log.getCurrentFileName()): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:09:44,670 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:44,671 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010 java.io.IOException: Bad response ERROR for BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010 from datanode DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:44,671 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009 java.io.IOException: Bad response ERROR for BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009 from datanode DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:44,671 WARN [DataStreamer for file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 block BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006 in pipeline [DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK], DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]) is bad. 2024-12-02T01:09:44,672 WARN [DataStreamer for file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta block BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010 in pipeline [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK], DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]) is bad. 2024-12-02T01:09:44,672 WARN [DataStreamer for file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 block BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009 in pipeline [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK], DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:37433,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]) is bad. 2024-12-02T01:09:44,672 WARN [PacketResponder: BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:37433] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,672 WARN [PacketResponder: BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:37433] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Broken pipe at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,672 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1509379763_22 at /127.0.0.1:41662 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:37433:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41662 dst: /127.0.0.1:37433 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,673 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1509379763_22 at /127.0.0.1:47546 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:38409:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47546 dst: /127.0.0.1:38409 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,673 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:47574 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:38409:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47574 dst: /127.0.0.1:38409 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,673 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:41688 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:37433:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41688 dst: /127.0.0.1:37433 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,673 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:47600 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:38409:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:47600 dst: /127.0.0.1:38409 java.net.SocketException: Connection reset at sun.nio.ch.SocketChannelImpl.throwConnectionReset(SocketChannelImpl.java:394) ~[?:?] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:426) ~[?:?] at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,674 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:41698 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:37433:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:41698 dst: /127.0.0.1:37433 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,723 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1350002905-172.17.0.2-1733101769022 (Datanode Uuid 2b70d962-240b-4745-bfbd-25674b8d24ff) service to localhost/127.0.0.1:33641 2024-12-02T01:09:44,725 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data3/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:44,725 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data4/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:44,727 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7a20cb9f{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:44,727 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@75bcdbb0{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:09:44,728 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:09:44,728 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7837e91d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:09:44,728 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@52c4b57c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:09:44,732 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:09:44,741 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:44,744 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:09:44,745 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:09:44,745 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:09:44,745 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:09:44,745 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@97053c7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:09:44,745 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@41a0f667{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:09:44,834 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@14df5fce{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-33763-hadoop-hdfs-3_4_1-tests_jar-_-any-3511425741054889672/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:44,834 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@c03e161{HTTP/1.1, (http/1.1)}{localhost:33763} 2024-12-02T01:09:44,834 INFO [Time-limited test {}] server.Server(415): Started @191963ms 2024-12-02T01:09:44,836 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:09:44,849 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1015 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1015 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:44,849 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1017 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1017 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:44,849 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1016 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1016 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:44,850 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:42766 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:38409:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:42766 dst: /127.0.0.1:38409 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,850 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1509379763_22 at /127.0.0.1:42760 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:38409:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:42760 dst: /127.0.0.1:38409 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,850 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:42778 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:38409:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:42778 dst: /127.0.0.1:38409 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:09:44,851 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1ffe3d71{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:44,852 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@32216334{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:09:44,852 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:09:44,852 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@52941675{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:09:44,852 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@30668f10{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:09:44,853 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:09:44,853 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:09:44,853 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:09:44,853 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1350002905-172.17.0.2-1733101769022 (Datanode Uuid defaffc4-c594-427d-96cb-b1299bacf9e2) service to localhost/127.0.0.1:33641 2024-12-02T01:09:44,854 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data1/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:44,854 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data2/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:09:44,854 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:09:44,864 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:09:44,867 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:09:44,867 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:09:44,867 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:09:44,868 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:09:44,868 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7c5eda2b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:09:44,868 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3135f9f9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:09:44,958 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@89935a2{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-39883-hadoop-hdfs-3_4_1-tests_jar-_-any-3645323683921769765/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:09:44,958 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@ad2e865{HTTP/1.1, (http/1.1)}{localhost:39883} 2024-12-02T01:09:44,958 INFO [Time-limited test {}] server.Server(415): Started @192087ms 2024-12-02T01:09:44,959 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:09:45,137 WARN [Thread-1098 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:09:45,139 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc3fa5c9f5f157275 with lease ID 0x72717cc0d7589caf: from storage DS-d00c427d-07db-48c7-b2d9-c26fb965db5e node DatanodeRegistration(127.0.0.1:41439, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=42355, infoSecurePort=0, ipcPort=37389, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 7, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T01:09:45,140 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc3fa5c9f5f157275 with lease ID 0x72717cc0d7589caf: from storage DS-37271db0-c377-4193-bb8c-fd8ded95f57d node DatanodeRegistration(127.0.0.1:41439, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=42355, infoSecurePort=0, ipcPort=37389, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:45,328 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:45,347 WARN [Thread-1118 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:09:45,349 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbdd9036c9298f588 with lease ID 0x72717cc0d7589cb0: from storage DS-6bb55151-2f94-47f3-815c-1718d5945e72 node DatanodeRegistration(127.0.0.1:36903, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=44143, infoSecurePort=0, ipcPort=37167, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:45,349 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbdd9036c9298f588 with lease ID 0x72717cc0d7589cb0: from storage DS-2d512eb9-320e-4e9b-9be6-9e98cf10c5e5 node DatanodeRegistration(127.0.0.1:36903, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=44143, infoSecurePort=0, ipcPort=37167, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:09:45,977 INFO [Time-limited test {}] wal.TestLogRolling(349): Data Nodes restarted 2024-12-02T01:09:45,982 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1002 2024-12-02T01:09:45,984 WARN [RS:0;a82f277fe18c:38681.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=5, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:45,985 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C38681%2C1733101770983:(num 1733101771490) roll requested 2024-12-02T01:09:45,985 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38681 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=5, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:45,985 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:09:45,986 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38681 {}] ipc.CallRunner(138): callId: 11 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:58512 deadline: 1733101795983, exception=org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=5, requesting roll of WAL 2024-12-02T01:09:45,992 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 newFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:09:45,992 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=5, requesting roll of WAL 2024-12-02T01:09:45,993 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 with entries=5, filesize=2.09 KB; new WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:09:45,993 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44143:44143),(127.0.0.1/127.0.0.1:42355:42355)] 2024-12-02T01:09:45,993 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 is not closed yet, will try archiving it next time 2024-12-02T01:09:45,993 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:45,993 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:09:45,993 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:09:45,994 WARN [IPC Server handler 1 on default port 33641 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 has not been closed. Lease recovery is in progress. RecoveryId = 1019 for block blk_1073741833_1016 2024-12-02T01:09:45,994 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 after 1ms 2024-12-02T01:09:46,330 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:47,331 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:48,141 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741833_1016: GenerationStamp not matched, existing replica is blk_1073741833_1009 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T01:09:48,333 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:49,334 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:49,995 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 after 4002ms 2024-12-02T01:09:50,335 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:51,335 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:52,336 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:53,337 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:54,337 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:55,338 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:56,339 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:57,340 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:58,085 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1003 2024-12-02T01:09:58,341 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:09:59,342 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:00,091 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:00,091 WARN [DataStreamer for file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 block BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36903,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK], DatanodeInfoWithStorage[127.0.0.1:41439,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:36903,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]) is bad. 2024-12-02T01:10:00,091 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:55168 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018] {}] datanode.DataXceiver(331): 127.0.0.1:36903:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:55168 dst: /127.0.0.1:36903 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:10:00,092 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:60896 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018] {}] datanode.DataXceiver(331): 127.0.0.1:41439:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:60896 dst: /127.0.0.1:41439 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:10:00,140 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@89935a2{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:00,140 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@ad2e865{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:10:00,140 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:10:00,140 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3135f9f9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:10:00,140 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7c5eda2b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:10:00,142 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:10:00,142 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:10:00,142 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:10:00,142 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1350002905-172.17.0.2-1733101769022 (Datanode Uuid defaffc4-c594-427d-96cb-b1299bacf9e2) service to localhost/127.0.0.1:33641 2024-12-02T01:10:00,142 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data1/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:00,143 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data2/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:00,143 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:10:00,150 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:10:00,153 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:10:00,153 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:10:00,153 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:10:00,154 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:10:00,154 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@940e980{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:10:00,155 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4a7dd207{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:10:00,244 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6400f8b0{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-46801-hadoop-hdfs-3_4_1-tests_jar-_-any-16493938918337686782/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:00,245 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1c7c62c8{HTTP/1.1, (http/1.1)}{localhost:46801} 2024-12-02T01:10:00,245 INFO [Time-limited test {}] server.Server(415): Started @207373ms 2024-12-02T01:10:00,247 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:10:00,262 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1020 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1020 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:00,263 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-1504474152_22 at /127.0.0.1:56208 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741839_1018] {}] datanode.DataXceiver(331): 127.0.0.1:41439:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:56208 dst: /127.0.0.1:41439 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:10:00,265 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@14df5fce{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:00,265 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@c03e161{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:10:00,265 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:10:00,265 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@41a0f667{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:10:00,265 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@97053c7{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:10:00,266 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:10:00,266 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:10:00,267 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1350002905-172.17.0.2-1733101769022 (Datanode Uuid 2b70d962-240b-4745-bfbd-25674b8d24ff) service to localhost/127.0.0.1:33641 2024-12-02T01:10:00,267 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:10:00,267 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data3/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:00,267 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data4/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:00,267 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:10:00,275 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:10:00,279 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:10:00,280 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:10:00,280 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:10:00,280 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:10:00,281 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@414849b2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:10:00,281 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2973b329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:10:00,343 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:00,374 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4692c3b3{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/java.io.tmpdir/jetty-localhost-43419-hadoop-hdfs-3_4_1-tests_jar-_-any-15700506380780068878/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:00,375 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1cd578ad{HTTP/1.1, (http/1.1)}{localhost:43419} 2024-12-02T01:10:00,375 INFO [Time-limited test {}] server.Server(415): Started @207503ms 2024-12-02T01:10:00,376 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:10:00,632 WARN [Thread-1173 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:10:00,634 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xca60e802e514bfcf with lease ID 0x72717cc0d7589cb1: from storage DS-6bb55151-2f94-47f3-815c-1718d5945e72 node DatanodeRegistration(127.0.0.1:33325, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=38063, infoSecurePort=0, ipcPort=39333, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 7, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T01:10:00,635 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xca60e802e514bfcf with lease ID 0x72717cc0d7589cb1: from storage DS-2d512eb9-320e-4e9b-9be6-9e98cf10c5e5 node DatanodeRegistration(127.0.0.1:33325, datanodeUuid=defaffc4-c594-427d-96cb-b1299bacf9e2, infoPort=38063, infoSecurePort=0, ipcPort=39333, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 8, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:00,753 WARN [Thread-1193 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:10:00,756 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe450dbe86389632e with lease ID 0x72717cc0d7589cb2: from storage DS-d00c427d-07db-48c7-b2d9-c26fb965db5e node DatanodeRegistration(127.0.0.1:35235, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=40965, infoSecurePort=0, ipcPort=39259, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:00,756 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xe450dbe86389632e with lease ID 0x72717cc0d7589cb2: from storage DS-37271db0-c377-4193-bb8c-fd8ded95f57d node DatanodeRegistration(127.0.0.1:35235, datanodeUuid=2b70d962-240b-4745-bfbd-25674b8d24ff, infoPort=40965, infoSecurePort=0, ipcPort=39259, storageInfo=lv=-57;cid=testClusterID;nsid=189119714;c=1733101769022), blocks: 8, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:00,809 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:10:01,278 WARN [master/a82f277fe18c:0:becomeActiveMaster.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=95, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,278 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C38809%2C1733101770828:(num 1733101771104) roll requested 2024-12-02T01:10:01,278 ERROR [ProcExecTimeout {}] region.RegionProcedureStore(422): Failed to delete pids=[4, 7, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,278 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38809%2C1733101770828.1733101801278 2024-12-02T01:10:01,278 ERROR [ProcExecTimeout {}] procedure2.TimeoutExecutorThread(124): Ignoring pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner exception: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL java.io.UncheckedIOException: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL at org.apache.hadoop.hbase.procedure2.store.region.RegionProcedureStore.delete(RegionProcedureStore.java:423) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner.periodicExecute(CompletedProcedureCleaner.java:135) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.executeInMemoryChore(TimeoutExecutorThread.java:122) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.execDelayedProcedure(TimeoutExecutorThread.java:101) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.run(TimeoutExecutorThread.java:68) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] Caused by: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,286 WARN [master:store-WAL-Roller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL 2024-12-02T01:10:01,287 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 with entries=92, filesize=45.98 KB; new WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101801278 2024-12-02T01:10:01,287 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38063:38063),(127.0.0.1/127.0.0.1:40965:40965)] 2024-12-02T01:10:01,287 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(751): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 is not closed yet, will try archiving it next time 2024-12-02T01:10:01,287 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,287 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,288 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 2024-12-02T01:10:01,288 WARN [IPC Server handler 1 on default port 33641 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 has not been closed. Lease recovery is in progress. RecoveryId = 1022 for block blk_1073741830_1015 2024-12-02T01:10:01,289 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 after 1ms 2024-12-02T01:10:01,343 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:01,393 INFO [Time-limited test {}] wal.TestLogRolling(366): Data Nodes restarted 2024-12-02T01:10:01,396 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1004 2024-12-02T01:10:01,398 WARN [RS:0;a82f277fe18c:38681.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=8, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:41439,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,398 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C38681%2C1733101770983:(num 1733101785985) roll requested 2024-12-02T01:10:01,399 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38681 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=8, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:41439,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,399 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.1733101801399 2024-12-02T01:10:01,399 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38681 {}] ipc.CallRunner(138): callId: 18 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:58512 deadline: 1733101811397, exception=org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=8, requesting roll of WAL 2024-12-02T01:10:01,406 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 newFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 2024-12-02T01:10:01,407 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=8, requesting roll of WAL 2024-12-02T01:10:01,407 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 with entries=2, filesize=2.37 KB; new WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 2024-12-02T01:10:01,407 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40965:40965),(127.0.0.1/127.0.0.1:38063:38063)] 2024-12-02T01:10:01,407 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 is not closed yet, will try archiving it next time 2024-12-02T01:10:01,407 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:41439,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,407 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:41439,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:01,407 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:10:01,408 WARN [IPC Server handler 1 on default port 33641 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 has not been closed. Lease recovery is in progress. RecoveryId = 1024 for block blk_1073741839_1020 2024-12-02T01:10:01,408 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 after 1ms 2024-12-02T01:10:02,345 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:02,637 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741839_1020: GenerationStamp not matched, existing replica is blk_1073741839_1018 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T01:10:03,336 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:10:03,338 INFO [RS-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36488, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:10:03,346 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:04,347 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:05,055 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 61b9a0fcb731850a98b7b912948ae7ef, had cached 0 bytes from a total of 23930 2024-12-02T01:10:05,289 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828/a82f277fe18c%2C38809%2C1733101770828.1733101771104 after 4001ms 2024-12-02T01:10:05,348 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:05,409 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 after 4002ms 2024-12-02T01:10:05,757 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 2) replica(s): 0) Failed to delete replica blk_1073741830_1015: GenerationStamp not matched, existing replica is blk_1073741830_1006 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T01:10:06,348 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:07,349 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:08,351 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:09,352 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:10,353 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:11,353 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:12,354 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:13,093 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T01:10:13,093 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T01:10:13,355 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:13,444 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:13,453 DEBUG [Time-limited test {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 newFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:13,454 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 with entries=1, filesize=1.23 KB; new WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:13,454 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:38063:38063),(127.0.0.1/127.0.0.1:40965:40965)] 2024-12-02T01:10:13,454 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 is not closed yet, will try archiving it next time 2024-12-02T01:10:13,455 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:10:13,455 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:10:13,456 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 after 0ms 2024-12-02T01:10:13,456 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:10:13,457 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741841_1023 (size=1264) 2024-12-02T01:10:13,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741841_1023 (size=1264) 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #3: [\x00/METAFAMILY:HBASE::REGION_EVENT::REGION_OPEN/1733101772419/Put/vlen=162/seqid=0] 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #4: [default/info:d/1733101772457/Put/vlen=9/seqid=0] 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #5: [hbase/info:d/1733101772494/Put/vlen=7/seqid=0] 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #3: [\x00/METAFAMILY:HBASE::REGION_EVENT::REGION_OPEN/1733101773023/Put/vlen=218/seqid=0] 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #4: [row1002/info:/1733101782666/Put/vlen=1045/seqid=0] 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(396): EOF reading file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101771490 2024-12-02T01:10:13,466 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:10:13,466 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:10:13,467 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 after 1ms 2024-12-02T01:10:13,467 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:10:13,470 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #6: [row1003/info:/1733101796079/Put/vlen=1045/seqid=0] 2024-12-02T01:10:13,470 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #7: [row1004/info:/1733101798087/Put/vlen=1045/seqid=0] 2024-12-02T01:10:13,470 DEBUG [Time-limited test {}] wal.TestLogRolling(396): EOF reading file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101785985 2024-12-02T01:10:13,470 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 2024-12-02T01:10:13,470 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 2024-12-02T01:10:13,470 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 after 0ms 2024-12-02T01:10:13,471 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101801399 2024-12-02T01:10:13,474 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #9: [row1005/info:/1733101811441/Put/vlen=1045/seqid=0] 2024-12-02T01:10:13,474 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:13,474 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:13,474 WARN [IPC Server handler 0 on default port 33641 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 has not been closed. Lease recovery is in progress. RecoveryId = 1026 for block blk_1073741842_1025 2024-12-02T01:10:13,475 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 after 1ms 2024-12-02T01:10:13,764 WARN [ResponseProcessor for block BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:13,764 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1509379763_22 at /127.0.0.1:58664 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025] {}] datanode.DataXceiver(331): 127.0.0.1:35235:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:58664 dst: /127.0.0.1:35235 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:10:13,764 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1509379763_22 at /127.0.0.1:56236 [Receiving block BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025] {}] datanode.DataXceiver(331): 127.0.0.1:33325:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:56236 dst: /127.0.0.1:33325 java.io.InterruptedIOException: Interrupted while waiting for IO on channel java.nio.channels.SocketChannel[connected local=localhost/127.0.0.1:33325 remote=/127.0.0.1:56236]. Total timeout mills is 60000, 59688 millis timeout left. at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:350) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:10:13,764 WARN [DataStreamer for file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 block BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 {}] hdfs.DataStreamer(1731): Error Recovery for BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 in pipeline [DatanodeInfoWithStorage[127.0.0.1:33325,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK], DatanodeInfoWithStorage[127.0.0.1:35235,DS-d00c427d-07db-48c7-b2d9-c26fb965db5e,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:33325,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]) is bad. 2024-12-02T01:10:13,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741842_1026 (size=85) 2024-12-02T01:10:13,770 WARN [DataStreamer for file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 block BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 {}] hdfs.DataStreamer(859): DataStreamer Exception org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy43.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:14,356 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:15,358 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:16,359 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:17,360 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:17,476 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 after 4002ms 2024-12-02T01:10:17,477 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:17,486 DEBUG [Time-limited test {}] wal.TestLogRolling(396): EOF reading file /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:17,486 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing bb4bec346b76ea94b266353bd821d234 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-02T01:10:17,487 WARN [RS:0;a82f277fe18c:38681.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=7, requesting roll of WAL org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy43.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,487 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C38681%2C1733101770983:(num 1733101813444) roll requested 2024-12-02T01:10:17,487 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for bb4bec346b76ea94b266353bd821d234: 2024-12-02T01:10:17,488 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.1733101817488 2024-12-02T01:10:17,488 INFO [Time-limited test {}] wal.TestLogRolling(416): org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=7, requesting roll of WAL org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=7, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy43.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,489 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.90 KB heapSize=5.42 KB 2024-12-02T01:10:17,489 WARN [RS_OPEN_META-regionserver/a82f277fe18c:0-0.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=15, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,490 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 1588230740: 2024-12-02T01:10:17,490 INFO [Time-limited test {}] wal.TestLogRolling(416): org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,490 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing baede2678bc7a8d2454cccc8b7cb1738 1/1 column families, dataSize=4.20 KB heapSize=4.75 KB 2024-12-02T01:10:17,490 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for baede2678bc7a8d2454cccc8b7cb1738: 2024-12-02T01:10:17,491 INFO [Time-limited test {}] wal.TestLogRolling(416): org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=7, requesting roll of WAL org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=7, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy43.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,493 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-02T01:10:17,493 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-02T01:10:17,493 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x56d05aee to 127.0.0.1:61902 2024-12-02T01:10:17,493 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:10:17,493 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T01:10:17,493 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1890160825, stopped=false 2024-12-02T01:10:17,493 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=a82f277fe18c,38809,1733101770828 2024-12-02T01:10:17,494 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 newFile=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101817488 2024-12-02T01:10:17,494 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=7, requesting roll of WAL 2024-12-02T01:10:17,495 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101817488 2024-12-02T01:10:17,495 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40965:40965),(127.0.0.1/127.0.0.1:38063:38063)] 2024-12-02T01:10:17,495 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 is not closed yet, will try archiving it next time 2024-12-02T01:10:17,495 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog a82f277fe18c%2C38681%2C1733101770983.meta:.meta(num 1733101771941) roll requested 2024-12-02T01:10:17,495 INFO [regionserver/a82f277fe18c:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C38681%2C1733101770983.meta.1733101817495.meta 2024-12-02T01:10:17,495 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy43.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,495 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-1350002905-172.17.0.2-1733101769022:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy43.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy44.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor115.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,495 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:17,496 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 after 1ms 2024-12-02T01:10:17,496 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.1733101813444 to hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/oldWALs/a82f277fe18c%2C38681%2C1733101770983.1733101813444 2024-12-02T01:10:17,500 WARN [regionserver/a82f277fe18c:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL 2024-12-02T01:10:17,500 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta with entries=11, filesize=3.66 KB; new WAL /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101817495.meta 2024-12-02T01:10:17,500 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40965:40965),(127.0.0.1/127.0.0.1:38063:38063)] 2024-12-02T01:10:17,500 DEBUG [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta is not closed yet, will try archiving it next time 2024-12-02T01:10:17,500 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,500 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:38409,DS-6bb55151-2f94-47f3-815c-1718d5945e72,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-02T01:10:17,500 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta 2024-12-02T01:10:17,501 WARN [IPC Server handler 1 on default port 33641 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta has not been closed. Lease recovery is in progress. RecoveryId = 1029 for block blk_1073741834_1017 2024-12-02T01:10:17,501 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta after 1ms 2024-12-02T01:10:17,550 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:10:17,550 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:10:17,551 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-02T01:10:17,551 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:17,551 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:17,551 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:10:17,551 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,38681,1733101770983' ***** 2024-12-02T01:10:17,551 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:10:17,551 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:10:17,551 INFO [RS:0;a82f277fe18c:38681 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:10:17,551 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:10:17,551 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:10:17,551 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:10:17,551 INFO [RS:0;a82f277fe18c:38681 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(3579): Received CLOSE for bb4bec346b76ea94b266353bd821d234 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(3579): Received CLOSE for baede2678bc7a8d2454cccc8b7cb1738 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,38681,1733101770983 2024-12-02T01:10:17,552 DEBUG [RS:0;a82f277fe18c:38681 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:10:17,552 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing bb4bec346b76ea94b266353bd821d234, disabling compactions & flushes 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:10:17,552 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:10:17,552 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:10:17,552 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:10:17,552 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. after waiting 0 ms 2024-12-02T01:10:17,552 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:10:17,552 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing bb4bec346b76ea94b266353bd821d234 1/1 column families, dataSize=78 B heapSize=728 B 2024-12-02T01:10:17,553 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-02T01:10:17,553 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1603): Online Regions={bb4bec346b76ea94b266353bd821d234=hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234., 1588230740=hbase:meta,,1.1588230740, baede2678bc7a8d2454cccc8b7cb1738=TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738.} 2024-12-02T01:10:17,553 WARN [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-02T01:10:17,553 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:10:17,553 DEBUG [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, baede2678bc7a8d2454cccc8b7cb1738, bb4bec346b76ea94b266353bd821d234 2024-12-02T01:10:17,553 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:10:17,553 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:10:17,553 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:10:17,553 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:10:17,553 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.90 KB heapSize=5.89 KB 2024-12-02T01:10:17,553 WARN [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-02T01:10:17,553 WARN [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-02T01:10:17,570 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/.tmp/info/0e681bc4590840f19b28685b4d699e74 is 45, key is default/info:d/1733101772457/Put/seqid=0 2024-12-02T01:10:17,570 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/.tmp/info/84d336ea2c4a499199a04e281487bd51 is 207, key is TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738./info:regioninfo/1733101773026/Put/seqid=0 2024-12-02T01:10:17,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741846_1031 (size=8268) 2024-12-02T01:10:17,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741845_1030 (size=5037) 2024-12-02T01:10:17,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741846_1031 (size=8268) 2024-12-02T01:10:17,575 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741845_1030 (size=5037) 2024-12-02T01:10:17,575 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=8 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/.tmp/info/0e681bc4590840f19b28685b4d699e74 2024-12-02T01:10:17,575 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.66 KB at sequenceid=16 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/.tmp/info/84d336ea2c4a499199a04e281487bd51 2024-12-02T01:10:17,582 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/.tmp/info/0e681bc4590840f19b28685b4d699e74 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/info/0e681bc4590840f19b28685b4d699e74 2024-12-02T01:10:17,588 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/info/0e681bc4590840f19b28685b4d699e74, entries=2, sequenceid=8, filesize=4.9 K 2024-12-02T01:10:17,589 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for bb4bec346b76ea94b266353bd821d234 in 36ms, sequenceid=8, compaction requested=false 2024-12-02T01:10:17,593 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/namespace/bb4bec346b76ea94b266353bd821d234/recovered.edits/11.seqid, newMaxSeqId=11, maxSeqId=1 2024-12-02T01:10:17,593 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:10:17,594 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for bb4bec346b76ea94b266353bd821d234: 2024-12-02T01:10:17,594 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733101772032.bb4bec346b76ea94b266353bd821d234. 2024-12-02T01:10:17,594 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing baede2678bc7a8d2454cccc8b7cb1738, disabling compactions & flushes 2024-12-02T01:10:17,594 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:10:17,594 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:10:17,594 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. after waiting 0 ms 2024-12-02T01:10:17,594 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:10:17,594 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing baede2678bc7a8d2454cccc8b7cb1738 1/1 column families, dataSize=4.20 KB heapSize=4.98 KB 2024-12-02T01:10:17,595 WARN [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-02T01:10:17,598 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/.tmp/table/7ec885cac484488ca063469154e37039 is 79, key is TestLogRolling-testLogRollOnPipelineRestart/table:state/1733101773032/Put/seqid=0 2024-12-02T01:10:17,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741847_1032 (size=5482) 2024-12-02T01:10:17,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741847_1032 (size=5482) 2024-12-02T01:10:17,609 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=244 B at sequenceid=16 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/.tmp/table/7ec885cac484488ca063469154e37039 2024-12-02T01:10:17,615 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/.tmp/info/84d336ea2c4a499199a04e281487bd51 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/info/84d336ea2c4a499199a04e281487bd51 2024-12-02T01:10:17,618 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/.tmp/info/b42408a3326f4b3a8f8f447c6ed09176 is 1080, key is row1002/info:/1733101782666/Put/seqid=0 2024-12-02T01:10:17,622 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/info/84d336ea2c4a499199a04e281487bd51, entries=20, sequenceid=16, filesize=8.1 K 2024-12-02T01:10:17,623 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/.tmp/table/7ec885cac484488ca063469154e37039 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/table/7ec885cac484488ca063469154e37039 2024-12-02T01:10:17,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741848_1033 (size=9270) 2024-12-02T01:10:17,630 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/table/7ec885cac484488ca063469154e37039, entries=4, sequenceid=16, filesize=5.4 K 2024-12-02T01:10:17,630 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741848_1033 (size=9270) 2024-12-02T01:10:17,630 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=4.20 KB at sequenceid=12 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/.tmp/info/b42408a3326f4b3a8f8f447c6ed09176 2024-12-02T01:10:17,631 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~2.90 KB/2972, heapSize ~5.14 KB/5264, currentSize=0 B/0 for 1588230740 in 78ms, sequenceid=16, compaction requested=false 2024-12-02T01:10:17,636 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/hbase/meta/1588230740/recovered.edits/19.seqid, newMaxSeqId=19, maxSeqId=1 2024-12-02T01:10:17,636 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T01:10:17,636 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:10:17,636 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:10:17,637 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T01:10:17,638 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/.tmp/info/b42408a3326f4b3a8f8f447c6ed09176 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/info/b42408a3326f4b3a8f8f447c6ed09176 2024-12-02T01:10:17,645 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/info/b42408a3326f4b3a8f8f447c6ed09176, entries=4, sequenceid=12, filesize=9.1 K 2024-12-02T01:10:17,646 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~4.20 KB/4304, heapSize ~4.73 KB/4848, currentSize=0 B/0 for baede2678bc7a8d2454cccc8b7cb1738 in 52ms, sequenceid=12, compaction requested=false 2024-12-02T01:10:17,650 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/data/default/TestLogRolling-testLogRollOnPipelineRestart/baede2678bc7a8d2454cccc8b7cb1738/recovered.edits/15.seqid, newMaxSeqId=15, maxSeqId=1 2024-12-02T01:10:17,651 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:10:17,651 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for baede2678bc7a8d2454cccc8b7cb1738: 2024-12-02T01:10:17,651 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733101772645.baede2678bc7a8d2454cccc8b7cb1738. 2024-12-02T01:10:17,753 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,38681,1733101770983; all regions closed. 2024-12-02T01:10:17,753 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983 2024-12-02T01:10:17,755 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741844_1028 (size=761) 2024-12-02T01:10:17,755 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741844_1028 (size=761) 2024-12-02T01:10:18,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:10:18,354 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart 2024-12-02T01:10:18,358 INFO [regionserver/a82f277fe18c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-02T01:10:18,358 INFO [regionserver/a82f277fe18c:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-02T01:10:18,361 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:19,360 INFO [regionserver/a82f277fe18c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:10:19,362 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:20,363 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:20,760 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741834_1017: GenerationStamp not matched, existing replica is blk_1073741834_1010 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-02T01:10:21,363 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:21,502 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta after 4002ms 2024-12-02T01:10:21,502 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta to hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/oldWALs/a82f277fe18c%2C38681%2C1733101770983.meta.1733101771941.meta 2024-12-02T01:10:21,506 DEBUG [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/oldWALs 2024-12-02T01:10:21,506 INFO [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C38681%2C1733101770983.meta:.meta(num 1733101817495) 2024-12-02T01:10:21,507 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/WALs/a82f277fe18c,38681,1733101770983 2024-12-02T01:10:21,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741843_1027 (size=1979) 2024-12-02T01:10:21,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741843_1027 (size=1979) 2024-12-02T01:10:21,517 DEBUG [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(1071): Moved 4 WAL file(s) to /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/oldWALs 2024-12-02T01:10:21,517 INFO [RS:0;a82f277fe18c:38681 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C38681%2C1733101770983:(num 1733101817488) 2024-12-02T01:10:21,517 DEBUG [RS:0;a82f277fe18c:38681 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:10:21,517 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:10:21,518 INFO [RS:0;a82f277fe18c:38681 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-02T01:10:21,518 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:10:21,518 INFO [RS:0;a82f277fe18c:38681 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:38681 2024-12-02T01:10:21,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,38681,1733101770983 2024-12-02T01:10:21,546 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:10:21,559 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,38681,1733101770983] 2024-12-02T01:10:21,559 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,38681,1733101770983; numProcessing=1 2024-12-02T01:10:21,567 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,38681,1733101770983 already deleted, retry=false 2024-12-02T01:10:21,567 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,38681,1733101770983 expired; onlineServers=0 2024-12-02T01:10:21,567 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,38809,1733101770828' ***** 2024-12-02T01:10:21,567 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T01:10:21,568 DEBUG [M:0;a82f277fe18c:38809 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@67879a97, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:10:21,568 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,38809,1733101770828 2024-12-02T01:10:21,568 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,38809,1733101770828; all regions closed. 2024-12-02T01:10:21,568 DEBUG [M:0;a82f277fe18c:38809 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:10:21,568 DEBUG [M:0;a82f277fe18c:38809 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T01:10:21,568 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T01:10:21,568 DEBUG [M:0;a82f277fe18c:38809 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T01:10:21,568 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101771281 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101771281,5,FailOnTimeoutGroup] 2024-12-02T01:10:21,568 INFO [M:0;a82f277fe18c:38809 {}] hbase.ChoreService(370): Chore service for: master/a82f277fe18c:0 had [] on shutdown 2024-12-02T01:10:21,568 DEBUG [M:0;a82f277fe18c:38809 {}] master.HMaster(1733): Stopping service threads 2024-12-02T01:10:21,568 INFO [M:0;a82f277fe18c:38809 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T01:10:21,569 INFO [M:0;a82f277fe18c:38809 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T01:10:21,569 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101771281 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101771281,5,FailOnTimeoutGroup] 2024-12-02T01:10:21,569 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T01:10:21,579 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T01:10:21,579 DEBUG [M:0;a82f277fe18c:38809 {}] zookeeper.ZKUtil(347): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T01:10:21,579 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:21,579 WARN [M:0;a82f277fe18c:38809 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T01:10:21,579 INFO [M:0;a82f277fe18c:38809 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-02T01:10:21,579 INFO [M:0;a82f277fe18c:38809 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T01:10:21,580 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:10:21,580 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:10:21,580 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:21,580 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:21,580 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:10:21,580 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:21,580 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=40.08 KB heapSize=49.23 KB 2024-12-02T01:10:21,596 DEBUG [M:0;a82f277fe18c:38809 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e169c25c5277432aba09264a59cae3fe is 82, key is hbase:meta,,1/info:regioninfo/1733101771964/Put/seqid=0 2024-12-02T01:10:21,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741849_1034 (size=5672) 2024-12-02T01:10:21,600 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741849_1034 (size=5672) 2024-12-02T01:10:21,601 INFO [M:0;a82f277fe18c:38809 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e169c25c5277432aba09264a59cae3fe 2024-12-02T01:10:21,624 DEBUG [M:0;a82f277fe18c:38809 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/bafe106e266548208f861c794941ca93 is 777, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733101773045/Put/seqid=0 2024-12-02T01:10:21,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741850_1035 (size=7468) 2024-12-02T01:10:21,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741850_1035 (size=7468) 2024-12-02T01:10:21,635 INFO [M:0;a82f277fe18c:38809 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=39.48 KB at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/bafe106e266548208f861c794941ca93 2024-12-02T01:10:21,658 DEBUG [M:0;a82f277fe18c:38809 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6fc172a52ca94e9cbdc670e7b2ca8a28 is 69, key is a82f277fe18c,38681,1733101770983/rs:state/1733101771341/Put/seqid=0 2024-12-02T01:10:21,659 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:10:21,659 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:38681-0x10194dffab30001, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:10:21,659 INFO [RS:0;a82f277fe18c:38681 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,38681,1733101770983; zookeeper connection closed. 2024-12-02T01:10:21,659 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@d03e16a {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@d03e16a 2024-12-02T01:10:21,660 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T01:10:21,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741851_1036 (size=5156) 2024-12-02T01:10:21,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741851_1036 (size=5156) 2024-12-02T01:10:21,664 INFO [M:0;a82f277fe18c:38809 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6fc172a52ca94e9cbdc670e7b2ca8a28 2024-12-02T01:10:21,692 DEBUG [M:0;a82f277fe18c:38809 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/81d7188ac65c4182be3b0137326b28d9 is 52, key is load_balancer_on/state:d/1733101772635/Put/seqid=0 2024-12-02T01:10:21,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741852_1037 (size=5056) 2024-12-02T01:10:21,698 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741852_1037 (size=5056) 2024-12-02T01:10:21,698 INFO [M:0;a82f277fe18c:38809 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/81d7188ac65c4182be3b0137326b28d9 2024-12-02T01:10:21,704 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/e169c25c5277432aba09264a59cae3fe as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/e169c25c5277432aba09264a59cae3fe 2024-12-02T01:10:21,710 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/e169c25c5277432aba09264a59cae3fe, entries=8, sequenceid=96, filesize=5.5 K 2024-12-02T01:10:21,712 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/bafe106e266548208f861c794941ca93 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/bafe106e266548208f861c794941ca93 2024-12-02T01:10:21,720 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/bafe106e266548208f861c794941ca93, entries=11, sequenceid=96, filesize=7.3 K 2024-12-02T01:10:21,722 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6fc172a52ca94e9cbdc670e7b2ca8a28 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/6fc172a52ca94e9cbdc670e7b2ca8a28 2024-12-02T01:10:21,729 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/6fc172a52ca94e9cbdc670e7b2ca8a28, entries=1, sequenceid=96, filesize=5.0 K 2024-12-02T01:10:21,731 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/81d7188ac65c4182be3b0137326b28d9 as hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/81d7188ac65c4182be3b0137326b28d9 2024-12-02T01:10:21,737 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33641/user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/81d7188ac65c4182be3b0137326b28d9, entries=1, sequenceid=96, filesize=4.9 K 2024-12-02T01:10:21,738 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(3040): Finished flush of dataSize ~40.08 KB/41040, heapSize ~49.16 KB/50344, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 158ms, sequenceid=96, compaction requested=false 2024-12-02T01:10:21,740 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:21,740 DEBUG [M:0;a82f277fe18c:38809 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:10:21,740 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9263deb4-ea0f-b6ed-6291-8fe16c41af3c/MasterData/WALs/a82f277fe18c,38809,1733101770828 2024-12-02T01:10:21,742 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35235 is added to blk_1073741840_1021 (size=757) 2024-12-02T01:10:21,742 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:33325 is added to blk_1073741840_1021 (size=757) 2024-12-02T01:10:21,742 INFO [M:0;a82f277fe18c:38809 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-02T01:10:21,742 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:10:21,743 INFO [M:0;a82f277fe18c:38809 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:38809 2024-12-02T01:10:21,750 DEBUG [M:0;a82f277fe18c:38809 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/a82f277fe18c,38809,1733101770828 already deleted, retry=false 2024-12-02T01:10:21,859 INFO [M:0;a82f277fe18c:38809 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,38809,1733101770828; zookeeper connection closed. 2024-12-02T01:10:21,859 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:10:21,859 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38809-0x10194dffab30000, quorum=127.0.0.1:61902, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:10:21,862 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4692c3b3{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:21,862 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1cd578ad{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:10:21,862 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:10:21,862 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2973b329{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:10:21,863 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@414849b2{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:10:21,864 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:10:21,864 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:10:21,864 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1350002905-172.17.0.2-1733101769022 (Datanode Uuid 2b70d962-240b-4745-bfbd-25674b8d24ff) service to localhost/127.0.0.1:33641 2024-12-02T01:10:21,864 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:10:21,865 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data3/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:21,865 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data4/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:21,866 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:10:21,872 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6400f8b0{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:21,872 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1c7c62c8{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:10:21,872 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:10:21,872 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4a7dd207{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:10:21,873 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@940e980{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:10:21,874 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:10:21,874 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:10:21,874 WARN [BP-1350002905-172.17.0.2-1733101769022 heartbeating to localhost/127.0.0.1:33641 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1350002905-172.17.0.2-1733101769022 (Datanode Uuid defaffc4-c594-427d-96cb-b1299bacf9e2) service to localhost/127.0.0.1:33641 2024-12-02T01:10:21,874 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:10:21,875 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data1/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:21,875 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/cluster_d18183f1-9d2e-849f-2350-ed6a83a30aad/dfs/data/data2/current/BP-1350002905-172.17.0.2-1733101769022 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:10:21,875 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:10:21,881 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@28247350{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:10:21,882 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5b812b08{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:10:21,882 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:10:21,882 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@49fc0599{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:10:21,882 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@40383e9{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir/,STOPPED} 2024-12-02T01:10:21,889 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-02T01:10:21,912 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-02T01:10:21,917 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnPipelineRestart Thread=103 (was 89) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:33641 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33641 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-26-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-29-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-9-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-26-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-28-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-6 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:33641 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33641 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-29-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-8-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-28-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-29-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-27-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:33641 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RS-EventLoopGroup-8-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-27-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-8-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-27-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-9-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (356647748) connection to localhost/127.0.0.1:33641 from jenkins.hfs.3 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-5 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-26-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.3@localhost:33641 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-9-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-28-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Command processor java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33641 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=450 (was 424) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=181 (was 80) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=6082 (was 5259) - AvailableMemoryMB LEAK? - 2024-12-02T01:10:21,923 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testCompactionRecordDoesntBlockRolling Thread=103, OpenFileDescriptor=450, MaxFileDescriptor=1048576, SystemLoadAverage=181, ProcessCount=11, AvailableMemoryMB=6082 2024-12-02T01:10:21,923 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T01:10:21,923 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.log.dir so I do NOT create it in target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556 2024-12-02T01:10:21,923 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/92371595-69da-0572-bc84-0e0c68dd32f2/hadoop.tmp.dir so I do NOT create it in target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556 2024-12-02T01:10:21,923 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360, deleteOnExit=true 2024-12-02T01:10:21,923 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/test.cache.data in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-02T01:10:21,924 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T01:10:21,924 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/nfs.dump.dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/java.io.tmpdir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T01:10:21,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T01:10:21,937 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:10:22,287 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:10:22,293 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:10:22,295 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:10:22,295 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:10:22,296 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:10:22,299 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:10:22,299 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5c883b08{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:10:22,300 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@dc1c4c3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:10:22,364 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:22,389 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2896003d{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/java.io.tmpdir/jetty-localhost-34827-hadoop-hdfs-3_4_1-tests_jar-_-any-6555076149298536283/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:10:22,389 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5f1edc2d{HTTP/1.1, (http/1.1)}{localhost:34827} 2024-12-02T01:10:22,389 INFO [Time-limited test {}] server.Server(415): Started @229518ms 2024-12-02T01:10:22,401 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:10:22,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,603 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,603 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,606 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,606 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,606 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,607 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,615 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,615 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,616 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,616 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,667 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:10:22,668 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,668 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,668 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,668 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,701 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,702 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,702 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,702 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,703 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,703 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,708 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,708 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,708 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,711 WARN [Time-limited test {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:22,721 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:10:22,725 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:10:22,726 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:10:22,726 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:10:22,726 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:10:22,727 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@23e7f2aa{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:10:22,727 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@30d4a328{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:10:22,829 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4cd1618{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/java.io.tmpdir/jetty-localhost-45361-hadoop-hdfs-3_4_1-tests_jar-_-any-10824790019955316186/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:22,829 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4f062b37{HTTP/1.1, (http/1.1)}{localhost:45361} 2024-12-02T01:10:22,830 INFO [Time-limited test {}] server.Server(415): Started @229958ms 2024-12-02T01:10:22,831 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:10:22,872 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:10:22,876 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:10:22,877 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:10:22,877 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:10:22,877 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:10:22,878 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5c2ef9ae{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:10:22,879 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2b0f6823{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:10:22,980 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@62cfa257{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/java.io.tmpdir/jetty-localhost-43841-hadoop-hdfs-3_4_1-tests_jar-_-any-3085131149189322554/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:10:22,981 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4bc05900{HTTP/1.1, (http/1.1)}{localhost:43841} 2024-12-02T01:10:22,981 INFO [Time-limited test {}] server.Server(415): Started @230109ms 2024-12-02T01:10:22,982 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:10:23,365 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:23,912 WARN [Thread-1411 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data2/current/BP-2004375065-172.17.0.2-1733101821951/current, will proceed with Du for space computation calculation, 2024-12-02T01:10:23,912 WARN [Thread-1410 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data1/current/BP-2004375065-172.17.0.2-1733101821951/current, will proceed with Du for space computation calculation, 2024-12-02T01:10:23,927 WARN [Thread-1374 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:10:23,930 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x178fe930a24249ab with lease ID 0xf54f39d097e76fb0: Processing first storage report for DS-ec391860-6cec-4745-8590-fb5cf46b94f0 from datanode DatanodeRegistration(127.0.0.1:46591, datanodeUuid=bf2c66f6-ccb2-45df-8a73-eaac12153c2d, infoPort=44943, infoSecurePort=0, ipcPort=38773, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951) 2024-12-02T01:10:23,930 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x178fe930a24249ab with lease ID 0xf54f39d097e76fb0: from storage DS-ec391860-6cec-4745-8590-fb5cf46b94f0 node DatanodeRegistration(127.0.0.1:46591, datanodeUuid=bf2c66f6-ccb2-45df-8a73-eaac12153c2d, infoPort=44943, infoSecurePort=0, ipcPort=38773, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:23,930 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x178fe930a24249ab with lease ID 0xf54f39d097e76fb0: Processing first storage report for DS-7574d65c-71d6-4429-8a18-aa1d091ea636 from datanode DatanodeRegistration(127.0.0.1:46591, datanodeUuid=bf2c66f6-ccb2-45df-8a73-eaac12153c2d, infoPort=44943, infoSecurePort=0, ipcPort=38773, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951) 2024-12-02T01:10:23,930 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x178fe930a24249ab with lease ID 0xf54f39d097e76fb0: from storage DS-7574d65c-71d6-4429-8a18-aa1d091ea636 node DatanodeRegistration(127.0.0.1:46591, datanodeUuid=bf2c66f6-ccb2-45df-8a73-eaac12153c2d, infoPort=44943, infoSecurePort=0, ipcPort=38773, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:24,033 WARN [Thread-1421 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data3/current/BP-2004375065-172.17.0.2-1733101821951/current, will proceed with Du for space computation calculation, 2024-12-02T01:10:24,033 WARN [Thread-1422 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data4/current/BP-2004375065-172.17.0.2-1733101821951/current, will proceed with Du for space computation calculation, 2024-12-02T01:10:24,047 WARN [Thread-1397 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:10:24,050 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xeae41af1119fe2d5 with lease ID 0xf54f39d097e76fb1: Processing first storage report for DS-bc710a76-0aa4-46db-9438-a2293c779f50 from datanode DatanodeRegistration(127.0.0.1:41581, datanodeUuid=dd9ba3b6-9e54-4e3d-9d8b-4afc7a487faa, infoPort=39881, infoSecurePort=0, ipcPort=39891, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951) 2024-12-02T01:10:24,050 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xeae41af1119fe2d5 with lease ID 0xf54f39d097e76fb1: from storage DS-bc710a76-0aa4-46db-9438-a2293c779f50 node DatanodeRegistration(127.0.0.1:41581, datanodeUuid=dd9ba3b6-9e54-4e3d-9d8b-4afc7a487faa, infoPort=39881, infoSecurePort=0, ipcPort=39891, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:24,050 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xeae41af1119fe2d5 with lease ID 0xf54f39d097e76fb1: Processing first storage report for DS-d9dfa725-2caf-4e43-bc19-fdae857f2423 from datanode DatanodeRegistration(127.0.0.1:41581, datanodeUuid=dd9ba3b6-9e54-4e3d-9d8b-4afc7a487faa, infoPort=39881, infoSecurePort=0, ipcPort=39891, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951) 2024-12-02T01:10:24,050 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xeae41af1119fe2d5 with lease ID 0xf54f39d097e76fb1: from storage DS-d9dfa725-2caf-4e43-bc19-fdae857f2423 node DatanodeRegistration(127.0.0.1:41581, datanodeUuid=dd9ba3b6-9e54-4e3d-9d8b-4afc7a487faa, infoPort=39881, infoSecurePort=0, ipcPort=39891, storageInfo=lv=-57;cid=testClusterID;nsid=39249748;c=1733101821951), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:10:24,123 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556 2024-12-02T01:10:24,125 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/zookeeper_0, clientPort=51080, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T01:10:24,126 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=51080 2024-12-02T01:10:24,127 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,129 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:10:24,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:10:24,139 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9 with version=8 2024-12-02T01:10:24,139 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase-staging 2024-12-02T01:10:24,141 INFO [Time-limited test {}] client.ConnectionUtils(129): master/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:10:24,141 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:10:24,142 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:39899 2024-12-02T01:10:24,142 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,143 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,145 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:39899 connecting to ZooKeeper ensemble=127.0.0.1:51080 2024-12-02T01:10:24,208 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:398990x0, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:10:24,208 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:39899-0x10194e0caf20000 connected 2024-12-02T01:10:24,275 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:10:24,275 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:10:24,276 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:10:24,276 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39899 2024-12-02T01:10:24,276 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39899 2024-12-02T01:10:24,277 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39899 2024-12-02T01:10:24,277 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39899 2024-12-02T01:10:24,277 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39899 2024-12-02T01:10:24,278 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9, hbase.cluster.distributed=false 2024-12-02T01:10:24,300 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:10:24,301 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:10:24,303 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:35559 2024-12-02T01:10:24,303 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:10:24,305 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:10:24,306 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,308 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,312 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:35559 connecting to ZooKeeper ensemble=127.0.0.1:51080 2024-12-02T01:10:24,322 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:355590x0, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:10:24,322 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:355590x0, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:10:24,322 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35559-0x10194e0caf20001 connected 2024-12-02T01:10:24,323 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:10:24,323 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:10:24,326 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35559 2024-12-02T01:10:24,326 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35559 2024-12-02T01:10:24,326 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35559 2024-12-02T01:10:24,327 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35559 2024-12-02T01:10:24,330 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35559 2024-12-02T01:10:24,331 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/a82f277fe18c,39899,1733101824140 2024-12-02T01:10:24,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:10:24,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:10:24,341 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a82f277fe18c,39899,1733101824140 2024-12-02T01:10:24,347 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a82f277fe18c:39899 2024-12-02T01:10:24,349 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:10:24,349 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:10:24,349 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,349 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,349 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:10:24,350 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a82f277fe18c,39899,1733101824140 from backup master directory 2024-12-02T01:10:24,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:10:24,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a82f277fe18c,39899,1733101824140 2024-12-02T01:10:24,357 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:10:24,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:10:24,357 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:10:24,357 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a82f277fe18c,39899,1733101824140 2024-12-02T01:10:24,366 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:24,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:10:24,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:10:24,373 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/hbase.id with ID: 06520d21-b675-40f6-84db-f3d8019e072b 2024-12-02T01:10:24,386 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:24,397 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,397 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,412 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:10:24,413 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:10:24,414 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:10:24,415 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T01:10:24,415 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:10:24,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:10:24,432 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:10:24,433 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store 2024-12-02T01:10:24,439 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:10:24,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:10:24,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:24,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:10:24,440 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:24,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:24,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:10:24,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:24,440 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:10:24,440 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:10:24,441 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/.initializing 2024-12-02T01:10:24,441 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/WALs/a82f277fe18c,39899,1733101824140 2024-12-02T01:10:24,444 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C39899%2C1733101824140, suffix=, logDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/WALs/a82f277fe18c,39899,1733101824140, archiveDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/oldWALs, maxLogs=10 2024-12-02T01:10:24,445 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C39899%2C1733101824140.1733101824444 2024-12-02T01:10:24,454 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/WALs/a82f277fe18c,39899,1733101824140/a82f277fe18c%2C39899%2C1733101824140.1733101824444 2024-12-02T01:10:24,454 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39881:39881),(127.0.0.1/127.0.0.1:44943:44943)] 2024-12-02T01:10:24,454 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:10:24,455 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:24,455 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,455 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,459 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,461 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T01:10:24,461 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:24,461 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:24,461 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,463 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T01:10:24,463 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:24,464 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:10:24,464 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,465 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T01:10:24,465 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:24,466 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:10:24,466 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,468 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T01:10:24,468 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:24,469 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:10:24,469 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,470 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,472 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T01:10:24,473 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:10:24,475 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:10:24,476 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=735208, jitterRate=-0.06513555347919464}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T01:10:24,476 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:10:24,476 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T01:10:24,479 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7eb098b7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:10:24,480 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-02T01:10:24,480 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T01:10:24,480 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T01:10:24,480 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T01:10:24,481 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T01:10:24,481 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-02T01:10:24,481 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T01:10:24,483 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T01:10:24,484 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T01:10:24,499 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-02T01:10:24,499 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T01:10:24,500 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T01:10:24,542 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-02T01:10:24,542 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T01:10:24,543 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T01:10:24,592 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-02T01:10:24,593 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T01:10:24,634 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T01:10:24,636 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T01:10:24,676 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T01:10:24,691 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:10:24,691 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:10:24,691 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,691 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,692 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=a82f277fe18c,39899,1733101824140, sessionid=0x10194e0caf20000, setting cluster-up flag (Was=false) 2024-12-02T01:10:24,751 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,751 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,807 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T01:10:24,808 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,39899,1733101824140 2024-12-02T01:10:24,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:24,931 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:25,049 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T01:10:25,051 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,39899,1733101824140 2024-12-02T01:10:25,053 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-02T01:10:25,053 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-02T01:10:25,054 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a82f277fe18c,39899,1733101824140 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a82f277fe18c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T01:10:25,054 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,055 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:10:25,055 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,056 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:10:25,057 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-02T01:10:25,058 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,058 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733101855059 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T01:10:25,059 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,060 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T01:10:25,060 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T01:10:25,060 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T01:10:25,060 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T01:10:25,060 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T01:10:25,062 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101825060,5,FailOnTimeoutGroup] 2024-12-02T01:10:25,063 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101825062,5,FailOnTimeoutGroup] 2024-12-02T01:10:25,063 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,063 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T01:10:25,063 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,063 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,068 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:10:25,068 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:10:25,069 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-02T01:10:25,069 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9 2024-12-02T01:10:25,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:10:25,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:10:25,080 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:25,083 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:10:25,084 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:10:25,084 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,084 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:25,085 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:10:25,086 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:10:25,086 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,086 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:25,086 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:10:25,088 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:10:25,088 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,088 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:25,089 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740 2024-12-02T01:10:25,090 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740 2024-12-02T01:10:25,091 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:10:25,092 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:10:25,094 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:10:25,095 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=789404, jitterRate=0.003779307007789612}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:10:25,095 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:10:25,095 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:10:25,095 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:10:25,095 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:10:25,095 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:10:25,095 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:10:25,096 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:10:25,096 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:10:25,097 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:10:25,097 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-02T01:10:25,097 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T01:10:25,098 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T01:10:25,100 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T01:10:25,146 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a82f277fe18c:35559 2024-12-02T01:10:25,147 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1008): ClusterId : 06520d21-b675-40f6-84db-f3d8019e072b 2024-12-02T01:10:25,147 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:10:25,155 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:10:25,155 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:10:25,166 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:10:25,167 DEBUG [RS:0;a82f277fe18c:35559 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5e63c754, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:10:25,167 DEBUG [RS:0;a82f277fe18c:35559 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3355c7c6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:10:25,167 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:10:25,167 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:10:25,167 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:10:25,168 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,39899,1733101824140 with isa=a82f277fe18c/172.17.0.2:35559, startcode=1733101824300 2024-12-02T01:10:25,168 DEBUG [RS:0;a82f277fe18c:35559 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:10:25,170 INFO [RS-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34177, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:10:25,170 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39899 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,170 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39899 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,172 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9 2024-12-02T01:10:25,172 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:43627 2024-12-02T01:10:25,172 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:10:25,182 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:10:25,183 DEBUG [RS:0;a82f277fe18c:35559 {}] zookeeper.ZKUtil(111): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,183 WARN [RS:0;a82f277fe18c:35559 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:10:25,183 INFO [RS:0;a82f277fe18c:35559 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:10:25,183 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,184 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,35559,1733101824300] 2024-12-02T01:10:25,186 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:10:25,187 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:10:25,189 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:10:25,191 INFO [RS:0;a82f277fe18c:35559 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:10:25,191 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,194 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:10:25,195 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:10:25,195 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,196 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,196 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,196 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,196 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:10:25,196 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:10:25,196 DEBUG [RS:0;a82f277fe18c:35559 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:10:25,198 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,198 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,198 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,198 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,198 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35559,1733101824300-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:10:25,213 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:10:25,213 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35559,1733101824300-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,227 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.Replication(204): a82f277fe18c,35559,1733101824300 started 2024-12-02T01:10:25,227 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,35559,1733101824300, RpcServer on a82f277fe18c/172.17.0.2:35559, sessionid=0x10194e0caf20001 2024-12-02T01:10:25,227 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:10:25,227 DEBUG [RS:0;a82f277fe18c:35559 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,227 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,35559,1733101824300' 2024-12-02T01:10:25,227 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,35559,1733101824300' 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:10:25,228 DEBUG [RS:0;a82f277fe18c:35559 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:10:25,228 INFO [RS:0;a82f277fe18c:35559 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:10:25,228 INFO [RS:0;a82f277fe18c:35559 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:10:25,250 WARN [a82f277fe18c:39899 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-02T01:10:25,330 INFO [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C35559%2C1733101824300, suffix=, logDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300, archiveDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs, maxLogs=32 2024-12-02T01:10:25,331 INFO [RS:0;a82f277fe18c:35559 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35559%2C1733101824300.1733101825331 2024-12-02T01:10:25,343 INFO [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101825331 2024-12-02T01:10:25,343 DEBUG [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44943:44943),(127.0.0.1/127.0.0.1:39881:39881)] 2024-12-02T01:10:25,367 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:25,500 DEBUG [a82f277fe18c:39899 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T01:10:25,501 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,502 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,35559,1733101824300, state=OPENING 2024-12-02T01:10:25,507 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T01:10:25,516 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:25,516 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:25,516 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:10:25,516 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:10:25,517 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=a82f277fe18c,35559,1733101824300}] 2024-12-02T01:10:25,670 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,670 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T01:10:25,672 INFO [RS-EventLoopGroup-11-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33446, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T01:10:25,675 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-02T01:10:25,675 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:10:25,677 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C35559%2C1733101824300.meta, suffix=.meta, logDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300, archiveDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs, maxLogs=32 2024-12-02T01:10:25,678 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35559%2C1733101824300.meta.1733101825678.meta 2024-12-02T01:10:25,684 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.meta.1733101825678.meta 2024-12-02T01:10:25,684 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39881:39881),(127.0.0.1/127.0.0.1:44943:44943)] 2024-12-02T01:10:25,684 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:10:25,684 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T01:10:25,684 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T01:10:25,685 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T01:10:25,685 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T01:10:25,685 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:25,685 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-02T01:10:25,685 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-02T01:10:25,687 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:10:25,688 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:10:25,688 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,689 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:25,689 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:10:25,690 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:10:25,690 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,690 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:25,690 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:10:25,691 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:10:25,692 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,692 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:10:25,693 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740 2024-12-02T01:10:25,695 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740 2024-12-02T01:10:25,698 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:10:25,700 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:10:25,701 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=781941, jitterRate=-0.005711272358894348}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:10:25,701 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:10:25,702 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733101825670 2024-12-02T01:10:25,704 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T01:10:25,705 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-02T01:10:25,705 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,706 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,35559,1733101824300, state=OPEN 2024-12-02T01:10:25,746 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:10:25,746 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:10:25,746 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:10:25,746 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:10:25,752 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T01:10:25,752 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=a82f277fe18c,35559,1733101824300 in 229 msec 2024-12-02T01:10:25,757 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T01:10:25,757 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 655 msec 2024-12-02T01:10:25,761 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 707 msec 2024-12-02T01:10:25,761 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733101825761, completionTime=-1 2024-12-02T01:10:25,761 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T01:10:25,761 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-02T01:10:25,762 DEBUG [hconnection-0x6b23704-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:10:25,763 INFO [RS-EventLoopGroup-11-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33448, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:10:25,765 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-02T01:10:25,765 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733101885765 2024-12-02T01:10:25,765 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733101945765 2024-12-02T01:10:25,765 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 3 msec 2024-12-02T01:10:25,791 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39899,1733101824140-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,791 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39899,1733101824140-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,791 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39899,1733101824140-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,791 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a82f277fe18c:39899, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,791 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:25,791 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-02T01:10:25,792 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:10:25,793 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-02T01:10:25,793 DEBUG [master/a82f277fe18c:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-02T01:10:25,794 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:10:25,794 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:25,796 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:10:25,804 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:10:25,805 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:10:25,806 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 855a3378d355b0c131db2b8b340ff7e9, NAME => 'hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9 2024-12-02T01:10:25,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:10:25,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:10:25,813 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:25,813 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 855a3378d355b0c131db2b8b340ff7e9, disabling compactions & flushes 2024-12-02T01:10:25,813 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:25,813 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:25,813 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. after waiting 0 ms 2024-12-02T01:10:25,813 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:25,813 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:25,813 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 855a3378d355b0c131db2b8b340ff7e9: 2024-12-02T01:10:25,814 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:10:25,814 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733101825814"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101825814"}]},"ts":"1733101825814"} 2024-12-02T01:10:25,816 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:10:25,817 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:10:25,817 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101825817"}]},"ts":"1733101825817"} 2024-12-02T01:10:25,819 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-02T01:10:25,838 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=855a3378d355b0c131db2b8b340ff7e9, ASSIGN}] 2024-12-02T01:10:25,839 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=855a3378d355b0c131db2b8b340ff7e9, ASSIGN 2024-12-02T01:10:25,840 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=855a3378d355b0c131db2b8b340ff7e9, ASSIGN; state=OFFLINE, location=a82f277fe18c,35559,1733101824300; forceNewPlan=false, retain=false 2024-12-02T01:10:25,991 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=855a3378d355b0c131db2b8b340ff7e9, regionState=OPENING, regionLocation=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:25,993 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 855a3378d355b0c131db2b8b340ff7e9, server=a82f277fe18c,35559,1733101824300}] 2024-12-02T01:10:26,146 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:10:26,149 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:26,150 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 855a3378d355b0c131db2b8b340ff7e9, NAME => 'hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:10:26,150 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,150 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:26,150 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,150 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,152 INFO [StoreOpener-855a3378d355b0c131db2b8b340ff7e9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,153 INFO [StoreOpener-855a3378d355b0c131db2b8b340ff7e9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 855a3378d355b0c131db2b8b340ff7e9 columnFamilyName info 2024-12-02T01:10:26,153 DEBUG [StoreOpener-855a3378d355b0c131db2b8b340ff7e9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:26,154 INFO [StoreOpener-855a3378d355b0c131db2b8b340ff7e9-1 {}] regionserver.HStore(327): Store=855a3378d355b0c131db2b8b340ff7e9/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:10:26,154 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,155 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,157 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:10:26,159 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:10:26,159 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 855a3378d355b0c131db2b8b340ff7e9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=839110, jitterRate=0.06698362529277802}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:10:26,160 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 855a3378d355b0c131db2b8b340ff7e9: 2024-12-02T01:10:26,161 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9., pid=6, masterSystemTime=1733101826145 2024-12-02T01:10:26,163 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:26,163 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:26,163 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=855a3378d355b0c131db2b8b340ff7e9, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:26,167 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T01:10:26,168 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 855a3378d355b0c131db2b8b340ff7e9, server=a82f277fe18c,35559,1733101824300 in 172 msec 2024-12-02T01:10:26,170 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T01:10:26,170 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=855a3378d355b0c131db2b8b340ff7e9, ASSIGN in 329 msec 2024-12-02T01:10:26,170 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:10:26,170 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101826170"}]},"ts":"1733101826170"} 2024-12-02T01:10:26,172 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-02T01:10:26,183 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:10:26,185 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 391 msec 2024-12-02T01:10:26,194 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-02T01:10:26,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:26,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:10:26,205 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:10:26,209 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-02T01:10:26,224 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:10:26,234 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 25 msec 2024-12-02T01:10:26,241 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-02T01:10:26,257 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:10:26,268 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 27 msec 2024-12-02T01:10:26,291 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-02T01:10:26,307 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-02T01:10:26,307 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.950sec 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39899,1733101824140-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:10:26,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39899,1733101824140-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T01:10:26,309 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-02T01:10:26,310 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T01:10:26,310 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39899,1733101824140-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:10:26,334 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x74d3406a to 127.0.0.1:51080 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@55294bed 2024-12-02T01:10:26,341 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1dca1277, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:10:26,343 DEBUG [hconnection-0x4636debe-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:10:26,345 INFO [RS-EventLoopGroup-11-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:33460, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:10:26,346 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=a82f277fe18c,39899,1733101824140 2024-12-02T01:10:26,346 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:10:26,348 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-02T01:10:26,349 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T01:10:26,351 INFO [RS-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:55648, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T01:10:26,352 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T01:10:26,352 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T01:10:26,352 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testCompactionRecordDoesntBlockRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:10:26,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:26,354 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:10:26,355 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:26,355 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testCompactionRecordDoesntBlockRolling" procId is: 9 2024-12-02T01:10:26,355 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:10:26,355 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:10:26,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741837_1013 (size=405) 2024-12-02T01:10:26,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741837_1013 (size=405) 2024-12-02T01:10:26,364 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => fcce99fb7fa2556a6a6938e11c30f9b5, NAME => 'TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testCompactionRecordDoesntBlockRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9 2024-12-02T01:10:26,367 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:26,373 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741838_1014 (size=88) 2024-12-02T01:10:26,373 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741838_1014 (size=88) 2024-12-02T01:10:26,374 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:26,374 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1681): Closing fcce99fb7fa2556a6a6938e11c30f9b5, disabling compactions & flushes 2024-12-02T01:10:26,374 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,374 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,374 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. after waiting 0 ms 2024-12-02T01:10:26,374 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,374 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,374 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1635): Region close journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:10:26,375 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:10:26,376 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.","families":{"info":[{"qualifier":"regioninfo","vlen":87,"tag":[],"timestamp":"1733101826376"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101826376"}]},"ts":"1733101826376"} 2024-12-02T01:10:26,377 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:10:26,378 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:10:26,378 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101826378"}]},"ts":"1733101826378"} 2024-12-02T01:10:26,380 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testCompactionRecordDoesntBlockRolling, state=ENABLING in hbase:meta 2024-12-02T01:10:26,399 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=fcce99fb7fa2556a6a6938e11c30f9b5, ASSIGN}] 2024-12-02T01:10:26,401 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=fcce99fb7fa2556a6a6938e11c30f9b5, ASSIGN 2024-12-02T01:10:26,401 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=fcce99fb7fa2556a6a6938e11c30f9b5, ASSIGN; state=OFFLINE, location=a82f277fe18c,35559,1733101824300; forceNewPlan=false, retain=false 2024-12-02T01:10:26,552 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=fcce99fb7fa2556a6a6938e11c30f9b5, regionState=OPENING, regionLocation=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:26,554 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure fcce99fb7fa2556a6a6938e11c30f9b5, server=a82f277fe18c,35559,1733101824300}] 2024-12-02T01:10:26,707 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:10:26,711 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,711 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => fcce99fb7fa2556a6a6938e11c30f9b5, NAME => 'TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:10:26,712 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testCompactionRecordDoesntBlockRolling fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,712 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:10:26,712 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,712 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,714 INFO [StoreOpener-fcce99fb7fa2556a6a6938e11c30f9b5-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,716 INFO [StoreOpener-fcce99fb7fa2556a6a6938e11c30f9b5-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region fcce99fb7fa2556a6a6938e11c30f9b5 columnFamilyName info 2024-12-02T01:10:26,716 DEBUG [StoreOpener-fcce99fb7fa2556a6a6938e11c30f9b5-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:10:26,717 INFO [StoreOpener-fcce99fb7fa2556a6a6938e11c30f9b5-1 {}] regionserver.HStore(327): Store=fcce99fb7fa2556a6a6938e11c30f9b5/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:10:26,718 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,719 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,722 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:10:26,724 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:10:26,725 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened fcce99fb7fa2556a6a6938e11c30f9b5; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=818296, jitterRate=0.040518417954444885}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:10:26,725 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:10:26,726 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5., pid=11, masterSystemTime=1733101826706 2024-12-02T01:10:26,728 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,728 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:26,729 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=fcce99fb7fa2556a6a6938e11c30f9b5, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,35559,1733101824300 2024-12-02T01:10:26,732 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-02T01:10:26,732 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure fcce99fb7fa2556a6a6938e11c30f9b5, server=a82f277fe18c,35559,1733101824300 in 177 msec 2024-12-02T01:10:26,734 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-02T01:10:26,734 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=fcce99fb7fa2556a6a6938e11c30f9b5, ASSIGN in 333 msec 2024-12-02T01:10:26,735 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:10:26,735 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101826735"}]},"ts":"1733101826735"} 2024-12-02T01:10:26,737 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testCompactionRecordDoesntBlockRolling, state=ENABLED in hbase:meta 2024-12-02T01:10:26,747 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:10:26,749 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 394 msec 2024-12-02T01:10:27,368 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:28,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:10:28,353 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2024-12-02T01:10:28,355 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:28,355 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling Metrics about Tables on a single HBase RegionServer 2024-12-02T01:10:28,369 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:29,371 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:30,371 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:30,372 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=2 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta after 68067ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor203.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:10:31,188 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:10:31,190 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,190 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,190 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,191 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,209 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,210 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,210 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,210 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,210 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,211 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,213 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,214 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,214 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,215 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:10:31,220 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T01:10:31,220 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-02T01:10:31,221 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testCompactionRecordDoesntBlockRolling' 2024-12-02T01:10:31,373 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:32,373 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:33,374 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:34,374 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:35,375 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:36,357 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:10:36,357 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 9 completed 2024-12-02T01:10:36,359 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:36,359 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:36,367 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush hbase:namespace 2024-12-02T01:10:36,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace 2024-12-02T01:10:36,373 INFO [PEWorker-4 {}] procedure.FlushTableProcedure(91): pid=12, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace execute state=FLUSH_TABLE_PREPARE 2024-12-02T01:10:36,374 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2024-12-02T01:10:36,374 INFO [PEWorker-4 {}] procedure.FlushTableProcedure(91): pid=12, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T01:10:36,376 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T01:10:36,376 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:36,533 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:10:36,534 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=35559 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=13 2024-12-02T01:10:36,534 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.FlushRegionCallable(51): Starting region operation on hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:36,535 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegion(2837): Flushing 855a3378d355b0c131db2b8b340ff7e9 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-02T01:10:36,551 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/.tmp/info/397221814eb747e38391dcda794a60f6 is 45, key is default/info:d/1733101826214/Put/seqid=0 2024-12-02T01:10:36,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741839_1015 (size=5037) 2024-12-02T01:10:36,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741839_1015 (size=5037) 2024-12-02T01:10:36,556 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/.tmp/info/397221814eb747e38391dcda794a60f6 2024-12-02T01:10:36,562 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/.tmp/info/397221814eb747e38391dcda794a60f6 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/info/397221814eb747e38391dcda794a60f6 2024-12-02T01:10:36,568 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/info/397221814eb747e38391dcda794a60f6, entries=2, sequenceid=6, filesize=4.9 K 2024-12-02T01:10:36,569 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 855a3378d355b0c131db2b8b340ff7e9 in 34ms, sequenceid=6, compaction requested=false 2024-12-02T01:10:36,570 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegion(2538): Flush status journal for 855a3378d355b0c131db2b8b340ff7e9: 2024-12-02T01:10:36,570 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.FlushRegionCallable(64): Closing region operation on hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:10:36,571 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=13 2024-12-02T01:10:36,572 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster(4106): Remote procedure done, pid=13 2024-12-02T01:10:36,575 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=13, resume processing ppid=12 2024-12-02T01:10:36,576 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 199 msec 2024-12-02T01:10:36,577 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace in 208 msec 2024-12-02T01:10:37,377 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:38,377 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:39,378 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:40,379 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:41,380 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:42,381 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:43,382 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:44,383 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:45,384 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:46,375 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2024-12-02T01:10:46,375 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: hbase:namespace, procId: 12 completed 2024-12-02T01:10:46,384 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:46,385 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:46,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] procedure2.ProcedureExecutor(1098): Stored pid=14, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:46,387 INFO [PEWorker-3 {}] procedure.FlushTableProcedure(91): pid=14, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T01:10:46,387 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=14 2024-12-02T01:10:46,387 INFO [PEWorker-3 {}] procedure.FlushTableProcedure(91): pid=14, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T01:10:46,388 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=14, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T01:10:46,543 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:10:46,544 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=35559 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=15 2024-12-02T01:10:46,545 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:46,545 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegion(2837): Flushing fcce99fb7fa2556a6a6938e11c30f9b5 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T01:10:46,565 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/86779663b67943ec934547ef33f74542 is 1080, key is row0001/info:/1733101846379/Put/seqid=0 2024-12-02T01:10:46,569 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741840_1016 (size=6033) 2024-12-02T01:10:46,570 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741840_1016 (size=6033) 2024-12-02T01:10:46,570 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/86779663b67943ec934547ef33f74542 2024-12-02T01:10:46,577 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/86779663b67943ec934547ef33f74542 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/86779663b67943ec934547ef33f74542 2024-12-02T01:10:46,582 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/86779663b67943ec934547ef33f74542, entries=1, sequenceid=5, filesize=5.9 K 2024-12-02T01:10:46,583 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for fcce99fb7fa2556a6a6938e11c30f9b5 in 38ms, sequenceid=5, compaction requested=false 2024-12-02T01:10:46,583 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegion(2538): Flush status journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:10:46,583 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:46,583 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=15 2024-12-02T01:10:46,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster(4106): Remote procedure done, pid=15 2024-12-02T01:10:46,586 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=15, resume processing ppid=14 2024-12-02T01:10:46,586 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=14, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 196 msec 2024-12-02T01:10:46,587 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 202 msec 2024-12-02T01:10:47,385 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:48,386 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:49,387 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:50,055 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 61b9a0fcb731850a98b7b912948ae7ef, had cached 0 bytes from a total of 23930 2024-12-02T01:10:50,388 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:51,389 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:52,390 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:53,390 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:54,122 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:10:54,391 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:55,391 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:56,388 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=14 2024-12-02T01:10:56,388 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 14 completed 2024-12-02T01:10:56,392 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:56,393 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:56,394 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] procedure2.ProcedureExecutor(1098): Stored pid=16, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:10:56,395 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=16 2024-12-02T01:10:56,395 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=16, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T01:10:56,396 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=16, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T01:10:56,396 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=17, ppid=16, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T01:10:56,548 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:10:56,549 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=35559 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=17 2024-12-02T01:10:56,550 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:56,550 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegion(2837): Flushing fcce99fb7fa2556a6a6938e11c30f9b5 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T01:10:56,556 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/c290cd5bb5ab4422a9d6ac35b72a10d6 is 1080, key is row0002/info:/1733101856389/Put/seqid=0 2024-12-02T01:10:56,564 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741841_1017 (size=6033) 2024-12-02T01:10:56,564 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741841_1017 (size=6033) 2024-12-02T01:10:56,565 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/c290cd5bb5ab4422a9d6ac35b72a10d6 2024-12-02T01:10:56,570 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/c290cd5bb5ab4422a9d6ac35b72a10d6 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/c290cd5bb5ab4422a9d6ac35b72a10d6 2024-12-02T01:10:56,576 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/c290cd5bb5ab4422a9d6ac35b72a10d6, entries=1, sequenceid=9, filesize=5.9 K 2024-12-02T01:10:56,577 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for fcce99fb7fa2556a6a6938e11c30f9b5 in 27ms, sequenceid=9, compaction requested=false 2024-12-02T01:10:56,577 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegion(2538): Flush status journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:10:56,577 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:10:56,577 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=17 2024-12-02T01:10:56,578 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster(4106): Remote procedure done, pid=17 2024-12-02T01:10:56,581 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=17, resume processing ppid=16 2024-12-02T01:10:56,581 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, ppid=16, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 183 msec 2024-12-02T01:10:56,583 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 188 msec 2024-12-02T01:10:57,393 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:58,393 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:10:59,167 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:10:59,168 INFO [RS-EventLoopGroup-10-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49580, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:10:59,394 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:00,395 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:01,396 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:02,397 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:03,398 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:04,399 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:05,400 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:06,351 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T01:11:06,351 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T01:11:06,396 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=16 2024-12-02T01:11:06,397 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 16 completed 2024-12-02T01:11:06,400 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35559%2C1733101824300.1733101866400 2024-12-02T01:11:06,401 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:06,411 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101825331 with entries=13, filesize=6.41 KB; new WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101866400 2024-12-02T01:11:06,412 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39881:39881),(127.0.0.1/127.0.0.1:44943:44943)] 2024-12-02T01:11:06,412 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101825331 is not closed yet, will try archiving it next time 2024-12-02T01:11:06,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741833_1009 (size=6574) 2024-12-02T01:11:06,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741833_1009 (size=6574) 2024-12-02T01:11:06,417 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:11:06,418 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] procedure2.ProcedureExecutor(1098): Stored pid=18, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:11:06,419 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=18, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T01:11:06,419 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=18 2024-12-02T01:11:06,420 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=18, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T01:11:06,420 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=19, ppid=18, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T01:11:06,572 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:11:06,574 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=35559 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=19 2024-12-02T01:11:06,574 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:06,574 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegion(2837): Flushing fcce99fb7fa2556a6a6938e11c30f9b5 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T01:11:06,585 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/801a628fcc184e92a0731a04eb3fe8fd is 1080, key is row0003/info:/1733101866398/Put/seqid=0 2024-12-02T01:11:06,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741843_1019 (size=6033) 2024-12-02T01:11:06,592 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741843_1019 (size=6033) 2024-12-02T01:11:06,593 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=13 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/801a628fcc184e92a0731a04eb3fe8fd 2024-12-02T01:11:06,600 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/801a628fcc184e92a0731a04eb3fe8fd as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/801a628fcc184e92a0731a04eb3fe8fd 2024-12-02T01:11:06,608 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/801a628fcc184e92a0731a04eb3fe8fd, entries=1, sequenceid=13, filesize=5.9 K 2024-12-02T01:11:06,609 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for fcce99fb7fa2556a6a6938e11c30f9b5 in 35ms, sequenceid=13, compaction requested=true 2024-12-02T01:11:06,609 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegion(2538): Flush status journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:11:06,609 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:06,609 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=19 2024-12-02T01:11:06,610 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster(4106): Remote procedure done, pid=19 2024-12-02T01:11:06,612 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=19, resume processing ppid=18 2024-12-02T01:11:06,612 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=19, ppid=18, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 191 msec 2024-12-02T01:11:06,613 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 196 msec 2024-12-02T01:11:07,402 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:08,403 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:09,404 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:10,405 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:11,150 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 855a3378d355b0c131db2b8b340ff7e9, had cached 0 bytes from a total of 5037 2024-12-02T01:11:11,406 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:11,712 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region fcce99fb7fa2556a6a6938e11c30f9b5, had cached 0 bytes from a total of 18099 2024-12-02T01:11:12,406 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:13,407 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:14,408 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:15,408 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:16,409 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:16,421 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=18 2024-12-02T01:11:16,422 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 18 completed 2024-12-02T01:11:16,422 DEBUG [Time-limited test {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:11:16,423 DEBUG [Time-limited test {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 18099 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:11:16,423 DEBUG [Time-limited test {}] regionserver.HStore(1540): fcce99fb7fa2556a6a6938e11c30f9b5/info is initiating minor compaction (all files) 2024-12-02T01:11:16,423 INFO [Time-limited test {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:11:16,423 INFO [Time-limited test {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:16,423 INFO [Time-limited test {}] regionserver.HRegion(2351): Starting compaction of fcce99fb7fa2556a6a6938e11c30f9b5/info in TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:16,423 INFO [Time-limited test {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/86779663b67943ec934547ef33f74542, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/c290cd5bb5ab4422a9d6ac35b72a10d6, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/801a628fcc184e92a0731a04eb3fe8fd] into tmpdir=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp, totalSize=17.7 K 2024-12-02T01:11:16,424 DEBUG [Time-limited test {}] compactions.Compactor(224): Compacting 86779663b67943ec934547ef33f74542, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=5, earliestPutTs=1733101846379 2024-12-02T01:11:16,424 DEBUG [Time-limited test {}] compactions.Compactor(224): Compacting c290cd5bb5ab4422a9d6ac35b72a10d6, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=9, earliestPutTs=1733101856389 2024-12-02T01:11:16,424 DEBUG [Time-limited test {}] compactions.Compactor(224): Compacting 801a628fcc184e92a0731a04eb3fe8fd, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=13, earliestPutTs=1733101866398 2024-12-02T01:11:16,435 INFO [Time-limited test {}] throttle.PressureAwareThroughputController(145): fcce99fb7fa2556a6a6938e11c30f9b5#info#compaction#32 average throughput is unlimited, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:11:16,436 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/1d01d8d29d2b46debcf69d08cd301aa4 is 1080, key is row0001/info:/1733101846379/Put/seqid=0 2024-12-02T01:11:16,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741844_1020 (size=8296) 2024-12-02T01:11:16,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741844_1020 (size=8296) 2024-12-02T01:11:16,448 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/1d01d8d29d2b46debcf69d08cd301aa4 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/1d01d8d29d2b46debcf69d08cd301aa4 2024-12-02T01:11:16,454 INFO [Time-limited test {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in fcce99fb7fa2556a6a6938e11c30f9b5/info of fcce99fb7fa2556a6a6938e11c30f9b5 into 1d01d8d29d2b46debcf69d08cd301aa4(size=8.1 K), total size for store is 8.1 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:11:16,454 DEBUG [Time-limited test {}] regionserver.HRegion(2381): Compaction status journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:11:16,456 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35559%2C1733101824300.1733101876456 2024-12-02T01:11:16,465 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101866400 with entries=4, filesize=2.45 KB; new WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101876456 2024-12-02T01:11:16,465 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44943:44943),(127.0.0.1/127.0.0.1:39881:39881)] 2024-12-02T01:11:16,465 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101866400 is not closed yet, will try archiving it next time 2024-12-02T01:11:16,465 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101825331 to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs/a82f277fe18c%2C35559%2C1733101824300.1733101825331 2024-12-02T01:11:16,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741842_1018 (size=2520) 2024-12-02T01:11:16,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741842_1018 (size=2520) 2024-12-02T01:11:16,468 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:11:16,469 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] procedure2.ProcedureExecutor(1098): Stored pid=20, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:11:16,470 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2024-12-02T01:11:16,470 INFO [PEWorker-5 {}] procedure.FlushTableProcedure(91): pid=20, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-02T01:11:16,471 INFO [PEWorker-5 {}] procedure.FlushTableProcedure(91): pid=20, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-02T01:11:16,471 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=21, ppid=20, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-02T01:11:16,623 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,35559,1733101824300 2024-12-02T01:11:16,624 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=35559 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=21 2024-12-02T01:11:16,624 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:16,624 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegion(2837): Flushing fcce99fb7fa2556a6a6938e11c30f9b5 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T01:11:16,628 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/dcbea4c3f6504836a284a495bfda12b8 is 1080, key is row0000/info:/1733101876455/Put/seqid=0 2024-12-02T01:11:16,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741846_1022 (size=6033) 2024-12-02T01:11:16,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741846_1022 (size=6033) 2024-12-02T01:11:16,638 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=18 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/dcbea4c3f6504836a284a495bfda12b8 2024-12-02T01:11:16,646 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/dcbea4c3f6504836a284a495bfda12b8 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/dcbea4c3f6504836a284a495bfda12b8 2024-12-02T01:11:16,653 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/dcbea4c3f6504836a284a495bfda12b8, entries=1, sequenceid=18, filesize=5.9 K 2024-12-02T01:11:16,654 INFO [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for fcce99fb7fa2556a6a6938e11c30f9b5 in 30ms, sequenceid=18, compaction requested=false 2024-12-02T01:11:16,654 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegion(2538): Flush status journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:11:16,654 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:16,654 DEBUG [RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=21 2024-12-02T01:11:16,654 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.HMaster(4106): Remote procedure done, pid=21 2024-12-02T01:11:16,657 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=21, resume processing ppid=20 2024-12-02T01:11:16,657 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=21, ppid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 184 msec 2024-12-02T01:11:16,659 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 188 msec 2024-12-02T01:11:17,410 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:18,412 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:19,413 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:20,414 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:21,415 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:22,416 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:23,417 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:24,123 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:11:24,418 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:25,418 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:26,419 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:26,452 DEBUG [master/a82f277fe18c:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 855a3378d355b0c131db2b8b340ff7e9 changed from -1.0 to 0.0, refreshing cache 2024-12-02T01:11:26,471 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39899 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2024-12-02T01:11:26,471 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 20 completed 2024-12-02T01:11:26,473 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35559%2C1733101824300.1733101886473 2024-12-02T01:11:26,481 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101876456 with entries=3, filesize=1.97 KB; new WAL /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101886473 2024-12-02T01:11:26,481 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:44943:44943),(127.0.0.1/127.0.0.1:39881:39881)] 2024-12-02T01:11:26,481 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101876456 is not closed yet, will try archiving it next time 2024-12-02T01:11:26,481 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101866400 to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs/a82f277fe18c%2C35559%2C1733101824300.1733101866400 2024-12-02T01:11:26,481 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-02T01:11:26,481 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-02T01:11:26,481 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x74d3406a to 127.0.0.1:51080 2024-12-02T01:11:26,481 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:11:26,482 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T01:11:26,482 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=26426551, stopped=false 2024-12-02T01:11:26,482 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=a82f277fe18c,39899,1733101824140 2024-12-02T01:11:26,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741845_1021 (size=2026) 2024-12-02T01:11:26,483 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741845_1021 (size=2026) 2024-12-02T01:11:26,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:11:26,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:11:26,491 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-02T01:11:26,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:26,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:26,491 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:11:26,491 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,35559,1733101824300' ***** 2024-12-02T01:11:26,491 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:11:26,491 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:11:26,491 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:11:26,491 INFO [RS:0;a82f277fe18c:35559 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:11:26,491 INFO [RS:0;a82f277fe18c:35559 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(3579): Received CLOSE for fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:11:26,492 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:11:26,492 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(3579): Received CLOSE for 855a3378d355b0c131db2b8b340ff7e9 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,35559,1733101824300 2024-12-02T01:11:26,492 DEBUG [RS:0;a82f277fe18c:35559 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing fcce99fb7fa2556a6a6938e11c30f9b5, disabling compactions & flushes 2024-12-02T01:11:26,492 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. after waiting 0 ms 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:11:26,492 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing fcce99fb7fa2556a6a6938e11c30f9b5 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-02T01:11:26,492 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-02T01:11:26,492 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1603): Online Regions={fcce99fb7fa2556a6a6938e11c30f9b5=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5., 1588230740=hbase:meta,,1.1588230740, 855a3378d355b0c131db2b8b340ff7e9=hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9.} 2024-12-02T01:11:26,492 DEBUG [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 855a3378d355b0c131db2b8b340ff7e9, fcce99fb7fa2556a6a6938e11c30f9b5 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:11:26,492 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:11:26,492 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:11:26,493 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=3.05 KB heapSize=5.55 KB 2024-12-02T01:11:26,496 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/ecc3da0b1b084c4d98795503491f0d5a is 1080, key is row0001/info:/1733101886472/Put/seqid=0 2024-12-02T01:11:26,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741848_1024 (size=6033) 2024-12-02T01:11:26,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741848_1024 (size=6033) 2024-12-02T01:11:26,501 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=22 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/ecc3da0b1b084c4d98795503491f0d5a 2024-12-02T01:11:26,506 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/.tmp/info/ecc3da0b1b084c4d98795503491f0d5a as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/ecc3da0b1b084c4d98795503491f0d5a 2024-12-02T01:11:26,507 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/.tmp/info/711e802cb34444508451868ba942640d is 227, key is TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5./info:regioninfo/1733101826729/Put/seqid=0 2024-12-02T01:11:26,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741849_1025 (size=8430) 2024-12-02T01:11:26,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741849_1025 (size=8430) 2024-12-02T01:11:26,512 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/ecc3da0b1b084c4d98795503491f0d5a, entries=1, sequenceid=22, filesize=5.9 K 2024-12-02T01:11:26,512 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.79 KB at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/.tmp/info/711e802cb34444508451868ba942640d 2024-12-02T01:11:26,513 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for fcce99fb7fa2556a6a6938e11c30f9b5 in 21ms, sequenceid=22, compaction requested=true 2024-12-02T01:11:26,514 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/86779663b67943ec934547ef33f74542, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/c290cd5bb5ab4422a9d6ac35b72a10d6, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/801a628fcc184e92a0731a04eb3fe8fd] to archive 2024-12-02T01:11:26,514 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T01:11:26,516 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/86779663b67943ec934547ef33f74542 to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/86779663b67943ec934547ef33f74542 2024-12-02T01:11:26,517 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/c290cd5bb5ab4422a9d6ac35b72a10d6 to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/c290cd5bb5ab4422a9d6ac35b72a10d6 2024-12-02T01:11:26,519 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/801a628fcc184e92a0731a04eb3fe8fd to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/info/801a628fcc184e92a0731a04eb3fe8fd 2024-12-02T01:11:26,524 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/fcce99fb7fa2556a6a6938e11c30f9b5/recovered.edits/25.seqid, newMaxSeqId=25, maxSeqId=1 2024-12-02T01:11:26,524 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:26,524 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for fcce99fb7fa2556a6a6938e11c30f9b5: 2024-12-02T01:11:26,524 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733101826352.fcce99fb7fa2556a6a6938e11c30f9b5. 2024-12-02T01:11:26,524 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 855a3378d355b0c131db2b8b340ff7e9, disabling compactions & flushes 2024-12-02T01:11:26,524 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:11:26,525 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:11:26,525 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. after waiting 0 ms 2024-12-02T01:11:26,525 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:11:26,529 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/namespace/855a3378d355b0c131db2b8b340ff7e9/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T01:11:26,529 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:11:26,529 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 855a3378d355b0c131db2b8b340ff7e9: 2024-12-02T01:11:26,529 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733101825791.855a3378d355b0c131db2b8b340ff7e9. 2024-12-02T01:11:26,539 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/.tmp/table/c4f814ac17624b5d99627371fd082b5e is 89, key is TestLogRolling-testCompactionRecordDoesntBlockRolling/table:state/1733101826735/Put/seqid=0 2024-12-02T01:11:26,543 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741850_1026 (size=5532) 2024-12-02T01:11:26,543 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741850_1026 (size=5532) 2024-12-02T01:11:26,544 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=264 B at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/.tmp/table/c4f814ac17624b5d99627371fd082b5e 2024-12-02T01:11:26,549 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/.tmp/info/711e802cb34444508451868ba942640d as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/info/711e802cb34444508451868ba942640d 2024-12-02T01:11:26,554 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/info/711e802cb34444508451868ba942640d, entries=20, sequenceid=14, filesize=8.2 K 2024-12-02T01:11:26,555 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/.tmp/table/c4f814ac17624b5d99627371fd082b5e as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/table/c4f814ac17624b5d99627371fd082b5e 2024-12-02T01:11:26,561 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/table/c4f814ac17624b5d99627371fd082b5e, entries=4, sequenceid=14, filesize=5.4 K 2024-12-02T01:11:26,562 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~3.05 KB/3122, heapSize ~5.27 KB/5400, currentSize=0 B/0 for 1588230740 in 69ms, sequenceid=14, compaction requested=false 2024-12-02T01:11:26,566 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/data/hbase/meta/1588230740/recovered.edits/17.seqid, newMaxSeqId=17, maxSeqId=1 2024-12-02T01:11:26,566 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T01:11:26,567 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:11:26,567 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:11:26,567 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T01:11:26,693 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,35559,1733101824300; all regions closed. 2024-12-02T01:11:26,693 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300 2024-12-02T01:11:26,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741834_1010 (size=4570) 2024-12-02T01:11:26,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741834_1010 (size=4570) 2024-12-02T01:11:26,703 DEBUG [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs 2024-12-02T01:11:26,704 INFO [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C35559%2C1733101824300.meta:.meta(num 1733101825678) 2024-12-02T01:11:26,704 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300 2024-12-02T01:11:26,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741847_1023 (size=1545) 2024-12-02T01:11:26,706 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741847_1023 (size=1545) 2024-12-02T01:11:26,885 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/WALs/a82f277fe18c,35559,1733101824300/a82f277fe18c%2C35559%2C1733101824300.1733101876456 to hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs/a82f277fe18c%2C35559%2C1733101824300.1733101876456 2024-12-02T01:11:26,892 DEBUG [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/oldWALs 2024-12-02T01:11:26,892 INFO [RS:0;a82f277fe18c:35559 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C35559%2C1733101824300:(num 1733101886473) 2024-12-02T01:11:26,893 DEBUG [RS:0;a82f277fe18c:35559 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:11:26,893 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:11:26,893 INFO [RS:0;a82f277fe18c:35559 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-02T01:11:26,894 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:11:26,895 INFO [RS:0;a82f277fe18c:35559 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:35559 2024-12-02T01:11:26,902 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,35559,1733101824300 2024-12-02T01:11:26,902 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:11:26,911 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,35559,1733101824300] 2024-12-02T01:11:26,911 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,35559,1733101824300; numProcessing=1 2024-12-02T01:11:26,919 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,35559,1733101824300 already deleted, retry=false 2024-12-02T01:11:26,919 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,35559,1733101824300 expired; onlineServers=0 2024-12-02T01:11:26,919 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,39899,1733101824140' ***** 2024-12-02T01:11:26,919 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T01:11:26,920 DEBUG [M:0;a82f277fe18c:39899 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6777c521, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:11:26,920 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,39899,1733101824140 2024-12-02T01:11:26,920 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,39899,1733101824140; all regions closed. 2024-12-02T01:11:26,920 DEBUG [M:0;a82f277fe18c:39899 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:11:26,920 DEBUG [M:0;a82f277fe18c:39899 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T01:11:26,920 DEBUG [M:0;a82f277fe18c:39899 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T01:11:26,920 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T01:11:26,920 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101825062 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101825062,5,FailOnTimeoutGroup] 2024-12-02T01:11:26,920 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101825060 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101825060,5,FailOnTimeoutGroup] 2024-12-02T01:11:26,921 INFO [M:0;a82f277fe18c:39899 {}] hbase.ChoreService(370): Chore service for: master/a82f277fe18c:0 had [] on shutdown 2024-12-02T01:11:26,921 DEBUG [M:0;a82f277fe18c:39899 {}] master.HMaster(1733): Stopping service threads 2024-12-02T01:11:26,921 INFO [M:0;a82f277fe18c:39899 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T01:11:26,921 INFO [M:0;a82f277fe18c:39899 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T01:11:26,921 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T01:11:26,928 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T01:11:26,928 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:26,928 DEBUG [M:0;a82f277fe18c:39899 {}] zookeeper.ZKUtil(347): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T01:11:26,928 WARN [M:0;a82f277fe18c:39899 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T01:11:26,928 INFO [M:0;a82f277fe18c:39899 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-02T01:11:26,928 INFO [M:0;a82f277fe18c:39899 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T01:11:26,928 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:11:26,928 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:26,929 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:26,929 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:11:26,929 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:26,929 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:11:26,929 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=65.02 KB heapSize=81.62 KB 2024-12-02T01:11:26,950 DEBUG [M:0;a82f277fe18c:39899 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/3b0e1089fdbd4669bb20ade5d12b39d8 is 82, key is hbase:meta,,1/info:regioninfo/1733101825705/Put/seqid=0 2024-12-02T01:11:26,954 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741851_1027 (size=5672) 2024-12-02T01:11:26,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741851_1027 (size=5672) 2024-12-02T01:11:26,955 INFO [M:0;a82f277fe18c:39899 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/3b0e1089fdbd4669bb20ade5d12b39d8 2024-12-02T01:11:26,973 DEBUG [M:0;a82f277fe18c:39899 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/7dc7dd7d80c44e06832266270cdbddb2 is 798, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733101826748/Put/seqid=0 2024-12-02T01:11:26,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741852_1028 (size=8350) 2024-12-02T01:11:26,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741852_1028 (size=8350) 2024-12-02T01:11:26,978 INFO [M:0;a82f277fe18c:39899 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=64.42 KB at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/7dc7dd7d80c44e06832266270cdbddb2 2024-12-02T01:11:26,982 INFO [M:0;a82f277fe18c:39899 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 7dc7dd7d80c44e06832266270cdbddb2 2024-12-02T01:11:26,994 DEBUG [M:0;a82f277fe18c:39899 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c4ee041fe57a480abd34669224d6b99e is 69, key is a82f277fe18c,35559,1733101824300/rs:state/1733101825171/Put/seqid=0 2024-12-02T01:11:26,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741853_1029 (size=5156) 2024-12-02T01:11:26,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741853_1029 (size=5156) 2024-12-02T01:11:27,000 INFO [M:0;a82f277fe18c:39899 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c4ee041fe57a480abd34669224d6b99e 2024-12-02T01:11:27,011 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:11:27,011 INFO [RS:0;a82f277fe18c:35559 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,35559,1733101824300; zookeeper connection closed. 2024-12-02T01:11:27,011 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35559-0x10194e0caf20001, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:11:27,011 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@16aba7b8 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@16aba7b8 2024-12-02T01:11:27,011 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T01:11:27,016 DEBUG [M:0;a82f277fe18c:39899 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/79f722bcab594e51a0dca74ea536ad38 is 52, key is load_balancer_on/state:d/1733101826347/Put/seqid=0 2024-12-02T01:11:27,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741854_1030 (size=5056) 2024-12-02T01:11:27,021 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741854_1030 (size=5056) 2024-12-02T01:11:27,021 INFO [M:0;a82f277fe18c:39899 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/79f722bcab594e51a0dca74ea536ad38 2024-12-02T01:11:27,026 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/3b0e1089fdbd4669bb20ade5d12b39d8 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/3b0e1089fdbd4669bb20ade5d12b39d8 2024-12-02T01:11:27,030 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/3b0e1089fdbd4669bb20ade5d12b39d8, entries=8, sequenceid=184, filesize=5.5 K 2024-12-02T01:11:27,031 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/7dc7dd7d80c44e06832266270cdbddb2 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/7dc7dd7d80c44e06832266270cdbddb2 2024-12-02T01:11:27,035 INFO [M:0;a82f277fe18c:39899 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 7dc7dd7d80c44e06832266270cdbddb2 2024-12-02T01:11:27,035 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/7dc7dd7d80c44e06832266270cdbddb2, entries=21, sequenceid=184, filesize=8.2 K 2024-12-02T01:11:27,036 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c4ee041fe57a480abd34669224d6b99e as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/c4ee041fe57a480abd34669224d6b99e 2024-12-02T01:11:27,040 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/c4ee041fe57a480abd34669224d6b99e, entries=1, sequenceid=184, filesize=5.0 K 2024-12-02T01:11:27,041 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/79f722bcab594e51a0dca74ea536ad38 as hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/79f722bcab594e51a0dca74ea536ad38 2024-12-02T01:11:27,045 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43627/user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/79f722bcab594e51a0dca74ea536ad38, entries=1, sequenceid=184, filesize=4.9 K 2024-12-02T01:11:27,046 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(3040): Finished flush of dataSize ~65.02 KB/66583, heapSize ~81.55 KB/83512, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 117ms, sequenceid=184, compaction requested=false 2024-12-02T01:11:27,048 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:27,048 DEBUG [M:0;a82f277fe18c:39899 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:11:27,048 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/a7668323-8c93-8712-06f2-bd25e01604a9/MasterData/WALs/a82f277fe18c,39899,1733101824140 2024-12-02T01:11:27,050 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46591 is added to blk_1073741830_1006 (size=79104) 2024-12-02T01:11:27,050 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41581 is added to blk_1073741830_1006 (size=79104) 2024-12-02T01:11:27,050 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:11:27,050 INFO [M:0;a82f277fe18c:39899 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-02T01:11:27,051 INFO [M:0;a82f277fe18c:39899 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:39899 2024-12-02T01:11:27,061 DEBUG [M:0;a82f277fe18c:39899 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/a82f277fe18c,39899,1733101824140 already deleted, retry=false 2024-12-02T01:11:27,170 INFO [M:0;a82f277fe18c:39899 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,39899,1733101824140; zookeeper connection closed. 2024-12-02T01:11:27,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:11:27,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39899-0x10194e0caf20000, quorum=127.0.0.1:51080, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:11:27,202 INFO [regionserver/a82f277fe18c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:11:27,204 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@62cfa257{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:11:27,205 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4bc05900{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:11:27,205 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:11:27,205 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2b0f6823{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:11:27,205 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5c2ef9ae{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir/,STOPPED} 2024-12-02T01:11:27,206 WARN [BP-2004375065-172.17.0.2-1733101821951 heartbeating to localhost/127.0.0.1:43627 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:11:27,206 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:11:27,206 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:11:27,206 WARN [BP-2004375065-172.17.0.2-1733101821951 heartbeating to localhost/127.0.0.1:43627 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2004375065-172.17.0.2-1733101821951 (Datanode Uuid dd9ba3b6-9e54-4e3d-9d8b-4afc7a487faa) service to localhost/127.0.0.1:43627 2024-12-02T01:11:27,207 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data3/current/BP-2004375065-172.17.0.2-1733101821951 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:11:27,207 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data4/current/BP-2004375065-172.17.0.2-1733101821951 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:11:27,207 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:11:27,209 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4cd1618{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:11:27,209 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4f062b37{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:11:27,209 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:11:27,209 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@30d4a328{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:11:27,209 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@23e7f2aa{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir/,STOPPED} 2024-12-02T01:11:27,210 WARN [BP-2004375065-172.17.0.2-1733101821951 heartbeating to localhost/127.0.0.1:43627 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:11:27,210 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:11:27,210 WARN [BP-2004375065-172.17.0.2-1733101821951 heartbeating to localhost/127.0.0.1:43627 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2004375065-172.17.0.2-1733101821951 (Datanode Uuid bf2c66f6-ccb2-45df-8a73-eaac12153c2d) service to localhost/127.0.0.1:43627 2024-12-02T01:11:27,210 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:11:27,210 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data1/current/BP-2004375065-172.17.0.2-1733101821951 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:11:27,211 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/cluster_3ebb1344-6c5d-a625-6580-ad8441cc3360/dfs/data/data2/current/BP-2004375065-172.17.0.2-1733101821951 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:11:27,211 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:11:27,216 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2896003d{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:11:27,217 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5f1edc2d{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:11:27,217 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:11:27,217 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@dc1c4c3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:11:27,217 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5c883b08{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir/,STOPPED} 2024-12-02T01:11:27,223 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-02T01:11:27,251 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-02T01:11:27,256 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testCompactionRecordDoesntBlockRolling Thread=111 (was 103) - Thread LEAK? -, OpenFileDescriptor=466 (was 450) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=212 (was 181) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=5320 (was 6082) 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRolling Thread=112, OpenFileDescriptor=466, MaxFileDescriptor=1048576, SystemLoadAverage=212, ProcessCount=11, AvailableMemoryMB=5320 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.log.dir so I do NOT create it in target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5fa197ed-6f27-8059-e8ee-5421fc182556/hadoop.tmp.dir so I do NOT create it in target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33, deleteOnExit=true 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/test.cache.data in system properties and HBase conf 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir in system properties and HBase conf 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T01:11:27,262 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-02T01:11:27,263 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/nfs.dump.dir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/java.io.tmpdir in system properties and HBase conf 2024-12-02T01:11:27,263 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:11:27,264 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T01:11:27,264 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T01:11:27,276 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:11:27,420 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:27,521 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:11:27,524 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:11:27,526 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:11:27,526 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:11:27,526 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:11:27,526 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:11:27,526 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3741aa5c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:11:27,527 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6780485a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:11:27,617 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3a6af3f1{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/java.io.tmpdir/jetty-localhost-38741-hadoop-hdfs-3_4_1-tests_jar-_-any-15339208244186319206/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:11:27,618 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4525abcd{HTTP/1.1, (http/1.1)}{localhost:38741} 2024-12-02T01:11:27,618 INFO [Time-limited test {}] server.Server(415): Started @294746ms 2024-12-02T01:11:27,628 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:11:27,845 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:11:27,848 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:11:27,848 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:11:27,848 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:11:27,848 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:11:27,849 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@68604cef{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:11:27,849 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1df47ba6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:11:27,942 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4aa14da8{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/java.io.tmpdir/jetty-localhost-40231-hadoop-hdfs-3_4_1-tests_jar-_-any-8239857113748420759/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:11:27,943 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@55682800{HTTP/1.1, (http/1.1)}{localhost:40231} 2024-12-02T01:11:27,943 INFO [Time-limited test {}] server.Server(415): Started @295071ms 2024-12-02T01:11:27,943 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:11:27,971 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:11:27,974 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:11:27,974 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:11:27,974 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:11:27,974 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:11:27,975 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3640f2d6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:11:27,975 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@21bfc9c0{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:11:28,066 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@723a744c{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/java.io.tmpdir/jetty-localhost-35251-hadoop-hdfs-3_4_1-tests_jar-_-any-17278782983499968808/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:11:28,066 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4dc1d62e{HTTP/1.1, (http/1.1)}{localhost:35251} 2024-12-02T01:11:28,066 INFO [Time-limited test {}] server.Server(415): Started @295194ms 2024-12-02T01:11:28,067 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:11:28,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-02T01:11:28,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-02T01:11:28,354 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:11:28,421 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:28,618 WARN [Thread-1732 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data2/current/BP-365843615-172.17.0.2-1733101887286/current, will proceed with Du for space computation calculation, 2024-12-02T01:11:28,618 WARN [Thread-1731 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data1/current/BP-365843615-172.17.0.2-1733101887286/current, will proceed with Du for space computation calculation, 2024-12-02T01:11:28,633 WARN [Thread-1696 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:11:28,636 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2c747922c1f07c83 with lease ID 0x1108a581d52681f: Processing first storage report for DS-fc4db1d9-7fda-4749-aa99-2b5888f878b1 from datanode DatanodeRegistration(127.0.0.1:43899, datanodeUuid=1451500d-ad1d-4e1c-8a7f-5a80ce5c779d, infoPort=45051, infoSecurePort=0, ipcPort=36663, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286) 2024-12-02T01:11:28,636 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2c747922c1f07c83 with lease ID 0x1108a581d52681f: from storage DS-fc4db1d9-7fda-4749-aa99-2b5888f878b1 node DatanodeRegistration(127.0.0.1:43899, datanodeUuid=1451500d-ad1d-4e1c-8a7f-5a80ce5c779d, infoPort=45051, infoSecurePort=0, ipcPort=36663, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:11:28,636 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x2c747922c1f07c83 with lease ID 0x1108a581d52681f: Processing first storage report for DS-407cdf51-06f8-41f4-8169-d454375aca50 from datanode DatanodeRegistration(127.0.0.1:43899, datanodeUuid=1451500d-ad1d-4e1c-8a7f-5a80ce5c779d, infoPort=45051, infoSecurePort=0, ipcPort=36663, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286) 2024-12-02T01:11:28,636 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x2c747922c1f07c83 with lease ID 0x1108a581d52681f: from storage DS-407cdf51-06f8-41f4-8169-d454375aca50 node DatanodeRegistration(127.0.0.1:43899, datanodeUuid=1451500d-ad1d-4e1c-8a7f-5a80ce5c779d, infoPort=45051, infoSecurePort=0, ipcPort=36663, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:11:28,891 WARN [Thread-1743 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data3/current/BP-365843615-172.17.0.2-1733101887286/current, will proceed with Du for space computation calculation, 2024-12-02T01:11:28,891 WARN [Thread-1744 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data4/current/BP-365843615-172.17.0.2-1733101887286/current, will proceed with Du for space computation calculation, 2024-12-02T01:11:28,908 WARN [Thread-1719 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:11:28,910 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x24059c15c2b4ecac with lease ID 0x1108a581d526820: Processing first storage report for DS-8a79cfbb-f00e-452e-a69d-61e40884bb27 from datanode DatanodeRegistration(127.0.0.1:39789, datanodeUuid=c36b8b12-0dce-4030-b8cd-5b907b6b0f7b, infoPort=34647, infoSecurePort=0, ipcPort=36533, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286) 2024-12-02T01:11:28,910 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x24059c15c2b4ecac with lease ID 0x1108a581d526820: from storage DS-8a79cfbb-f00e-452e-a69d-61e40884bb27 node DatanodeRegistration(127.0.0.1:39789, datanodeUuid=c36b8b12-0dce-4030-b8cd-5b907b6b0f7b, infoPort=34647, infoSecurePort=0, ipcPort=36533, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:11:28,910 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x24059c15c2b4ecac with lease ID 0x1108a581d526820: Processing first storage report for DS-c8770cb1-0dbb-430e-a0b4-5f7136e13b0f from datanode DatanodeRegistration(127.0.0.1:39789, datanodeUuid=c36b8b12-0dce-4030-b8cd-5b907b6b0f7b, infoPort=34647, infoSecurePort=0, ipcPort=36533, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286) 2024-12-02T01:11:28,910 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x24059c15c2b4ecac with lease ID 0x1108a581d526820: from storage DS-c8770cb1-0dbb-430e-a0b4-5f7136e13b0f node DatanodeRegistration(127.0.0.1:39789, datanodeUuid=c36b8b12-0dce-4030-b8cd-5b907b6b0f7b, infoPort=34647, infoSecurePort=0, ipcPort=36533, storageInfo=lv=-57;cid=testClusterID;nsid=1675551369;c=1733101887286), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:11:28,994 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd 2024-12-02T01:11:28,997 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/zookeeper_0, clientPort=64314, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T01:11:28,998 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=64314 2024-12-02T01:11:28,999 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,002 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:11:29,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:11:29,013 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7 with version=8 2024-12-02T01:11:29,013 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase-staging 2024-12-02T01:11:29,015 INFO [Time-limited test {}] client.ConnectionUtils(129): master/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:11:29,015 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:11:29,016 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:35029 2024-12-02T01:11:29,016 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,018 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,020 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:35029 connecting to ZooKeeper ensemble=127.0.0.1:64314 2024-12-02T01:11:29,092 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:350290x0, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:11:29,093 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:35029-0x10194e1c8590000 connected 2024-12-02T01:11:29,153 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:11:29,153 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:11:29,153 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:11:29,154 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35029 2024-12-02T01:11:29,154 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35029 2024-12-02T01:11:29,154 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35029 2024-12-02T01:11:29,155 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35029 2024-12-02T01:11:29,155 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35029 2024-12-02T01:11:29,155 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7, hbase.cluster.distributed=false 2024-12-02T01:11:29,168 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:11:29,169 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:32775 2024-12-02T01:11:29,170 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:11:29,170 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:11:29,171 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,172 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,174 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:32775 connecting to ZooKeeper ensemble=127.0.0.1:64314 2024-12-02T01:11:29,182 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:327750x0, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:11:29,183 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:32775-0x10194e1c8590001 connected 2024-12-02T01:11:29,183 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:11:29,183 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:11:29,184 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:11:29,184 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=32775 2024-12-02T01:11:29,184 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=32775 2024-12-02T01:11:29,184 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=32775 2024-12-02T01:11:29,185 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=32775 2024-12-02T01:11:29,185 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=32775 2024-12-02T01:11:29,186 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,194 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:11:29,194 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:11:29,194 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,197 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a82f277fe18c:35029 2024-12-02T01:11:29,202 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:11:29,202 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:11:29,202 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,202 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,203 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:11:29,203 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a82f277fe18c,35029,1733101889015 from backup master directory 2024-12-02T01:11:29,211 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:11:29,211 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,211 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:11:29,211 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:11:29,211 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,211 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:11:29,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:11:29,219 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:11:29,220 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/hbase.id with ID: c3649e28-ac58-44da-b5f5-024714f5c99c 2024-12-02T01:11:29,230 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:29,241 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,241 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,249 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:11:29,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:11:29,251 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:11:29,252 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T01:11:29,252 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:11:29,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:11:29,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:11:29,264 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store 2024-12-02T01:11:29,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:11:29,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:11:29,272 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:29,272 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:11:29,272 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:29,272 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:29,272 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:11:29,272 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:29,272 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:11:29,272 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:11:29,273 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/.initializing 2024-12-02T01:11:29,273 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/WALs/a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,276 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C35029%2C1733101889015, suffix=, logDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/WALs/a82f277fe18c,35029,1733101889015, archiveDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/oldWALs, maxLogs=10 2024-12-02T01:11:29,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C35029%2C1733101889015.1733101889277 2024-12-02T01:11:29,285 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/WALs/a82f277fe18c,35029,1733101889015/a82f277fe18c%2C35029%2C1733101889015.1733101889277 2024-12-02T01:11:29,286 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34647:34647),(127.0.0.1/127.0.0.1:45051:45051)] 2024-12-02T01:11:29,286 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:11:29,286 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:29,286 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,286 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,287 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,289 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T01:11:29,289 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,289 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:29,289 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,291 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T01:11:29,291 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,291 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:29,291 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,293 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T01:11:29,293 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,293 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:29,293 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,295 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T01:11:29,295 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,295 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:29,296 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,296 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,298 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T01:11:29,299 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:11:29,301 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:11:29,301 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=805822, jitterRate=0.024655818939208984}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T01:11:29,302 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:11:29,302 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T01:11:29,305 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@68b2abbb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:11:29,305 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-02T01:11:29,305 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T01:11:29,305 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T01:11:29,306 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T01:11:29,306 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T01:11:29,306 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-02T01:11:29,306 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T01:11:29,308 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T01:11:29,309 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T01:11:29,319 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-02T01:11:29,319 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T01:11:29,320 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T01:11:29,327 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-02T01:11:29,328 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T01:11:29,329 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T01:11:29,336 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-02T01:11:29,336 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T01:11:29,344 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T01:11:29,347 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T01:11:29,352 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T01:11:29,361 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:11:29,361 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:11:29,361 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,361 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,361 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=a82f277fe18c,35029,1733101889015, sessionid=0x10194e1c8590000, setting cluster-up flag (Was=false) 2024-12-02T01:11:29,377 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,377 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,402 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T01:11:29,405 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,422 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:29,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,453 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T01:11:29,455 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,35029,1733101889015 2024-12-02T01:11:29,457 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-02T01:11:29,457 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-02T01:11:29,457 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T01:11:29,457 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a82f277fe18c,35029,1733101889015 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a82f277fe18c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:11:29,458 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,459 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:11:29,459 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-02T01:11:29,460 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,460 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:11:29,461 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733101919461 2024-12-02T01:11:29,461 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T01:11:29,461 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T01:11:29,461 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T01:11:29,461 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T01:11:29,461 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T01:11:29,462 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T01:11:29,462 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101889462,5,FailOnTimeoutGroup] 2024-12-02T01:11:29,463 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101889463,5,FailOnTimeoutGroup] 2024-12-02T01:11:29,463 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,463 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T01:11:29,463 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,463 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:11:29,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:11:29,466 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-02T01:11:29,466 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7 2024-12-02T01:11:29,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:11:29,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:11:29,472 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:29,473 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:11:29,474 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:11:29,474 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,474 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:29,475 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:11:29,476 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:11:29,476 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,476 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:29,476 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:11:29,477 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:11:29,477 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:29,478 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:29,478 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740 2024-12-02T01:11:29,478 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740 2024-12-02T01:11:29,480 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:11:29,481 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:11:29,482 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:11:29,483 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=815829, jitterRate=0.03738148510456085}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:11:29,483 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:11:29,483 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:11:29,483 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:11:29,483 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:11:29,483 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:11:29,483 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:11:29,483 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:11:29,483 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:11:29,484 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:11:29,484 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-02T01:11:29,484 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T01:11:29,485 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T01:11:29,486 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T01:11:29,500 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a82f277fe18c:32775 2024-12-02T01:11:29,501 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1008): ClusterId : c3649e28-ac58-44da-b5f5-024714f5c99c 2024-12-02T01:11:29,501 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:11:29,511 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:11:29,511 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:11:29,520 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:11:29,520 DEBUG [RS:0;a82f277fe18c:32775 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@28f70e0f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:11:29,520 DEBUG [RS:0;a82f277fe18c:32775 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7f54013, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:11:29,521 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:11:29,521 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:11:29,521 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:11:29,521 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,35029,1733101889015 with isa=a82f277fe18c/172.17.0.2:32775, startcode=1733101889168 2024-12-02T01:11:29,521 DEBUG [RS:0;a82f277fe18c:32775 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:11:29,523 INFO [RS-EventLoopGroup-12-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51933, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:11:29,523 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35029 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,523 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35029 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,524 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7 2024-12-02T01:11:29,524 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:35829 2024-12-02T01:11:29,524 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:11:29,532 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:11:29,533 DEBUG [RS:0;a82f277fe18c:32775 {}] zookeeper.ZKUtil(111): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,533 WARN [RS:0;a82f277fe18c:32775 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:11:29,533 INFO [RS:0;a82f277fe18c:32775 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:11:29,533 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,534 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,32775,1733101889168] 2024-12-02T01:11:29,538 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:11:29,538 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:11:29,540 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:11:29,541 INFO [RS:0;a82f277fe18c:32775 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:11:29,541 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,541 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:11:29,542 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,542 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,542 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,542 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,542 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,542 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,542 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:11:29,543 DEBUG [RS:0;a82f277fe18c:32775 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:11:29,543 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,543 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,543 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,543 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,543 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,32775,1733101889168-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:11:29,556 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:11:29,557 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,32775,1733101889168-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:29,567 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.Replication(204): a82f277fe18c,32775,1733101889168 started 2024-12-02T01:11:29,568 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,32775,1733101889168, RpcServer on a82f277fe18c/172.17.0.2:32775, sessionid=0x10194e1c8590001 2024-12-02T01:11:29,568 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:11:29,568 DEBUG [RS:0;a82f277fe18c:32775 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,568 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,32775,1733101889168' 2024-12-02T01:11:29,568 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:11:29,568 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,32775,1733101889168' 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:11:29,569 DEBUG [RS:0;a82f277fe18c:32775 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:11:29,569 INFO [RS:0;a82f277fe18c:32775 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:11:29,569 INFO [RS:0;a82f277fe18c:32775 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:11:29,636 WARN [a82f277fe18c:35029 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-02T01:11:29,673 INFO [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C32775%2C1733101889168, suffix=, logDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168, archiveDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/oldWALs, maxLogs=32 2024-12-02T01:11:29,674 INFO [RS:0;a82f277fe18c:32775 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C32775%2C1733101889168.1733101889674 2024-12-02T01:11:29,684 INFO [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101889674 2024-12-02T01:11:29,684 DEBUG [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34647:34647),(127.0.0.1/127.0.0.1:45051:45051)] 2024-12-02T01:11:29,887 DEBUG [a82f277fe18c:35029 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T01:11:29,887 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:29,889 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,32775,1733101889168, state=OPENING 2024-12-02T01:11:29,933 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T01:11:29,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,944 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:29,946 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:11:29,946 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:11:29,946 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=a82f277fe18c,32775,1733101889168}] 2024-12-02T01:11:30,101 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,32775,1733101889168 2024-12-02T01:11:30,101 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T01:11:30,104 INFO [RS-EventLoopGroup-13-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38850, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T01:11:30,112 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-02T01:11:30,112 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:11:30,116 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C32775%2C1733101889168.meta, suffix=.meta, logDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168, archiveDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/oldWALs, maxLogs=32 2024-12-02T01:11:30,117 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C32775%2C1733101889168.meta.1733101890117.meta 2024-12-02T01:11:30,124 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.meta.1733101890117.meta 2024-12-02T01:11:30,124 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45051:45051),(127.0.0.1/127.0.0.1:34647:34647)] 2024-12-02T01:11:30,124 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:11:30,124 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T01:11:30,124 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T01:11:30,124 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T01:11:30,124 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T01:11:30,124 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:30,125 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-02T01:11:30,125 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-02T01:11:30,126 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:11:30,127 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:11:30,127 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:30,127 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:30,127 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:11:30,128 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:11:30,128 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:30,128 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:30,128 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:11:30,129 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:11:30,129 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:30,129 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:11:30,130 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740 2024-12-02T01:11:30,131 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740 2024-12-02T01:11:30,132 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:11:30,133 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:11:30,134 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=805578, jitterRate=0.024346306920051575}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:11:30,134 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:11:30,134 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733101890101 2024-12-02T01:11:30,136 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T01:11:30,136 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-02T01:11:30,136 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:30,137 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,32775,1733101889168, state=OPEN 2024-12-02T01:11:30,161 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:11:30,161 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:11:30,162 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:11:30,162 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:11:30,163 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T01:11:30,163 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=a82f277fe18c,32775,1733101889168 in 215 msec 2024-12-02T01:11:30,165 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T01:11:30,165 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 679 msec 2024-12-02T01:11:30,167 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 710 msec 2024-12-02T01:11:30,168 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733101890167, completionTime=-1 2024-12-02T01:11:30,168 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T01:11:30,168 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-02T01:11:30,169 DEBUG [hconnection-0x17105236-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:11:30,170 INFO [RS-EventLoopGroup-13-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38864, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:11:30,171 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-02T01:11:30,171 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733101950171 2024-12-02T01:11:30,171 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733102010171 2024-12-02T01:11:30,171 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 3 msec 2024-12-02T01:11:30,194 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35029,1733101889015-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:30,194 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35029,1733101889015-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:30,195 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35029,1733101889015-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:30,195 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a82f277fe18c:35029, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:30,195 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:30,195 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-02T01:11:30,195 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:11:30,196 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-02T01:11:30,197 DEBUG [master/a82f277fe18c:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-02T01:11:30,197 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:11:30,197 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:30,198 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:11:30,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:11:30,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:11:30,209 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 5acee627f693bbca053dc704b61b658f, NAME => 'hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7 2024-12-02T01:11:30,217 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:11:30,217 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:11:30,221 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:30,221 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 5acee627f693bbca053dc704b61b658f, disabling compactions & flushes 2024-12-02T01:11:30,221 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,221 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,221 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. after waiting 0 ms 2024-12-02T01:11:30,221 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,221 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,221 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 5acee627f693bbca053dc704b61b658f: 2024-12-02T01:11:30,222 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:11:30,222 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733101890222"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101890222"}]},"ts":"1733101890222"} 2024-12-02T01:11:30,224 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:11:30,225 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:11:30,225 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101890225"}]},"ts":"1733101890225"} 2024-12-02T01:11:30,227 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-02T01:11:30,274 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=5acee627f693bbca053dc704b61b658f, ASSIGN}] 2024-12-02T01:11:30,276 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=5acee627f693bbca053dc704b61b658f, ASSIGN 2024-12-02T01:11:30,278 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=5acee627f693bbca053dc704b61b658f, ASSIGN; state=OFFLINE, location=a82f277fe18c,32775,1733101889168; forceNewPlan=false, retain=false 2024-12-02T01:11:30,423 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:30,429 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=5acee627f693bbca053dc704b61b658f, regionState=OPENING, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:30,431 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 5acee627f693bbca053dc704b61b658f, server=a82f277fe18c,32775,1733101889168}] 2024-12-02T01:11:30,583 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,32775,1733101889168 2024-12-02T01:11:30,588 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,589 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 5acee627f693bbca053dc704b61b658f, NAME => 'hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:11:30,589 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,589 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:30,589 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,589 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,591 INFO [StoreOpener-5acee627f693bbca053dc704b61b658f-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,593 INFO [StoreOpener-5acee627f693bbca053dc704b61b658f-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 5acee627f693bbca053dc704b61b658f columnFamilyName info 2024-12-02T01:11:30,593 DEBUG [StoreOpener-5acee627f693bbca053dc704b61b658f-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:30,594 INFO [StoreOpener-5acee627f693bbca053dc704b61b658f-1 {}] regionserver.HStore(327): Store=5acee627f693bbca053dc704b61b658f/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:30,595 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,595 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,598 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 5acee627f693bbca053dc704b61b658f 2024-12-02T01:11:30,601 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:11:30,602 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 5acee627f693bbca053dc704b61b658f; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=802806, jitterRate=0.02082096040248871}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:11:30,603 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 5acee627f693bbca053dc704b61b658f: 2024-12-02T01:11:30,604 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f., pid=6, masterSystemTime=1733101890583 2024-12-02T01:11:30,606 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,606 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:11:30,606 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=5acee627f693bbca053dc704b61b658f, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:30,610 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T01:11:30,610 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 5acee627f693bbca053dc704b61b658f, server=a82f277fe18c,32775,1733101889168 in 178 msec 2024-12-02T01:11:30,612 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T01:11:30,612 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=5acee627f693bbca053dc704b61b658f, ASSIGN in 336 msec 2024-12-02T01:11:30,612 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:11:30,613 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101890612"}]},"ts":"1733101890612"} 2024-12-02T01:11:30,614 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-02T01:11:30,634 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:11:30,636 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 439 msec 2024-12-02T01:11:30,698 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-02T01:11:30,708 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:11:30,708 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:30,708 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:11:30,716 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-02T01:11:30,732 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:11:30,743 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 28 msec 2024-12-02T01:11:30,747 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-02T01:11:30,761 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:11:30,771 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 23 msec 2024-12-02T01:11:30,794 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-02T01:11:30,811 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.600sec 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35029,1733101889015-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:11:30,811 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35029,1733101889015-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T01:11:30,813 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-02T01:11:30,813 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T01:11:30,813 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,35029,1733101889015-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:11:30,891 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x78fd5869 to 127.0.0.1:64314 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@70da4c6b 2024-12-02T01:11:30,903 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@abae079, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:11:30,906 DEBUG [hconnection-0x2a3d24ac-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:11:30,908 INFO [RS-EventLoopGroup-13-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38866, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:11:30,910 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=a82f277fe18c,35029,1733101889015 2024-12-02T01:11:30,911 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:11:30,914 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-02T01:11:30,915 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-02T01:11:30,919 INFO [RS-EventLoopGroup-12-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:34486, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-02T01:11:30,920 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-02T01:11:30,920 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-02T01:11:30,920 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testLogRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:11:30,921 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testLogRolling 2024-12-02T01:11:30,922 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:11:30,922 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:30,922 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRolling" procId is: 9 2024-12-02T01:11:30,923 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:11:30,923 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:11:30,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741837_1013 (size=381) 2024-12-02T01:11:30,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741837_1013 (size=381) 2024-12-02T01:11:30,932 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => fbd98b96465e3cae82d3686045e0a994, NAME => 'TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7 2024-12-02T01:11:30,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741838_1014 (size=64) 2024-12-02T01:11:30,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741838_1014 (size=64) 2024-12-02T01:11:30,940 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:30,940 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1681): Closing fbd98b96465e3cae82d3686045e0a994, disabling compactions & flushes 2024-12-02T01:11:30,940 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:30,940 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:30,940 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. after waiting 0 ms 2024-12-02T01:11:30,940 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:30,940 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:30,940 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1635): Region close journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:30,941 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:11:30,941 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.","families":{"info":[{"qualifier":"regioninfo","vlen":63,"tag":[],"timestamp":"1733101890941"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101890941"}]},"ts":"1733101890941"} 2024-12-02T01:11:30,942 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:11:30,943 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:11:30,943 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101890943"}]},"ts":"1733101890943"} 2024-12-02T01:11:30,944 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRolling, state=ENABLING in hbase:meta 2024-12-02T01:11:30,961 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, ASSIGN}] 2024-12-02T01:11:30,962 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, ASSIGN 2024-12-02T01:11:30,962 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, ASSIGN; state=OFFLINE, location=a82f277fe18c,32775,1733101889168; forceNewPlan=false, retain=false 2024-12-02T01:11:31,113 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=fbd98b96465e3cae82d3686045e0a994, regionState=OPENING, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:31,114 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168}] 2024-12-02T01:11:31,266 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to a82f277fe18c,32775,1733101889168 2024-12-02T01:11:31,269 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:31,269 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => fbd98b96465e3cae82d3686045e0a994, NAME => 'TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:11:31,270 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,270 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:31,270 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,270 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,271 INFO [StoreOpener-fbd98b96465e3cae82d3686045e0a994-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,272 INFO [StoreOpener-fbd98b96465e3cae82d3686045e0a994-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region fbd98b96465e3cae82d3686045e0a994 columnFamilyName info 2024-12-02T01:11:31,272 DEBUG [StoreOpener-fbd98b96465e3cae82d3686045e0a994-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:31,273 INFO [StoreOpener-fbd98b96465e3cae82d3686045e0a994-1 {}] regionserver.HStore(327): Store=fbd98b96465e3cae82d3686045e0a994/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:31,274 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,274 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,276 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:31,278 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:11:31,279 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened fbd98b96465e3cae82d3686045e0a994; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=790641, jitterRate=0.00535312294960022}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:11:31,279 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:31,280 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994., pid=11, masterSystemTime=1733101891266 2024-12-02T01:11:31,282 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:31,282 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:31,282 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=fbd98b96465e3cae82d3686045e0a994, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:31,286 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-02T01:11:31,286 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168 in 170 msec 2024-12-02T01:11:31,288 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-02T01:11:31,288 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, ASSIGN in 326 msec 2024-12-02T01:11:31,288 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:11:31,288 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101891288"}]},"ts":"1733101891288"} 2024-12-02T01:11:31,290 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRolling, state=ENABLED in hbase:meta 2024-12-02T01:11:31,305 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:11:31,306 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testLogRolling in 385 msec 2024-12-02T01:11:31,423 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:31,525 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,525 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,525 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,526 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,526 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,526 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,536 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,536 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,536 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,537 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,539 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,539 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,539 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:31,541 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,045 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:11:32,046 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,046 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,047 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,047 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,047 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,047 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,063 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,064 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,064 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,064 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,064 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,064 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,067 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,067 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,067 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,069 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:32,424 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:33,425 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:34,425 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:35,056 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 61b9a0fcb731850a98b7b912948ae7ef, had cached 0 bytes from a total of 23930 2024-12-02T01:11:35,426 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:35,538 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T01:11:35,540 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-02T01:11:35,541 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRolling' 2024-12-02T01:11:36,427 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:37,428 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:38,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:11:38,353 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2024-12-02T01:11:38,354 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-02T01:11:38,354 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-02T01:11:38,355 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling 2024-12-02T01:11:38,355 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling Metrics about Tables on a single HBase RegionServer 2024-12-02T01:11:38,428 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:38,823 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=1, created chunk count=11, reused chunk count=38, reuseRatio=77.55% 2024-12-02T01:11:38,823 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2024-12-02T01:11:39,429 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:40,430 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:40,926 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35029 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-02T01:11:40,926 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testLogRolling, procId: 9 completed 2024-12-02T01:11:40,929 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testLogRolling 2024-12-02T01:11:40,929 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:40,942 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:40,942 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing fbd98b96465e3cae82d3686045e0a994 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:11:40,960 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/ae1e70fcdb7e4b7bb76c9e9d2097708a is 1080, key is row0001/info:/1733101900933/Put/seqid=0 2024-12-02T01:11:40,965 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741839_1015 (size=12509) 2024-12-02T01:11:40,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741839_1015 (size=12509) 2024-12-02T01:11:40,966 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/ae1e70fcdb7e4b7bb76c9e9d2097708a 2024-12-02T01:11:40,968 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-02T01:11:40,969 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] ipc.CallRunner(138): callId: 38 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:38866 deadline: 1733101910968, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:40,972 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/ae1e70fcdb7e4b7bb76c9e9d2097708a as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/ae1e70fcdb7e4b7bb76c9e9d2097708a 2024-12-02T01:11:40,977 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/ae1e70fcdb7e4b7bb76c9e9d2097708a, entries=7, sequenceid=11, filesize=12.2 K 2024-12-02T01:11:40,978 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for fbd98b96465e3cae82d3686045e0a994 in 36ms, sequenceid=11, compaction requested=false 2024-12-02T01:11:40,978 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:41,431 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:42,432 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:43,433 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:44,434 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:45,435 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:45,472 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,473 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,473 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,473 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,473 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,474 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,490 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,491 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,491 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,491 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,491 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,491 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,493 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,493 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,493 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:45,495 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,001 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:11:46,002 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,003 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,003 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,003 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,003 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,003 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,017 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,018 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,018 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,018 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,018 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,018 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,021 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,021 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,022 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,024 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:11:46,435 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:47,436 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:48,437 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:49,438 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:50,440 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:51,063 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:51,063 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing fbd98b96465e3cae82d3686045e0a994 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-02T01:11:51,073 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/793526e53dbf4caab08dde710a4ee2c7 is 1080, key is row0008/info:/1733101900943/Put/seqid=0 2024-12-02T01:11:51,078 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741840_1016 (size=29761) 2024-12-02T01:11:51,078 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741840_1016 (size=29761) 2024-12-02T01:11:51,079 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=37 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/793526e53dbf4caab08dde710a4ee2c7 2024-12-02T01:11:51,084 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/793526e53dbf4caab08dde710a4ee2c7 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7 2024-12-02T01:11:51,089 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7, entries=23, sequenceid=37, filesize=29.1 K 2024-12-02T01:11:51,089 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=2.10 KB/2152 for fbd98b96465e3cae82d3686045e0a994 in 26ms, sequenceid=37, compaction requested=false 2024-12-02T01:11:51,089 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:51,090 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=41.3 K, sizeToCheck=16.0 K 2024-12-02T01:11:51,090 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:11:51,090 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7 because midkey is the same as first or last row 2024-12-02T01:11:51,440 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:52,442 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:53,083 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:53,084 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing fbd98b96465e3cae82d3686045e0a994 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:11:53,089 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/cc15baeab249430b988d627b26287160 is 1080, key is row0031/info:/1733101911065/Put/seqid=0 2024-12-02T01:11:53,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741841_1017 (size=12509) 2024-12-02T01:11:53,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741841_1017 (size=12509) 2024-12-02T01:11:53,094 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=47 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/cc15baeab249430b988d627b26287160 2024-12-02T01:11:53,101 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/cc15baeab249430b988d627b26287160 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/cc15baeab249430b988d627b26287160 2024-12-02T01:11:53,107 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/cc15baeab249430b988d627b26287160, entries=7, sequenceid=47, filesize=12.2 K 2024-12-02T01:11:53,108 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=16.81 KB/17216 for fbd98b96465e3cae82d3686045e0a994 in 24ms, sequenceid=47, compaction requested=true 2024-12-02T01:11:53,108 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:53,108 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=53.5 K, sizeToCheck=16.0 K 2024-12-02T01:11:53,108 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:11:53,108 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7 because midkey is the same as first or last row 2024-12-02T01:11:53,109 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store fbd98b96465e3cae82d3686045e0a994:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:11:53,109 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:53,109 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:11:53,109 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:53,109 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing fbd98b96465e3cae82d3686045e0a994 1/1 column families, dataSize=17.86 KB heapSize=19.38 KB 2024-12-02T01:11:53,110 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 54779 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:11:53,110 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): fbd98b96465e3cae82d3686045e0a994/info is initiating minor compaction (all files) 2024-12-02T01:11:53,110 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of fbd98b96465e3cae82d3686045e0a994/info in TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:53,111 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/ae1e70fcdb7e4b7bb76c9e9d2097708a, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/cc15baeab249430b988d627b26287160] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp, totalSize=53.5 K 2024-12-02T01:11:53,111 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting ae1e70fcdb7e4b7bb76c9e9d2097708a, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733101900933 2024-12-02T01:11:53,111 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 793526e53dbf4caab08dde710a4ee2c7, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=37, earliestPutTs=1733101900943 2024-12-02T01:11:53,112 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting cc15baeab249430b988d627b26287160, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=47, earliestPutTs=1733101911065 2024-12-02T01:11:53,114 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/c0d6b9d1ca384acabd0167021ead93ff is 1080, key is row0038/info:/1733101913084/Put/seqid=0 2024-12-02T01:11:53,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741842_1018 (size=23299) 2024-12-02T01:11:53,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741842_1018 (size=23299) 2024-12-02T01:11:53,129 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=17.86 KB at sequenceid=67 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/c0d6b9d1ca384acabd0167021ead93ff 2024-12-02T01:11:53,135 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/c0d6b9d1ca384acabd0167021ead93ff as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/c0d6b9d1ca384acabd0167021ead93ff 2024-12-02T01:11:53,137 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): fbd98b96465e3cae82d3686045e0a994#info#compaction#45 average throughput is 12.66 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:11:53,137 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/aeb262a5c6d249f68de026fa8019ca8f is 1080, key is row0001/info:/1733101900933/Put/seqid=0 2024-12-02T01:11:53,141 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/c0d6b9d1ca384acabd0167021ead93ff, entries=17, sequenceid=67, filesize=22.8 K 2024-12-02T01:11:53,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741843_1019 (size=44978) 2024-12-02T01:11:53,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741843_1019 (size=44978) 2024-12-02T01:11:53,142 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~17.86 KB/18292, heapSize ~19.36 KB/19824, currentSize=10.51 KB/10760 for fbd98b96465e3cae82d3686045e0a994 in 33ms, sequenceid=67, compaction requested=false 2024-12-02T01:11:53,142 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:53,142 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=76.2 K, sizeToCheck=16.0 K 2024-12-02T01:11:53,142 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:11:53,142 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7 because midkey is the same as first or last row 2024-12-02T01:11:53,149 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/aeb262a5c6d249f68de026fa8019ca8f as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f 2024-12-02T01:11:53,155 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in fbd98b96465e3cae82d3686045e0a994/info of fbd98b96465e3cae82d3686045e0a994 into aeb262a5c6d249f68de026fa8019ca8f(size=43.9 K), total size for store is 66.7 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:11:53,155 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:53,156 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994., storeName=fbd98b96465e3cae82d3686045e0a994/info, priority=13, startTime=1733101913108; duration=0sec 2024-12-02T01:11:53,156 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=66.7 K, sizeToCheck=16.0 K 2024-12-02T01:11:53,156 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:11:53,156 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f because midkey is the same as first or last row 2024-12-02T01:11:53,156 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:53,156 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: fbd98b96465e3cae82d3686045e0a994:info 2024-12-02T01:11:53,443 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:54,444 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:55,127 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,127 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing fbd98b96465e3cae82d3686045e0a994 1/1 column families, dataSize=11.56 KB heapSize=12.63 KB 2024-12-02T01:11:55,137 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/674ab235d6534bc4b6cb6ca15a03499b is 1080, key is row0055/info:/1733101913110/Put/seqid=0 2024-12-02T01:11:55,143 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741844_1020 (size=16817) 2024-12-02T01:11:55,143 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741844_1020 (size=16817) 2024-12-02T01:11:55,147 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=11.56 KB at sequenceid=82 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/674ab235d6534bc4b6cb6ca15a03499b 2024-12-02T01:11:55,154 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/674ab235d6534bc4b6cb6ca15a03499b as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/674ab235d6534bc4b6cb6ca15a03499b 2024-12-02T01:11:55,159 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/674ab235d6534bc4b6cb6ca15a03499b, entries=11, sequenceid=82, filesize=16.4 K 2024-12-02T01:11:55,159 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-02T01:11:55,159 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] ipc.CallRunner(138): callId: 94 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:38866 deadline: 1733101925159, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,160 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~11.56 KB/11836, heapSize ~12.61 KB/12912, currentSize=18.91 KB/19368 for fbd98b96465e3cae82d3686045e0a994 in 33ms, sequenceid=82, compaction requested=true 2024-12-02T01:11:55,160 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:55,160 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=83.1 K, sizeToCheck=16.0 K 2024-12-02T01:11:55,160 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:11:55,160 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f because midkey is the same as first or last row 2024-12-02T01:11:55,160 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store fbd98b96465e3cae82d3686045e0a994:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:11:55,160 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,160 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:11:55,161 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 85094 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:11:55,161 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): fbd98b96465e3cae82d3686045e0a994/info is initiating minor compaction (all files) 2024-12-02T01:11:55,161 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of fbd98b96465e3cae82d3686045e0a994/info in TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:55,161 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/c0d6b9d1ca384acabd0167021ead93ff, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/674ab235d6534bc4b6cb6ca15a03499b] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp, totalSize=83.1 K 2024-12-02T01:11:55,161 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting aeb262a5c6d249f68de026fa8019ca8f, keycount=37, bloomtype=ROW, size=43.9 K, encoding=NONE, compression=NONE, seqNum=47, earliestPutTs=1733101900933 2024-12-02T01:11:55,162 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting c0d6b9d1ca384acabd0167021ead93ff, keycount=17, bloomtype=ROW, size=22.8 K, encoding=NONE, compression=NONE, seqNum=67, earliestPutTs=1733101913084 2024-12-02T01:11:55,162 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 674ab235d6534bc4b6cb6ca15a03499b, keycount=11, bloomtype=ROW, size=16.4 K, encoding=NONE, compression=NONE, seqNum=82, earliestPutTs=1733101913110 2024-12-02T01:11:55,172 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): fbd98b96465e3cae82d3686045e0a994#info#compaction#47 average throughput is 33.35 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:11:55,172 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/a768bc17c29b4ff89ab23cfced6233eb is 1080, key is row0001/info:/1733101900933/Put/seqid=0 2024-12-02T01:11:55,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741845_1021 (size=75378) 2024-12-02T01:11:55,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741845_1021 (size=75378) 2024-12-02T01:11:55,181 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/a768bc17c29b4ff89ab23cfced6233eb as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb 2024-12-02T01:11:55,209 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in fbd98b96465e3cae82d3686045e0a994/info of fbd98b96465e3cae82d3686045e0a994 into a768bc17c29b4ff89ab23cfced6233eb(size=73.6 K), total size for store is 73.6 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:11:55,209 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:55,209 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994., storeName=fbd98b96465e3cae82d3686045e0a994/info, priority=13, startTime=1733101915160; duration=0sec 2024-12-02T01:11:55,209 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=73.6 K, sizeToCheck=16.0 K 2024-12-02T01:11:55,209 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-02T01:11:55,210 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit(239): Splitting TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994., compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,210 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,210 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: fbd98b96465e3cae82d3686045e0a994:info 2024-12-02T01:11:55,212 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35029 {}] assignment.AssignmentManager(1346): Split request from a82f277fe18c,32775,1733101889168, parent={ENCODED => fbd98b96465e3cae82d3686045e0a994, NAME => 'TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.', STARTKEY => '', ENDKEY => ''}, splitKey=row0062 2024-12-02T01:11:55,215 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35029 {}] assignment.SplitTableRegionProcedure(223): Splittable=true state=OPEN, location=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,219 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35029 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=fbd98b96465e3cae82d3686045e0a994, daughterA=38ed7bb40b7df5cbab5b70514657b066, daughterB=761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,220 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=fbd98b96465e3cae82d3686045e0a994, daughterA=38ed7bb40b7df5cbab5b70514657b066, daughterB=761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,220 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=fbd98b96465e3cae82d3686045e0a994, daughterA=38ed7bb40b7df5cbab5b70514657b066, daughterB=761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,220 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=fbd98b96465e3cae82d3686045e0a994, daughterA=38ed7bb40b7df5cbab5b70514657b066, daughterB=761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,226 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, UNASSIGN}] 2024-12-02T01:11:55,227 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, UNASSIGN 2024-12-02T01:11:55,228 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=fbd98b96465e3cae82d3686045e0a994, regionState=CLOSING, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,230 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2024-12-02T01:11:55,230 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=14, ppid=13, state=RUNNABLE; CloseRegionProcedure fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168}] 2024-12-02T01:11:55,386 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,387 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] handler.UnassignRegionHandler(124): Close fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,387 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2024-12-02T01:11:55,388 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1681): Closing fbd98b96465e3cae82d3686045e0a994, disabling compactions & flushes 2024-12-02T01:11:55,388 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:55,388 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:55,388 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. after waiting 0 ms 2024-12-02T01:11:55,388 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:55,388 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(2837): Flushing fbd98b96465e3cae82d3686045e0a994 1/1 column families, dataSize=18.91 KB heapSize=20.50 KB 2024-12-02T01:11:55,392 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/5f5f45f88d654c339e045094d137b5cc is 1080, key is row0066/info:/1733101915129/Put/seqid=0 2024-12-02T01:11:55,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741846_1022 (size=24376) 2024-12-02T01:11:55,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741846_1022 (size=24376) 2024-12-02T01:11:55,397 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=18.91 KB at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/5f5f45f88d654c339e045094d137b5cc 2024-12-02T01:11:55,403 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/.tmp/info/5f5f45f88d654c339e045094d137b5cc as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/5f5f45f88d654c339e045094d137b5cc 2024-12-02T01:11:55,409 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/5f5f45f88d654c339e045094d137b5cc, entries=18, sequenceid=104, filesize=23.8 K 2024-12-02T01:11:55,410 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(3040): Finished flush of dataSize ~18.91 KB/19368, heapSize ~20.48 KB/20976, currentSize=0 B/0 for fbd98b96465e3cae82d3686045e0a994 in 22ms, sequenceid=104, compaction requested=false 2024-12-02T01:11:55,411 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/ae1e70fcdb7e4b7bb76c9e9d2097708a, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/cc15baeab249430b988d627b26287160, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/c0d6b9d1ca384acabd0167021ead93ff, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/674ab235d6534bc4b6cb6ca15a03499b] to archive 2024-12-02T01:11:55,412 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T01:11:55,414 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/ae1e70fcdb7e4b7bb76c9e9d2097708a to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/ae1e70fcdb7e4b7bb76c9e9d2097708a 2024-12-02T01:11:55,415 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/793526e53dbf4caab08dde710a4ee2c7 2024-12-02T01:11:55,416 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/aeb262a5c6d249f68de026fa8019ca8f 2024-12-02T01:11:55,417 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/cc15baeab249430b988d627b26287160 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/cc15baeab249430b988d627b26287160 2024-12-02T01:11:55,418 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/c0d6b9d1ca384acabd0167021ead93ff to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/c0d6b9d1ca384acabd0167021ead93ff 2024-12-02T01:11:55,419 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/674ab235d6534bc4b6cb6ca15a03499b to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/674ab235d6534bc4b6cb6ca15a03499b 2024-12-02T01:11:55,423 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/recovered.edits/107.seqid, newMaxSeqId=107, maxSeqId=1 2024-12-02T01:11:55,424 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. 2024-12-02T01:11:55,424 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1635): Region close journal for fbd98b96465e3cae82d3686045e0a994: 2024-12-02T01:11:55,425 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] handler.UnassignRegionHandler(170): Closed fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,426 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=fbd98b96465e3cae82d3686045e0a994, regionState=CLOSED 2024-12-02T01:11:55,429 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=14, resume processing ppid=13 2024-12-02T01:11:55,429 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, ppid=13, state=SUCCESS; CloseRegionProcedure fbd98b96465e3cae82d3686045e0a994, server=a82f277fe18c,32775,1733101889168 in 197 msec 2024-12-02T01:11:55,430 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=13, resume processing ppid=12 2024-12-02T01:11:55,430 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=fbd98b96465e3cae82d3686045e0a994, UNASSIGN in 203 msec 2024-12-02T01:11:55,445 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:55,447 DEBUG [PEWorker-5 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:55,449 INFO [PEWorker-5 {}] assignment.SplitTableRegionProcedure(728): pid=12 splitting 2 storefiles, region=fbd98b96465e3cae82d3686045e0a994, threads=2 2024-12-02T01:11:55,449 DEBUG [StoreFileSplitter-pool-1 {}] assignment.SplitTableRegionProcedure(823): pid=12 splitting started for store file: hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb for region: fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,449 DEBUG [StoreFileSplitter-pool-0 {}] assignment.SplitTableRegionProcedure(823): pid=12 splitting started for store file: hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/5f5f45f88d654c339e045094d137b5cc for region: fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,458 DEBUG [StoreFileSplitter-pool-0 {}] regionserver.HRegionFileSystem(650): Will create HFileLink file for hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/5f5f45f88d654c339e045094d137b5cc, top=true 2024-12-02T01:11:55,462 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741847_1023 (size=27) 2024-12-02T01:11:55,463 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741847_1023 (size=27) 2024-12-02T01:11:55,463 INFO [StoreFileSplitter-pool-0 {}] regionserver.HRegionFileSystem(691): Created linkFile:hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc for child: 761c6c5fe18cb86212f23a9aac9b6e3c, parent: fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,463 DEBUG [StoreFileSplitter-pool-0 {}] assignment.SplitTableRegionProcedure(834): pid=12 splitting complete for store file: hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/5f5f45f88d654c339e045094d137b5cc for region: fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741848_1024 (size=27) 2024-12-02T01:11:55,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741848_1024 (size=27) 2024-12-02T01:11:55,472 DEBUG [StoreFileSplitter-pool-1 {}] assignment.SplitTableRegionProcedure(834): pid=12 splitting complete for store file: hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb for region: fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:11:55,472 DEBUG [PEWorker-5 {}] assignment.SplitTableRegionProcedure(802): pid=12 split storefiles for region fbd98b96465e3cae82d3686045e0a994 Daughter A: [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994] storefiles, Daughter B: [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994] storefiles. 2024-12-02T01:11:55,481 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741849_1025 (size=71) 2024-12-02T01:11:55,481 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741849_1025 (size=71) 2024-12-02T01:11:55,483 DEBUG [PEWorker-5 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:55,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741850_1026 (size=71) 2024-12-02T01:11:55,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741850_1026 (size=71) 2024-12-02T01:11:55,495 DEBUG [PEWorker-5 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:55,505 DEBUG [PEWorker-5 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/recovered.edits/107.seqid, newMaxSeqId=107, maxSeqId=-1 2024-12-02T01:11:55,507 DEBUG [PEWorker-5 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/recovered.edits/107.seqid, newMaxSeqId=107, maxSeqId=-1 2024-12-02T01:11:55,509 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994.","families":{"info":[{"qualifier":"regioninfo","vlen":63,"tag":[],"timestamp":"1733101915508"},{"qualifier":"splitA","vlen":70,"tag":[],"timestamp":"1733101915508"},{"qualifier":"splitB","vlen":70,"tag":[],"timestamp":"1733101915508"}]},"ts":"1733101915508"} 2024-12-02T01:11:55,509 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.","families":{"info":[{"qualifier":"regioninfo","vlen":70,"tag":[],"timestamp":"1733101915508"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101915508"},{"qualifier":"seqnumDuringOpen","vlen":8,"tag":[],"timestamp":"1733101915508"}]},"ts":"1733101915508"} 2024-12-02T01:11:55,509 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.","families":{"info":[{"qualifier":"regioninfo","vlen":70,"tag":[],"timestamp":"1733101915508"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101915508"},{"qualifier":"seqnumDuringOpen","vlen":8,"tag":[],"timestamp":"1733101915508"}]},"ts":"1733101915508"} 2024-12-02T01:11:55,534 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=32775 {}] regionserver.HRegion(8581): Flush requested on 1588230740 2024-12-02T01:11:55,535 DEBUG [MemStoreFlusher.0 {}] regionserver.FlushAllLargeStoresPolicy(69): Since none of the CFs were above the size, flushing all. 2024-12-02T01:11:55,535 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=4.75 KB heapSize=8.29 KB 2024-12-02T01:11:55,539 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=38ed7bb40b7df5cbab5b70514657b066, ASSIGN}, {pid=16, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=761c6c5fe18cb86212f23a9aac9b6e3c, ASSIGN}] 2024-12-02T01:11:55,540 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=15, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=38ed7bb40b7df5cbab5b70514657b066, ASSIGN 2024-12-02T01:11:55,540 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=16, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=761c6c5fe18cb86212f23a9aac9b6e3c, ASSIGN 2024-12-02T01:11:55,541 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=15, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=38ed7bb40b7df5cbab5b70514657b066, ASSIGN; state=SPLITTING_NEW, location=a82f277fe18c,32775,1733101889168; forceNewPlan=false, retain=false 2024-12-02T01:11:55,541 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=16, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=761c6c5fe18cb86212f23a9aac9b6e3c, ASSIGN; state=SPLITTING_NEW, location=a82f277fe18c,32775,1733101889168; forceNewPlan=false, retain=false 2024-12-02T01:11:55,551 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/info/9ae43222baf44e22b64df8a41361086b is 193, key is TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c./info:regioninfo/1733101915508/Put/seqid=0 2024-12-02T01:11:55,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741851_1027 (size=9423) 2024-12-02T01:11:55,556 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741851_1027 (size=9423) 2024-12-02T01:11:55,556 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=4.54 KB at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/info/9ae43222baf44e22b64df8a41361086b 2024-12-02T01:11:55,580 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/table/55e472cd24034bdfbba4670ecd167867 is 65, key is TestLogRolling-testLogRolling/table:state/1733101891288/Put/seqid=0 2024-12-02T01:11:55,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741852_1028 (size=5412) 2024-12-02T01:11:55,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741852_1028 (size=5412) 2024-12-02T01:11:55,585 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=216 B at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/table/55e472cd24034bdfbba4670ecd167867 2024-12-02T01:11:55,592 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/info/9ae43222baf44e22b64df8a41361086b as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/info/9ae43222baf44e22b64df8a41361086b 2024-12-02T01:11:55,597 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/info/9ae43222baf44e22b64df8a41361086b, entries=29, sequenceid=17, filesize=9.2 K 2024-12-02T01:11:55,597 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/table/55e472cd24034bdfbba4670ecd167867 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/table/55e472cd24034bdfbba4670ecd167867 2024-12-02T01:11:55,602 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/table/55e472cd24034bdfbba4670ecd167867, entries=4, sequenceid=17, filesize=5.3 K 2024-12-02T01:11:55,603 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~4.75 KB/4869, heapSize ~8.01 KB/8200, currentSize=0 B/0 for 1588230740 in 68ms, sequenceid=17, compaction requested=false 2024-12-02T01:11:55,603 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 1588230740: 2024-12-02T01:11:55,691 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=15 updating hbase:meta row=38ed7bb40b7df5cbab5b70514657b066, regionState=OPENING, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,691 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=16 updating hbase:meta row=761c6c5fe18cb86212f23a9aac9b6e3c, regionState=OPENING, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,694 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=17, ppid=16, state=RUNNABLE; OpenRegionProcedure 761c6c5fe18cb86212f23a9aac9b6e3c, server=a82f277fe18c,32775,1733101889168}] 2024-12-02T01:11:55,696 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=18, ppid=15, state=RUNNABLE; OpenRegionProcedure 38ed7bb40b7df5cbab5b70514657b066, server=a82f277fe18c,32775,1733101889168}] 2024-12-02T01:11:55,848 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,851 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:11:55,851 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7285): Opening region: {ENCODED => 38ed7bb40b7df5cbab5b70514657b066, NAME => 'TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.', STARTKEY => '', ENDKEY => 'row0062'} 2024-12-02T01:11:55,851 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling 38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,851 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:55,851 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7327): checking encryption for 38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,851 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7330): checking classloading for 38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,853 INFO [StoreOpener-38ed7bb40b7df5cbab5b70514657b066-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,854 INFO [StoreOpener-38ed7bb40b7df5cbab5b70514657b066-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 38ed7bb40b7df5cbab5b70514657b066 columnFamilyName info 2024-12-02T01:11:55,854 DEBUG [StoreOpener-38ed7bb40b7df5cbab5b70514657b066-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:55,864 DEBUG [StoreOpener-38ed7bb40b7df5cbab5b70514657b066-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994->hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb-bottom 2024-12-02T01:11:55,864 INFO [StoreOpener-38ed7bb40b7df5cbab5b70514657b066-1 {}] regionserver.HStore(327): Store=38ed7bb40b7df5cbab5b70514657b066/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:55,865 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,867 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,869 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1085): writing seq id for 38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:11:55,870 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1102): Opened 38ed7bb40b7df5cbab5b70514657b066; next sequenceid=108; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=754549, jitterRate=-0.04054178297519684}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:11:55,870 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1001): Region open journal for 38ed7bb40b7df5cbab5b70514657b066: 2024-12-02T01:11:55,871 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066., pid=18, masterSystemTime=1733101915848 2024-12-02T01:11:55,871 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.CompactSplit(403): Add compact mark for store 38ed7bb40b7df5cbab5b70514657b066:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:11:55,871 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: Opening Region; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,871 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 1 store files, 0 compacting, 1 eligible, 16 blocking 2024-12-02T01:11:55,872 INFO [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HStore(1526): Keeping/Overriding Compaction request priority to -2147482648 for CF info since it belongs to recently split daughter region TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:11:55,872 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HStore(1540): 38ed7bb40b7df5cbab5b70514657b066/info is initiating minor compaction (all files) 2024-12-02T01:11:55,872 INFO [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 38ed7bb40b7df5cbab5b70514657b066/info in TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:11:55,873 INFO [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994->hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb-bottom] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/.tmp, totalSize=73.6 K 2024-12-02T01:11:55,873 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:11:55,873 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:11:55,873 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:11:55,873 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(7285): Opening region: {ENCODED => 761c6c5fe18cb86212f23a9aac9b6e3c, NAME => 'TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.', STARTKEY => 'row0062', ENDKEY => ''} 2024-12-02T01:11:55,873 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] compactions.Compactor(224): Compacting a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994, keycount=32, bloomtype=ROW, size=73.6 K, encoding=NONE, compression=NONE, seqNum=82, earliestPutTs=1733101900933 2024-12-02T01:11:55,873 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=15 updating hbase:meta row=38ed7bb40b7df5cbab5b70514657b066, regionState=OPEN, openSeqNum=108, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,874 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,874 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:11:55,874 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(7327): checking encryption for 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,874 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(7330): checking classloading for 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,875 INFO [StoreOpener-761c6c5fe18cb86212f23a9aac9b6e3c-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,876 INFO [StoreOpener-761c6c5fe18cb86212f23a9aac9b6e3c-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 761c6c5fe18cb86212f23a9aac9b6e3c columnFamilyName info 2024-12-02T01:11:55,876 DEBUG [StoreOpener-761c6c5fe18cb86212f23a9aac9b6e3c-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:11:55,877 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=18, resume processing ppid=15 2024-12-02T01:11:55,877 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, ppid=15, state=SUCCESS; OpenRegionProcedure 38ed7bb40b7df5cbab5b70514657b066, server=a82f277fe18c,32775,1733101889168 in 179 msec 2024-12-02T01:11:55,878 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=38ed7bb40b7df5cbab5b70514657b066, ASSIGN in 338 msec 2024-12-02T01:11:55,884 DEBUG [StoreOpener-761c6c5fe18cb86212f23a9aac9b6e3c-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc 2024-12-02T01:11:55,889 DEBUG [StoreOpener-761c6c5fe18cb86212f23a9aac9b6e3c-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994->hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb-top 2024-12-02T01:11:55,889 INFO [StoreOpener-761c6c5fe18cb86212f23a9aac9b6e3c-1 {}] regionserver.HStore(327): Store=761c6c5fe18cb86212f23a9aac9b6e3c/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:11:55,890 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,891 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,891 INFO [RS:0;a82f277fe18c:32775-longCompactions-0 {}] throttle.PressureAwareThroughputController(145): 38ed7bb40b7df5cbab5b70514657b066#info#compaction#51 average throughput is 31.30 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:11:55,892 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/.tmp/info/d33884626e454a97866cbbe68bf3fa2e is 1080, key is row0001/info:/1733101900933/Put/seqid=0 2024-12-02T01:11:55,894 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(1085): writing seq id for 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:11:55,894 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(1102): Opened 761c6c5fe18cb86212f23a9aac9b6e3c; next sequenceid=108; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=714013, jitterRate=-0.09208577871322632}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:11:55,895 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(1001): Region open journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:11:55,895 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., pid=17, masterSystemTime=1733101915848 2024-12-02T01:11:55,895 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 2 store files, 0 compacting, 2 eligible, 16 blocking 2024-12-02T01:11:55,896 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 2 2024-12-02T01:11:55,896 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: Opening Region; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,896 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1526): Keeping/Overriding Compaction request priority to -2147482648 for CF info since it belongs to recently split daughter region TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:11:55,896 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:11:55,896 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:11:55,897 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994->hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb-top, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=97.4 K 2024-12-02T01:11:55,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741853_1029 (size=70862) 2024-12-02T01:11:55,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741853_1029 (size=70862) 2024-12-02T01:11:55,898 DEBUG [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:11:55,898 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994, keycount=32, bloomtype=ROW, size=73.6 K, encoding=NONE, compression=NONE, seqNum=83, earliestPutTs=1733101900933 2024-12-02T01:11:55,898 INFO [RS_OPEN_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:11:55,898 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc, keycount=18, bloomtype=ROW, size=23.8 K, encoding=NONE, compression=NONE, seqNum=104, earliestPutTs=1733101915129 2024-12-02T01:11:55,899 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=16 updating hbase:meta row=761c6c5fe18cb86212f23a9aac9b6e3c, regionState=OPEN, openSeqNum=108, regionLocation=a82f277fe18c,32775,1733101889168 2024-12-02T01:11:55,902 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=17, resume processing ppid=16 2024-12-02T01:11:55,903 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, ppid=16, state=SUCCESS; OpenRegionProcedure 761c6c5fe18cb86212f23a9aac9b6e3c, server=a82f277fe18c,32775,1733101889168 in 206 msec 2024-12-02T01:11:55,904 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/.tmp/info/d33884626e454a97866cbbe68bf3fa2e as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/d33884626e454a97866cbbe68bf3fa2e 2024-12-02T01:11:55,904 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=16, resume processing ppid=12 2024-12-02T01:11:55,904 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=761c6c5fe18cb86212f23a9aac9b6e3c, ASSIGN in 363 msec 2024-12-02T01:11:55,906 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=fbd98b96465e3cae82d3686045e0a994, daughterA=38ed7bb40b7df5cbab5b70514657b066, daughterB=761c6c5fe18cb86212f23a9aac9b6e3c in 688 msec 2024-12-02T01:11:55,917 INFO [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 1 (all) file(s) in 38ed7bb40b7df5cbab5b70514657b066/info of 38ed7bb40b7df5cbab5b70514657b066 into d33884626e454a97866cbbe68bf3fa2e(size=69.2 K), total size for store is 69.2 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:11:55,917 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 38ed7bb40b7df5cbab5b70514657b066: 2024-12-02T01:11:55,917 INFO [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066., storeName=38ed7bb40b7df5cbab5b70514657b066/info, priority=15, startTime=1733101915871; duration=0sec 2024-12-02T01:11:55,917 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,918 DEBUG [RS:0;a82f277fe18c:32775-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 38ed7bb40b7df5cbab5b70514657b066:info 2024-12-02T01:11:55,921 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#52 average throughput is 22.58 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:11:55,921 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/e844981557294726a8242e7962752dc5 is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:11:55,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741854_1030 (size=28850) 2024-12-02T01:11:55,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741854_1030 (size=28850) 2024-12-02T01:11:55,936 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/e844981557294726a8242e7962752dc5 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e844981557294726a8242e7962752dc5 2024-12-02T01:11:55,942 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 2 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into e844981557294726a8242e7962752dc5(size=28.2 K), total size for store is 28.2 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:11:55,942 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:11:55,942 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=14, startTime=1733101915895; duration=0sec 2024-12-02T01:11:55,942 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:11:55,942 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:11:56,445 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:57,447 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:58,447 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:11:58,994 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:11:59,448 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:00,424 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,425 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,425 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,425 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,425 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,425 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,438 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,438 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,438 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,438 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,438 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,438 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,440 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,441 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,441 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,442 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,449 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:00,521 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:12:00,522 INFO [RS-EventLoopGroup-12-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48976, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:12:00,951 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-02T01:12:00,954 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,955 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,955 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,956 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,957 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,958 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,974 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,974 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,974 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,975 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,975 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,975 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,977 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,977 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,977 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:00,979 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-02T01:12:01,450 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:02,450 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:03,451 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:04,453 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:05,217 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] ipc.CallRunner(138): callId: 96 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:38866 deadline: 1733101935217, exception=org.apache.hadoop.hbase.NotServingRegionException: TestLogRolling-testLogRolling,,1733101890920.fbd98b96465e3cae82d3686045e0a994. is not online on a82f277fe18c,32775,1733101889168 2024-12-02T01:12:05,454 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:06,454 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:07,455 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:08,455 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:09,456 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:10,456 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:11,206 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-02T01:12:11,206 INFO [master/a82f277fe18c:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-02T01:12:11,457 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:12,458 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:13,458 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:14,459 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:15,125 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 1588230740, had cached 0 bytes from a total of 14835 2024-12-02T01:12:15,460 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:16,460 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:17,461 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:18,461 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:19,462 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:20,056 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 61b9a0fcb731850a98b7b912948ae7ef, had cached 0 bytes from a total of 23930 2024-12-02T01:12:20,463 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:21,463 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:22,464 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:23,465 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:24,466 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:25,363 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:25,363 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:12:25,369 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/aff4712981754a02b4cf4efee2d5caeb is 1080, key is row0084/info:/1733101945353/Put/seqid=0 2024-12-02T01:12:25,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741855_1031 (size=12509) 2024-12-02T01:12:25,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741855_1031 (size=12509) 2024-12-02T01:12:25,376 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=118 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/aff4712981754a02b4cf4efee2d5caeb 2024-12-02T01:12:25,381 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/aff4712981754a02b4cf4efee2d5caeb as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aff4712981754a02b4cf4efee2d5caeb 2024-12-02T01:12:25,385 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aff4712981754a02b4cf4efee2d5caeb, entries=7, sequenceid=118, filesize=12.2 K 2024-12-02T01:12:25,386 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=6.30 KB/6456 for 761c6c5fe18cb86212f23a9aac9b6e3c in 23ms, sequenceid=118, compaction requested=false 2024-12-02T01:12:25,386 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:25,467 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:26,467 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:27,376 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:27,376 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:12:27,387 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/f37a1ca5850a44daa4f4930ef71dc91e is 1080, key is row0091/info:/1733101945364/Put/seqid=0 2024-12-02T01:12:27,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741856_1032 (size=12509) 2024-12-02T01:12:27,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741856_1032 (size=12509) 2024-12-02T01:12:27,392 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=128 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/f37a1ca5850a44daa4f4930ef71dc91e 2024-12-02T01:12:27,398 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/f37a1ca5850a44daa4f4930ef71dc91e as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/f37a1ca5850a44daa4f4930ef71dc91e 2024-12-02T01:12:27,403 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/f37a1ca5850a44daa4f4930ef71dc91e, entries=7, sequenceid=128, filesize=12.2 K 2024-12-02T01:12:27,404 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=17.86 KB/18292 for 761c6c5fe18cb86212f23a9aac9b6e3c in 28ms, sequenceid=128, compaction requested=true 2024-12-02T01:12:27,404 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:27,404 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:12:27,404 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:27,404 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:12:27,405 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:27,405 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=18.91 KB heapSize=20.50 KB 2024-12-02T01:12:27,405 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 53868 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:12:27,405 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:12:27,405 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:27,405 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e844981557294726a8242e7962752dc5, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aff4712981754a02b4cf4efee2d5caeb, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/f37a1ca5850a44daa4f4930ef71dc91e] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=52.6 K 2024-12-02T01:12:27,406 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting e844981557294726a8242e7962752dc5, keycount=22, bloomtype=ROW, size=28.2 K, encoding=NONE, compression=NONE, seqNum=104, earliestPutTs=1733101913120 2024-12-02T01:12:27,406 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting aff4712981754a02b4cf4efee2d5caeb, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=118, earliestPutTs=1733101945353 2024-12-02T01:12:27,407 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting f37a1ca5850a44daa4f4930ef71dc91e, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=128, earliestPutTs=1733101945364 2024-12-02T01:12:27,409 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/a934843dfa184e179f6f050d01ee2803 is 1080, key is row0098/info:/1733101947378/Put/seqid=0 2024-12-02T01:12:27,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741857_1033 (size=24394) 2024-12-02T01:12:27,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741857_1033 (size=24394) 2024-12-02T01:12:27,416 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=18.91 KB at sequenceid=149 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/a934843dfa184e179f6f050d01ee2803 2024-12-02T01:12:27,422 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#56 average throughput is 36.94 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:12:27,422 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=761c6c5fe18cb86212f23a9aac9b6e3c, server=a82f277fe18c,32775,1733101889168 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-02T01:12:27,422 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] ipc.CallRunner(138): callId: 143 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:38866 deadline: 1733101957422, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=761c6c5fe18cb86212f23a9aac9b6e3c, server=a82f277fe18c,32775,1733101889168 2024-12-02T01:12:27,422 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/e607caa580a4496d8161861bfaaffb8d is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:12:27,422 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/a934843dfa184e179f6f050d01ee2803 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a934843dfa184e179f6f050d01ee2803 2024-12-02T01:12:27,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741858_1034 (size=44066) 2024-12-02T01:12:27,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741858_1034 (size=44066) 2024-12-02T01:12:27,428 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a934843dfa184e179f6f050d01ee2803, entries=18, sequenceid=149, filesize=23.8 K 2024-12-02T01:12:27,428 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~18.91 KB/19368, heapSize ~20.48 KB/20976, currentSize=11.56 KB/11836 for 761c6c5fe18cb86212f23a9aac9b6e3c in 23ms, sequenceid=149, compaction requested=false 2024-12-02T01:12:27,429 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:27,431 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/e607caa580a4496d8161861bfaaffb8d as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e607caa580a4496d8161861bfaaffb8d 2024-12-02T01:12:27,436 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into e607caa580a4496d8161861bfaaffb8d(size=43.0 K), total size for store is 66.9 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:12:27,436 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:27,436 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=13, startTime=1733101947404; duration=0sec 2024-12-02T01:12:27,436 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:27,436 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:12:27,468 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:28,469 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:28,994 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-02T01:12:29,470 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:30,470 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:30,780 DEBUG [master/a82f277fe18c:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 5acee627f693bbca053dc704b61b658f changed from -1.0 to 0.0, refreshing cache 2024-12-02T01:12:31,471 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:32,472 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:33,473 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:34,474 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:35,475 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:36,476 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:37,445 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:37,445 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=12.61 KB heapSize=13.75 KB 2024-12-02T01:12:37,450 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/23fad1102d0845058ebfb9473fc33a63 is 1080, key is row0116/info:/1733101947405/Put/seqid=0 2024-12-02T01:12:37,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741859_1035 (size=17906) 2024-12-02T01:12:37,456 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741859_1035 (size=17906) 2024-12-02T01:12:37,457 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=12.61 KB at sequenceid=165 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/23fad1102d0845058ebfb9473fc33a63 2024-12-02T01:12:37,465 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/23fad1102d0845058ebfb9473fc33a63 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/23fad1102d0845058ebfb9473fc33a63 2024-12-02T01:12:37,469 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/23fad1102d0845058ebfb9473fc33a63, entries=12, sequenceid=165, filesize=17.5 K 2024-12-02T01:12:37,470 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~12.61 KB/12912, heapSize ~13.73 KB/14064, currentSize=1.05 KB/1076 for 761c6c5fe18cb86212f23a9aac9b6e3c in 25ms, sequenceid=165, compaction requested=true 2024-12-02T01:12:37,470 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:37,470 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:12:37,470 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:37,470 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:12:37,471 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 86366 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:12:37,471 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:12:37,471 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:37,471 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e607caa580a4496d8161861bfaaffb8d, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a934843dfa184e179f6f050d01ee2803, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/23fad1102d0845058ebfb9473fc33a63] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=84.3 K 2024-12-02T01:12:37,472 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting e607caa580a4496d8161861bfaaffb8d, keycount=36, bloomtype=ROW, size=43.0 K, encoding=NONE, compression=NONE, seqNum=128, earliestPutTs=1733101913120 2024-12-02T01:12:37,472 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting a934843dfa184e179f6f050d01ee2803, keycount=18, bloomtype=ROW, size=23.8 K, encoding=NONE, compression=NONE, seqNum=149, earliestPutTs=1733101947378 2024-12-02T01:12:37,473 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 23fad1102d0845058ebfb9473fc33a63, keycount=12, bloomtype=ROW, size=17.5 K, encoding=NONE, compression=NONE, seqNum=165, earliestPutTs=1733101947405 2024-12-02T01:12:37,476 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:37,483 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#58 average throughput is 67.73 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:12:37,484 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/fdaccb3c10804953a4bb57921b4d9e3a is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:12:37,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741860_1036 (size=76649) 2024-12-02T01:12:37,489 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741860_1036 (size=76649) 2024-12-02T01:12:37,494 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/fdaccb3c10804953a4bb57921b4d9e3a as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/fdaccb3c10804953a4bb57921b4d9e3a 2024-12-02T01:12:37,499 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into fdaccb3c10804953a4bb57921b4d9e3a(size=74.9 K), total size for store is 74.9 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:12:37,499 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:37,500 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=13, startTime=1733101957470; duration=0sec 2024-12-02T01:12:37,500 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:37,500 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:12:38,477 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:38,477 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=3 on file=hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta after 196172ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor203.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-02T01:12:39,461 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:39,461 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:12:39,466 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/a7fa121e4f9a41d8b8140389cfc11d01 is 1080, key is row0128/info:/1733101957446/Put/seqid=0 2024-12-02T01:12:39,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741861_1037 (size=12516) 2024-12-02T01:12:39,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741861_1037 (size=12516) 2024-12-02T01:12:39,473 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=176 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/a7fa121e4f9a41d8b8140389cfc11d01 2024-12-02T01:12:39,478 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:39,480 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/a7fa121e4f9a41d8b8140389cfc11d01 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a7fa121e4f9a41d8b8140389cfc11d01 2024-12-02T01:12:39,486 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a7fa121e4f9a41d8b8140389cfc11d01, entries=7, sequenceid=176, filesize=12.2 K 2024-12-02T01:12:39,487 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=19.96 KB/20444 for 761c6c5fe18cb86212f23a9aac9b6e3c in 26ms, sequenceid=176, compaction requested=false 2024-12-02T01:12:39,487 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:39,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:39,488 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=21.02 KB heapSize=22.75 KB 2024-12-02T01:12:39,491 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f is 1080, key is row0135/info:/1733101959461/Put/seqid=0 2024-12-02T01:12:39,496 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741862_1038 (size=26550) 2024-12-02T01:12:39,496 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741862_1038 (size=26550) 2024-12-02T01:12:39,496 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=21.02 KB at sequenceid=199 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f 2024-12-02T01:12:39,501 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f 2024-12-02T01:12:39,506 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f, entries=20, sequenceid=199, filesize=25.9 K 2024-12-02T01:12:39,507 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~21.02 KB/21520, heapSize ~22.73 KB/23280, currentSize=6.30 KB/6456 for 761c6c5fe18cb86212f23a9aac9b6e3c in 19ms, sequenceid=199, compaction requested=true 2024-12-02T01:12:39,507 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:39,507 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:12:39,507 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:39,507 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:12:39,509 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 115715 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:12:39,509 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:12:39,509 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:39,509 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/fdaccb3c10804953a4bb57921b4d9e3a, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a7fa121e4f9a41d8b8140389cfc11d01, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=113.0 K 2024-12-02T01:12:39,509 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting fdaccb3c10804953a4bb57921b4d9e3a, keycount=66, bloomtype=ROW, size=74.9 K, encoding=NONE, compression=NONE, seqNum=165, earliestPutTs=1733101913120 2024-12-02T01:12:39,510 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting a7fa121e4f9a41d8b8140389cfc11d01, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=176, earliestPutTs=1733101957446 2024-12-02T01:12:39,510 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting d3e02b3a14ca4f0e9883cc1d4a3f5b8f, keycount=20, bloomtype=ROW, size=25.9 K, encoding=NONE, compression=NONE, seqNum=199, earliestPutTs=1733101959461 2024-12-02T01:12:39,522 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#61 average throughput is 47.72 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:12:39,523 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/8bdd39b29c3c4aa582b51bdf7fb55810 is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:12:39,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741863_1039 (size=105865) 2024-12-02T01:12:39,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741863_1039 (size=105865) 2024-12-02T01:12:39,532 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/8bdd39b29c3c4aa582b51bdf7fb55810 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8bdd39b29c3c4aa582b51bdf7fb55810 2024-12-02T01:12:39,537 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into 8bdd39b29c3c4aa582b51bdf7fb55810(size=103.4 K), total size for store is 103.4 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:12:39,537 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:39,537 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=13, startTime=1733101959507; duration=0sec 2024-12-02T01:12:39,537 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:39,538 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:12:40,478 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:40,852 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 38ed7bb40b7df5cbab5b70514657b066, had cached 0 bytes from a total of 70862 2024-12-02T01:12:40,874 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 761c6c5fe18cb86212f23a9aac9b6e3c, had cached 0 bytes from a total of 105865 2024-12-02T01:12:41,479 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:41,498 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:41,499 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:12:41,507 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/6936eaea5ec443f5adf74679d5a1e971 is 1080, key is row0155/info:/1733101959488/Put/seqid=0 2024-12-02T01:12:41,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741864_1040 (size=12516) 2024-12-02T01:12:41,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741864_1040 (size=12516) 2024-12-02T01:12:41,514 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=210 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/6936eaea5ec443f5adf74679d5a1e971 2024-12-02T01:12:41,519 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/6936eaea5ec443f5adf74679d5a1e971 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/6936eaea5ec443f5adf74679d5a1e971 2024-12-02T01:12:41,525 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/6936eaea5ec443f5adf74679d5a1e971, entries=7, sequenceid=210, filesize=12.2 K 2024-12-02T01:12:41,525 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=15.76 KB/16140 for 761c6c5fe18cb86212f23a9aac9b6e3c in 27ms, sequenceid=210, compaction requested=false 2024-12-02T01:12:41,525 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:41,526 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:41,527 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=17.86 KB heapSize=19.38 KB 2024-12-02T01:12:41,530 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/c6725f250b5a4f5eb3b2500c9ffd749e is 1080, key is row0162/info:/1733101961500/Put/seqid=0 2024-12-02T01:12:41,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741865_1041 (size=23316) 2024-12-02T01:12:41,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741865_1041 (size=23316) 2024-12-02T01:12:41,536 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=17.86 KB at sequenceid=230 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/c6725f250b5a4f5eb3b2500c9ffd749e 2024-12-02T01:12:41,541 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/c6725f250b5a4f5eb3b2500c9ffd749e as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/c6725f250b5a4f5eb3b2500c9ffd749e 2024-12-02T01:12:41,543 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=761c6c5fe18cb86212f23a9aac9b6e3c, server=a82f277fe18c,32775,1733101889168 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-02T01:12:41,543 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] ipc.CallRunner(138): callId: 209 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:38866 deadline: 1733101971543, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=761c6c5fe18cb86212f23a9aac9b6e3c, server=a82f277fe18c,32775,1733101889168 2024-12-02T01:12:41,546 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/c6725f250b5a4f5eb3b2500c9ffd749e, entries=17, sequenceid=230, filesize=22.8 K 2024-12-02T01:12:41,547 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~17.86 KB/18292, heapSize ~19.36 KB/19824, currentSize=12.61 KB/12912 for 761c6c5fe18cb86212f23a9aac9b6e3c in 19ms, sequenceid=230, compaction requested=true 2024-12-02T01:12:41,547 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:41,547 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:12:41,547 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:41,547 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:12:41,548 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 141697 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:12:41,548 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:12:41,548 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:41,548 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8bdd39b29c3c4aa582b51bdf7fb55810, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/6936eaea5ec443f5adf74679d5a1e971, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/c6725f250b5a4f5eb3b2500c9ffd749e] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=138.4 K 2024-12-02T01:12:41,548 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 8bdd39b29c3c4aa582b51bdf7fb55810, keycount=93, bloomtype=ROW, size=103.4 K, encoding=NONE, compression=NONE, seqNum=199, earliestPutTs=1733101913120 2024-12-02T01:12:41,548 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 6936eaea5ec443f5adf74679d5a1e971, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=210, earliestPutTs=1733101959488 2024-12-02T01:12:41,549 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting c6725f250b5a4f5eb3b2500c9ffd749e, keycount=17, bloomtype=ROW, size=22.8 K, encoding=NONE, compression=NONE, seqNum=230, earliestPutTs=1733101961500 2024-12-02T01:12:41,558 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#64 average throughput is 60.03 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:12:41,558 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/2edbbac3993c458a8ed13a716cd70982 is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:12:41,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741866_1042 (size=131971) 2024-12-02T01:12:41,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741866_1042 (size=131971) 2024-12-02T01:12:41,566 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/2edbbac3993c458a8ed13a716cd70982 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/2edbbac3993c458a8ed13a716cd70982 2024-12-02T01:12:41,571 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into 2edbbac3993c458a8ed13a716cd70982(size=128.9 K), total size for store is 128.9 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:12:41,571 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:41,571 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=13, startTime=1733101961547; duration=0sec 2024-12-02T01:12:41,571 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:41,571 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:12:42,480 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:43,482 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:44,483 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:45,484 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:46,485 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:47,485 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:48,486 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:49,487 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:50,488 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:51,489 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:51,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:51,592 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=13.66 KB heapSize=14.88 KB 2024-12-02T01:12:51,600 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/34987475de8f4b509b3631a5efedbd42 is 1080, key is row0179/info:/1733101961527/Put/seqid=0 2024-12-02T01:12:51,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741867_1043 (size=19000) 2024-12-02T01:12:51,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741867_1043 (size=19000) 2024-12-02T01:12:51,607 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=13.66 KB at sequenceid=247 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/34987475de8f4b509b3631a5efedbd42 2024-12-02T01:12:51,615 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/34987475de8f4b509b3631a5efedbd42 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/34987475de8f4b509b3631a5efedbd42 2024-12-02T01:12:51,622 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/34987475de8f4b509b3631a5efedbd42, entries=13, sequenceid=247, filesize=18.6 K 2024-12-02T01:12:51,623 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~13.66 KB/13988, heapSize ~14.86 KB/15216, currentSize=1.05 KB/1076 for 761c6c5fe18cb86212f23a9aac9b6e3c in 31ms, sequenceid=247, compaction requested=false 2024-12-02T01:12:51,623 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:52,490 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:53,491 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:53,611 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:53,611 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:12:53,615 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/45ddc6055e304539b5b6a7543415d739 is 1080, key is row0192/info:/1733101971593/Put/seqid=0 2024-12-02T01:12:53,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741868_1044 (size=12517) 2024-12-02T01:12:53,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741868_1044 (size=12517) 2024-12-02T01:12:53,624 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=257 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/45ddc6055e304539b5b6a7543415d739 2024-12-02T01:12:53,629 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/45ddc6055e304539b5b6a7543415d739 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/45ddc6055e304539b5b6a7543415d739 2024-12-02T01:12:53,634 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/45ddc6055e304539b5b6a7543415d739, entries=7, sequenceid=257, filesize=12.2 K 2024-12-02T01:12:53,635 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=21.02 KB/21520 for 761c6c5fe18cb86212f23a9aac9b6e3c in 23ms, sequenceid=257, compaction requested=true 2024-12-02T01:12:53,635 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:53,635 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:12:53,635 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:53,635 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:12:53,635 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:53,636 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=23.12 KB heapSize=25 KB 2024-12-02T01:12:53,636 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 163488 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:12:53,636 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:12:53,636 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:53,636 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/2edbbac3993c458a8ed13a716cd70982, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/34987475de8f4b509b3631a5efedbd42, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/45ddc6055e304539b5b6a7543415d739] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=159.7 K 2024-12-02T01:12:53,636 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 2edbbac3993c458a8ed13a716cd70982, keycount=117, bloomtype=ROW, size=128.9 K, encoding=NONE, compression=NONE, seqNum=230, earliestPutTs=1733101913120 2024-12-02T01:12:53,637 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 34987475de8f4b509b3631a5efedbd42, keycount=13, bloomtype=ROW, size=18.6 K, encoding=NONE, compression=NONE, seqNum=247, earliestPutTs=1733101961527 2024-12-02T01:12:53,637 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 45ddc6055e304539b5b6a7543415d739, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=257, earliestPutTs=1733101971593 2024-12-02T01:12:53,639 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/8332ea3c24c04740873f8bc1d805acba is 1080, key is row0199/info:/1733101973611/Put/seqid=0 2024-12-02T01:12:53,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741869_1045 (size=28728) 2024-12-02T01:12:53,646 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741869_1045 (size=28728) 2024-12-02T01:12:53,646 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=23.12 KB at sequenceid=282 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/8332ea3c24c04740873f8bc1d805acba 2024-12-02T01:12:53,650 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#68 average throughput is 46.86 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:12:53,651 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/7ca338eda0a1438fadabbba8dbc7def0 is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:12:53,652 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/8332ea3c24c04740873f8bc1d805acba as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8332ea3c24c04740873f8bc1d805acba 2024-12-02T01:12:53,654 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741870_1046 (size=153707) 2024-12-02T01:12:53,654 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741870_1046 (size=153707) 2024-12-02T01:12:53,657 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8332ea3c24c04740873f8bc1d805acba, entries=22, sequenceid=282, filesize=28.1 K 2024-12-02T01:12:53,658 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~23.12 KB/23672, heapSize ~24.98 KB/25584, currentSize=4.20 KB/4304 for 761c6c5fe18cb86212f23a9aac9b6e3c in 23ms, sequenceid=282, compaction requested=false 2024-12-02T01:12:53,658 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:53,659 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/7ca338eda0a1438fadabbba8dbc7def0 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/7ca338eda0a1438fadabbba8dbc7def0 2024-12-02T01:12:53,664 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into 7ca338eda0a1438fadabbba8dbc7def0(size=150.1 K), total size for store is 178.2 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:12:53,664 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:53,664 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=13, startTime=1733101973635; duration=0sec 2024-12-02T01:12:53,664 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:53,664 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:12:54,492 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:55,494 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:55,649 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:55,649 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-02T01:12:55,655 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/aeca30517bab4ff6aac97dbc50b09db7 is 1080, key is row0221/info:/1733101973636/Put/seqid=0 2024-12-02T01:12:55,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741871_1047 (size=12523) 2024-12-02T01:12:55,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741871_1047 (size=12523) 2024-12-02T01:12:55,663 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=293 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/aeca30517bab4ff6aac97dbc50b09db7 2024-12-02T01:12:55,669 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/aeca30517bab4ff6aac97dbc50b09db7 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aeca30517bab4ff6aac97dbc50b09db7 2024-12-02T01:12:55,674 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aeca30517bab4ff6aac97dbc50b09db7, entries=7, sequenceid=293, filesize=12.2 K 2024-12-02T01:12:55,675 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=19.96 KB/20444 for 761c6c5fe18cb86212f23a9aac9b6e3c in 26ms, sequenceid=293, compaction requested=true 2024-12-02T01:12:55,675 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:55,675 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 761c6c5fe18cb86212f23a9aac9b6e3c:info, priority=-2147483648, current under compaction store size is 1 2024-12-02T01:12:55,675 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:55,675 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-02T01:12:55,676 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32775 {}] regionserver.HRegion(8581): Flush requested on 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:55,676 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 194958 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-02T01:12:55,676 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1540): 761c6c5fe18cb86212f23a9aac9b6e3c/info is initiating minor compaction (all files) 2024-12-02T01:12:55,676 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=22.07 KB heapSize=23.88 KB 2024-12-02T01:12:55,676 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 761c6c5fe18cb86212f23a9aac9b6e3c/info in TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:55,677 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/7ca338eda0a1438fadabbba8dbc7def0, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8332ea3c24c04740873f8bc1d805acba, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aeca30517bab4ff6aac97dbc50b09db7] into tmpdir=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp, totalSize=190.4 K 2024-12-02T01:12:55,677 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 7ca338eda0a1438fadabbba8dbc7def0, keycount=137, bloomtype=ROW, size=150.1 K, encoding=NONE, compression=NONE, seqNum=257, earliestPutTs=1733101913120 2024-12-02T01:12:55,677 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting 8332ea3c24c04740873f8bc1d805acba, keycount=22, bloomtype=ROW, size=28.1 K, encoding=NONE, compression=NONE, seqNum=282, earliestPutTs=1733101973611 2024-12-02T01:12:55,678 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] compactions.Compactor(224): Compacting aeca30517bab4ff6aac97dbc50b09db7, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=293, earliestPutTs=1733101973636 2024-12-02T01:12:55,680 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/1eccaf07c71e4b12940613e71c6febfe is 1080, key is row0228/info:/1733101975650/Put/seqid=0 2024-12-02T01:12:55,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741872_1048 (size=27649) 2024-12-02T01:12:55,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741872_1048 (size=27649) 2024-12-02T01:12:55,686 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=22.07 KB at sequenceid=317 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/1eccaf07c71e4b12940613e71c6febfe 2024-12-02T01:12:55,691 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 761c6c5fe18cb86212f23a9aac9b6e3c#info#compaction#71 average throughput is 56.78 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-02T01:12:55,692 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/d119fc803a514202a50bdf3e2e57fa43 is 1080, key is row0062/info:/1733101913120/Put/seqid=0 2024-12-02T01:12:55,692 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/1eccaf07c71e4b12940613e71c6febfe as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/1eccaf07c71e4b12940613e71c6febfe 2024-12-02T01:12:55,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741873_1049 (size=185108) 2024-12-02T01:12:55,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741873_1049 (size=185108) 2024-12-02T01:12:55,697 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/1eccaf07c71e4b12940613e71c6febfe, entries=21, sequenceid=317, filesize=27.0 K 2024-12-02T01:12:55,698 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~22.07 KB/22596, heapSize ~23.86 KB/24432, currentSize=8.41 KB/8608 for 761c6c5fe18cb86212f23a9aac9b6e3c in 22ms, sequenceid=317, compaction requested=false 2024-12-02T01:12:55,698 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:55,701 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/d119fc803a514202a50bdf3e2e57fa43 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d119fc803a514202a50bdf3e2e57fa43 2024-12-02T01:12:55,707 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 761c6c5fe18cb86212f23a9aac9b6e3c/info of 761c6c5fe18cb86212f23a9aac9b6e3c into d119fc803a514202a50bdf3e2e57fa43(size=180.8 K), total size for store is 207.8 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-02T01:12:55,707 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:55,707 INFO [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., storeName=761c6c5fe18cb86212f23a9aac9b6e3c/info, priority=13, startTime=1733101975675; duration=0sec 2024-12-02T01:12:55,708 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-02T01:12:55,708 DEBUG [RS:0;a82f277fe18c:32775-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 761c6c5fe18cb86212f23a9aac9b6e3c:info 2024-12-02T01:12:56,495 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:57,496 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:57,689 INFO [Time-limited test {}] wal.AbstractTestLogRolling(285): after writing there are 0 log files 2024-12-02T01:12:57,689 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C32775%2C1733101889168.1733101977689 2024-12-02T01:12:57,703 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101889674 with entries=311, filesize=307.43 KB; new WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101977689 2024-12-02T01:12:57,703 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45051:45051),(127.0.0.1/127.0.0.1:34647:34647)] 2024-12-02T01:12:57,704 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101889674 is not closed yet, will try archiving it next time 2024-12-02T01:12:57,706 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741833_1009 (size=314812) 2024-12-02T01:12:57,706 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741833_1009 (size=314812) 2024-12-02T01:12:57,708 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 5acee627f693bbca053dc704b61b658f 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-02T01:12:57,722 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/.tmp/info/799f4a3bfadc4fa4b821192027c76068 is 45, key is default/info:d/1733101890721/Put/seqid=0 2024-12-02T01:12:57,725 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741875_1051 (size=5037) 2024-12-02T01:12:57,726 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741875_1051 (size=5037) 2024-12-02T01:12:57,726 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/.tmp/info/799f4a3bfadc4fa4b821192027c76068 2024-12-02T01:12:57,730 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/.tmp/info/799f4a3bfadc4fa4b821192027c76068 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/info/799f4a3bfadc4fa4b821192027c76068 2024-12-02T01:12:57,735 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/info/799f4a3bfadc4fa4b821192027c76068, entries=2, sequenceid=6, filesize=4.9 K 2024-12-02T01:12:57,736 INFO [Time-limited test {}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 5acee627f693bbca053dc704b61b658f in 28ms, sequenceid=6, compaction requested=false 2024-12-02T01:12:57,736 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 5acee627f693bbca053dc704b61b658f: 2024-12-02T01:12:57,737 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 761c6c5fe18cb86212f23a9aac9b6e3c 1/1 column families, dataSize=8.41 KB heapSize=9.25 KB 2024-12-02T01:12:57,741 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/3f803ddf33e2437ebf6bfa6ecbc1aca2 is 1080, key is row0249/info:/1733101975677/Put/seqid=0 2024-12-02T01:12:57,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741876_1052 (size=13602) 2024-12-02T01:12:57,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741876_1052 (size=13602) 2024-12-02T01:12:57,746 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=8.41 KB at sequenceid=329 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/3f803ddf33e2437ebf6bfa6ecbc1aca2 2024-12-02T01:12:57,750 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/.tmp/info/3f803ddf33e2437ebf6bfa6ecbc1aca2 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/3f803ddf33e2437ebf6bfa6ecbc1aca2 2024-12-02T01:12:57,754 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/3f803ddf33e2437ebf6bfa6ecbc1aca2, entries=8, sequenceid=329, filesize=13.3 K 2024-12-02T01:12:57,755 INFO [Time-limited test {}] regionserver.HRegion(3040): Finished flush of dataSize ~8.41 KB/8608, heapSize ~9.23 KB/9456, currentSize=0 B/0 for 761c6c5fe18cb86212f23a9aac9b6e3c in 19ms, sequenceid=329, compaction requested=true 2024-12-02T01:12:57,755 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:57,755 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.21 KB heapSize=4.13 KB 2024-12-02T01:12:57,758 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/info/e9528c2dc22944d2bd2edcacab4d047f is 193, key is TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c./info:regioninfo/1733101915898/Put/seqid=0 2024-12-02T01:12:57,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741877_1053 (size=7803) 2024-12-02T01:12:57,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741877_1053 (size=7803) 2024-12-02T01:12:57,763 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.21 KB at sequenceid=24 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/info/e9528c2dc22944d2bd2edcacab4d047f 2024-12-02T01:12:57,768 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/.tmp/info/e9528c2dc22944d2bd2edcacab4d047f as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/info/e9528c2dc22944d2bd2edcacab4d047f 2024-12-02T01:12:57,773 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/info/e9528c2dc22944d2bd2edcacab4d047f, entries=16, sequenceid=24, filesize=7.6 K 2024-12-02T01:12:57,774 INFO [Time-limited test {}] regionserver.HRegion(3040): Finished flush of dataSize ~2.21 KB/2260, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 1588230740 in 18ms, sequenceid=24, compaction requested=false 2024-12-02T01:12:57,774 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 1588230740: 2024-12-02T01:12:57,774 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 38ed7bb40b7df5cbab5b70514657b066: 2024-12-02T01:12:57,774 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C32775%2C1733101889168.1733101977774 2024-12-02T01:12:57,780 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101977689 with entries=4, filesize=1.22 KB; new WAL /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101977774 2024-12-02T01:12:57,780 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34647:34647),(127.0.0.1/127.0.0.1:45051:45051)] 2024-12-02T01:12:57,780 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101977689 is not closed yet, will try archiving it next time 2024-12-02T01:12:57,781 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101889674 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/oldWALs/a82f277fe18c%2C32775%2C1733101889168.1733101889674 2024-12-02T01:12:57,781 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [5,000] milli-secs(wait.for.ratio=[1]) 2024-12-02T01:12:57,782 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741874_1050 (size=1255) 2024-12-02T01:12:57,782 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741874_1050 (size=1255) 2024-12-02T01:12:57,782 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168/a82f277fe18c%2C32775%2C1733101889168.1733101977689 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/oldWALs/a82f277fe18c%2C32775%2C1733101889168.1733101977689 2024-12-02T01:12:57,882 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-02T01:12:57,882 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-02T01:12:57,882 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x78fd5869 to 127.0.0.1:64314 2024-12-02T01:12:57,882 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:12:57,882 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T01:12:57,882 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1182394913, stopped=false 2024-12-02T01:12:57,883 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=a82f277fe18c,35029,1733101889015 2024-12-02T01:12:58,000 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:12:58,000 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:12:58,000 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:12:58,000 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:12:58,000 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-02T01:12:58,001 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:12:58,001 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,32775,1733101889168' ***** 2024-12-02T01:12:58,001 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:12:58,002 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:12:58,002 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:12:58,002 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:12:58,003 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:12:58,003 INFO [RS:0;a82f277fe18c:32775 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:12:58,003 INFO [RS:0;a82f277fe18c:32775 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:12:58,003 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(3579): Received CLOSE for 5acee627f693bbca053dc704b61b658f 2024-12-02T01:12:58,003 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(3579): Received CLOSE for 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(3579): Received CLOSE for 38ed7bb40b7df5cbab5b70514657b066 2024-12-02T01:12:58,004 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 5acee627f693bbca053dc704b61b658f, disabling compactions & flushes 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,32775,1733101889168 2024-12-02T01:12:58,004 DEBUG [RS:0;a82f277fe18c:32775 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:12:58,004 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:12:58,004 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:12:58,004 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. after waiting 0 ms 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:12:58,004 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:12:58,004 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1599): Waiting on 4 regions to close 2024-12-02T01:12:58,004 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1603): Online Regions={5acee627f693bbca053dc704b61b658f=hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f., 761c6c5fe18cb86212f23a9aac9b6e3c=TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c., 1588230740=hbase:meta,,1.1588230740, 38ed7bb40b7df5cbab5b70514657b066=TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.} 2024-12-02T01:12:58,005 DEBUG [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 38ed7bb40b7df5cbab5b70514657b066, 5acee627f693bbca053dc704b61b658f, 761c6c5fe18cb86212f23a9aac9b6e3c 2024-12-02T01:12:58,005 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:12:58,005 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:12:58,005 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:12:58,005 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:12:58,005 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:12:58,009 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/namespace/5acee627f693bbca053dc704b61b658f/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T01:12:58,009 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/hbase/meta/1588230740/recovered.edits/27.seqid, newMaxSeqId=27, maxSeqId=1 2024-12-02T01:12:58,010 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 5acee627f693bbca053dc704b61b658f: 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733101890195.5acee627f693bbca053dc704b61b658f. 2024-12-02T01:12:58,010 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 761c6c5fe18cb86212f23a9aac9b6e3c, disabling compactions & flushes 2024-12-02T01:12:58,010 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. after waiting 0 ms 2024-12-02T01:12:58,010 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:58,010 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994->hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb-top, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e844981557294726a8242e7962752dc5, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aff4712981754a02b4cf4efee2d5caeb, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e607caa580a4496d8161861bfaaffb8d, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/f37a1ca5850a44daa4f4930ef71dc91e, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a934843dfa184e179f6f050d01ee2803, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/fdaccb3c10804953a4bb57921b4d9e3a, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/23fad1102d0845058ebfb9473fc33a63, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a7fa121e4f9a41d8b8140389cfc11d01, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8bdd39b29c3c4aa582b51bdf7fb55810, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/6936eaea5ec443f5adf74679d5a1e971, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/2edbbac3993c458a8ed13a716cd70982, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/c6725f250b5a4f5eb3b2500c9ffd749e, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/34987475de8f4b509b3631a5efedbd42, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/7ca338eda0a1438fadabbba8dbc7def0, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/45ddc6055e304539b5b6a7543415d739, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8332ea3c24c04740873f8bc1d805acba, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aeca30517bab4ff6aac97dbc50b09db7] to archive 2024-12-02T01:12:58,011 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T01:12:58,013 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:12:58,014 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e844981557294726a8242e7962752dc5 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e844981557294726a8242e7962752dc5 2024-12-02T01:12:58,016 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/TestLogRolling-testLogRolling=fbd98b96465e3cae82d3686045e0a994-5f5f45f88d654c339e045094d137b5cc 2024-12-02T01:12:58,017 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aff4712981754a02b4cf4efee2d5caeb to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aff4712981754a02b4cf4efee2d5caeb 2024-12-02T01:12:58,018 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e607caa580a4496d8161861bfaaffb8d to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/e607caa580a4496d8161861bfaaffb8d 2024-12-02T01:12:58,019 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/f37a1ca5850a44daa4f4930ef71dc91e to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/f37a1ca5850a44daa4f4930ef71dc91e 2024-12-02T01:12:58,020 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a934843dfa184e179f6f050d01ee2803 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a934843dfa184e179f6f050d01ee2803 2024-12-02T01:12:58,022 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/fdaccb3c10804953a4bb57921b4d9e3a to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/fdaccb3c10804953a4bb57921b4d9e3a 2024-12-02T01:12:58,023 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/23fad1102d0845058ebfb9473fc33a63 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/23fad1102d0845058ebfb9473fc33a63 2024-12-02T01:12:58,024 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a7fa121e4f9a41d8b8140389cfc11d01 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/a7fa121e4f9a41d8b8140389cfc11d01 2024-12-02T01:12:58,025 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8bdd39b29c3c4aa582b51bdf7fb55810 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8bdd39b29c3c4aa582b51bdf7fb55810 2024-12-02T01:12:58,026 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/d3e02b3a14ca4f0e9883cc1d4a3f5b8f 2024-12-02T01:12:58,027 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/6936eaea5ec443f5adf74679d5a1e971 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/6936eaea5ec443f5adf74679d5a1e971 2024-12-02T01:12:58,029 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/2edbbac3993c458a8ed13a716cd70982 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/2edbbac3993c458a8ed13a716cd70982 2024-12-02T01:12:58,030 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/c6725f250b5a4f5eb3b2500c9ffd749e to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/c6725f250b5a4f5eb3b2500c9ffd749e 2024-12-02T01:12:58,031 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/34987475de8f4b509b3631a5efedbd42 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/34987475de8f4b509b3631a5efedbd42 2024-12-02T01:12:58,033 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/7ca338eda0a1438fadabbba8dbc7def0 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/7ca338eda0a1438fadabbba8dbc7def0 2024-12-02T01:12:58,034 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/45ddc6055e304539b5b6a7543415d739 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/45ddc6055e304539b5b6a7543415d739 2024-12-02T01:12:58,036 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8332ea3c24c04740873f8bc1d805acba to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/8332ea3c24c04740873f8bc1d805acba 2024-12-02T01:12:58,037 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aeca30517bab4ff6aac97dbc50b09db7 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/info/aeca30517bab4ff6aac97dbc50b09db7 2024-12-02T01:12:58,043 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/761c6c5fe18cb86212f23a9aac9b6e3c/recovered.edits/332.seqid, newMaxSeqId=332, maxSeqId=107 2024-12-02T01:12:58,043 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:58,043 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 761c6c5fe18cb86212f23a9aac9b6e3c: 2024-12-02T01:12:58,044 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRolling,row0062,1733101915216.761c6c5fe18cb86212f23a9aac9b6e3c. 2024-12-02T01:12:58,044 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 38ed7bb40b7df5cbab5b70514657b066, disabling compactions & flushes 2024-12-02T01:12:58,044 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:12:58,044 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:12:58,044 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. after waiting 0 ms 2024-12-02T01:12:58,044 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:12:58,044 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994->hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/fbd98b96465e3cae82d3686045e0a994/info/a768bc17c29b4ff89ab23cfced6233eb-bottom] to archive 2024-12-02T01:12:58,045 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-02T01:12:58,047 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994 to hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/archive/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/info/a768bc17c29b4ff89ab23cfced6233eb.fbd98b96465e3cae82d3686045e0a994 2024-12-02T01:12:58,052 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/data/default/TestLogRolling-testLogRolling/38ed7bb40b7df5cbab5b70514657b066/recovered.edits/112.seqid, newMaxSeqId=112, maxSeqId=107 2024-12-02T01:12:58,053 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:12:58,053 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 38ed7bb40b7df5cbab5b70514657b066: 2024-12-02T01:12:58,053 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRolling,,1733101915216.38ed7bb40b7df5cbab5b70514657b066. 2024-12-02T01:12:58,205 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,32775,1733101889168; all regions closed. 2024-12-02T01:12:58,206 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168 2024-12-02T01:12:58,210 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741834_1010 (size=9351) 2024-12-02T01:12:58,211 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741834_1010 (size=9351) 2024-12-02T01:12:58,215 DEBUG [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/oldWALs 2024-12-02T01:12:58,215 INFO [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C32775%2C1733101889168.meta:.meta(num 1733101890117) 2024-12-02T01:12:58,216 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/WALs/a82f277fe18c,32775,1733101889168 2024-12-02T01:12:58,218 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741878_1054 (size=1071) 2024-12-02T01:12:58,218 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741878_1054 (size=1071) 2024-12-02T01:12:58,221 DEBUG [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/oldWALs 2024-12-02T01:12:58,221 INFO [RS:0;a82f277fe18c:32775 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C32775%2C1733101889168:(num 1733101977774) 2024-12-02T01:12:58,221 DEBUG [RS:0;a82f277fe18c:32775 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:12:58,221 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:12:58,221 INFO [RS:0;a82f277fe18c:32775 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-02T01:12:58,222 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:12:58,222 INFO [RS:0;a82f277fe18c:32775 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:32775 2024-12-02T01:12:58,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:12:58,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,32775,1733101889168 2024-12-02T01:12:58,241 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,32775,1733101889168] 2024-12-02T01:12:58,241 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,32775,1733101889168; numProcessing=1 2024-12-02T01:12:58,249 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,32775,1733101889168 already deleted, retry=false 2024-12-02T01:12:58,249 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,32775,1733101889168 expired; onlineServers=0 2024-12-02T01:12:58,250 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,35029,1733101889015' ***** 2024-12-02T01:12:58,250 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T01:12:58,250 DEBUG [M:0;a82f277fe18c:35029 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6c38cd0a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:12:58,250 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,35029,1733101889015 2024-12-02T01:12:58,250 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,35029,1733101889015; all regions closed. 2024-12-02T01:12:58,250 DEBUG [M:0;a82f277fe18c:35029 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:12:58,250 DEBUG [M:0;a82f277fe18c:35029 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T01:12:58,250 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T01:12:58,250 DEBUG [M:0;a82f277fe18c:35029 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T01:12:58,250 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101889463 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101889463,5,FailOnTimeoutGroup] 2024-12-02T01:12:58,250 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101889462 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101889462,5,FailOnTimeoutGroup] 2024-12-02T01:12:58,251 INFO [M:0;a82f277fe18c:35029 {}] hbase.ChoreService(370): Chore service for: master/a82f277fe18c:0 had [] on shutdown 2024-12-02T01:12:58,251 DEBUG [M:0;a82f277fe18c:35029 {}] master.HMaster(1733): Stopping service threads 2024-12-02T01:12:58,251 INFO [M:0;a82f277fe18c:35029 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T01:12:58,251 INFO [M:0;a82f277fe18c:35029 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T01:12:58,251 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T01:12:58,258 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T01:12:58,258 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:12:58,258 DEBUG [M:0;a82f277fe18c:35029 {}] zookeeper.ZKUtil(347): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T01:12:58,258 WARN [M:0;a82f277fe18c:35029 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T01:12:58,258 INFO [M:0;a82f277fe18c:35029 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-02T01:12:58,258 INFO [M:0;a82f277fe18c:35029 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T01:12:58,258 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:12:58,258 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:12:58,259 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:12:58,259 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:12:58,259 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:12:58,259 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:12:58,259 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=66.42 KB heapSize=81.66 KB 2024-12-02T01:12:58,279 DEBUG [M:0;a82f277fe18c:35029 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f01a7f562c8c49038037dcf08899cc32 is 82, key is hbase:meta,,1/info:regioninfo/1733101890136/Put/seqid=0 2024-12-02T01:12:58,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741879_1055 (size=5672) 2024-12-02T01:12:58,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741879_1055 (size=5672) 2024-12-02T01:12:58,284 INFO [M:0;a82f277fe18c:35029 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f01a7f562c8c49038037dcf08899cc32 2024-12-02T01:12:58,302 DEBUG [M:0;a82f277fe18c:35029 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/cba9bcc30b00443995dfcc1772c1d8ae is 749, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733101891306/Put/seqid=0 2024-12-02T01:12:58,306 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741880_1056 (size=7284) 2024-12-02T01:12:58,306 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741880_1056 (size=7284) 2024-12-02T01:12:58,307 INFO [M:0;a82f277fe18c:35029 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65.82 KB at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/cba9bcc30b00443995dfcc1772c1d8ae 2024-12-02T01:12:58,311 INFO [M:0;a82f277fe18c:35029 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for cba9bcc30b00443995dfcc1772c1d8ae 2024-12-02T01:12:58,323 DEBUG [M:0;a82f277fe18c:35029 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/9c42191cf3514d06b4635bdb2693c4f2 is 69, key is a82f277fe18c,32775,1733101889168/rs:state/1733101889523/Put/seqid=0 2024-12-02T01:12:58,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741881_1057 (size=5156) 2024-12-02T01:12:58,328 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741881_1057 (size=5156) 2024-12-02T01:12:58,328 INFO [M:0;a82f277fe18c:35029 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/9c42191cf3514d06b4635bdb2693c4f2 2024-12-02T01:12:58,341 INFO [RS:0;a82f277fe18c:32775 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,32775,1733101889168; zookeeper connection closed. 2024-12-02T01:12:58,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:12:58,341 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32775-0x10194e1c8590001, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:12:58,341 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@386de44e {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@386de44e 2024-12-02T01:12:58,342 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T01:12:58,347 DEBUG [M:0;a82f277fe18c:35029 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/a35bf4de763e4d328a64638edbcc477f is 52, key is load_balancer_on/state:d/1733101890913/Put/seqid=0 2024-12-02T01:12:58,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741882_1058 (size=5056) 2024-12-02T01:12:58,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741882_1058 (size=5056) 2024-12-02T01:12:58,351 INFO [M:0;a82f277fe18c:35029 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/a35bf4de763e4d328a64638edbcc477f 2024-12-02T01:12:58,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-02T01:12:58,353 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling 2024-12-02T01:12:58,356 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f01a7f562c8c49038037dcf08899cc32 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/f01a7f562c8c49038037dcf08899cc32 2024-12-02T01:12:58,360 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/f01a7f562c8c49038037dcf08899cc32, entries=8, sequenceid=164, filesize=5.5 K 2024-12-02T01:12:58,360 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/cba9bcc30b00443995dfcc1772c1d8ae as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/cba9bcc30b00443995dfcc1772c1d8ae 2024-12-02T01:12:58,364 INFO [M:0;a82f277fe18c:35029 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for cba9bcc30b00443995dfcc1772c1d8ae 2024-12-02T01:12:58,364 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/cba9bcc30b00443995dfcc1772c1d8ae, entries=18, sequenceid=164, filesize=7.1 K 2024-12-02T01:12:58,365 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/9c42191cf3514d06b4635bdb2693c4f2 as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/9c42191cf3514d06b4635bdb2693c4f2 2024-12-02T01:12:58,369 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/9c42191cf3514d06b4635bdb2693c4f2, entries=1, sequenceid=164, filesize=5.0 K 2024-12-02T01:12:58,370 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/a35bf4de763e4d328a64638edbcc477f as hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/a35bf4de763e4d328a64638edbcc477f 2024-12-02T01:12:58,374 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35829/user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/a35bf4de763e4d328a64638edbcc477f, entries=1, sequenceid=164, filesize=4.9 K 2024-12-02T01:12:58,375 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(3040): Finished flush of dataSize ~66.42 KB/68019, heapSize ~81.60 KB/83560, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 116ms, sequenceid=164, compaction requested=false 2024-12-02T01:12:58,377 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:12:58,377 DEBUG [M:0;a82f277fe18c:35029 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:12:58,377 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/1be1b73d-5261-0795-081f-64a329bb5cc7/MasterData/WALs/a82f277fe18c,35029,1733101889015 2024-12-02T01:12:58,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43899 is added to blk_1073741830_1006 (size=79248) 2024-12-02T01:12:58,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39789 is added to blk_1073741830_1006 (size=79248) 2024-12-02T01:12:58,379 INFO [M:0;a82f277fe18c:35029 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-02T01:12:58,379 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:12:58,379 INFO [M:0;a82f277fe18c:35029 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:35029 2024-12-02T01:12:58,382 DEBUG [M:0;a82f277fe18c:35029 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/a82f277fe18c,35029,1733101889015 already deleted, retry=false 2024-12-02T01:12:58,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:12:58,491 INFO [M:0;a82f277fe18c:35029 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,35029,1733101889015; zookeeper connection closed. 2024-12-02T01:12:58,491 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35029-0x10194e1c8590000, quorum=127.0.0.1:64314, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:12:58,493 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@723a744c{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:12:58,493 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4dc1d62e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:12:58,493 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:12:58,493 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@21bfc9c0{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:12:58,493 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3640f2d6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir/,STOPPED} 2024-12-02T01:12:58,494 WARN [BP-365843615-172.17.0.2-1733101887286 heartbeating to localhost/127.0.0.1:35829 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:12:58,494 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:12:58,494 WARN [BP-365843615-172.17.0.2-1733101887286 heartbeating to localhost/127.0.0.1:35829 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-365843615-172.17.0.2-1733101887286 (Datanode Uuid c36b8b12-0dce-4030-b8cd-5b907b6b0f7b) service to localhost/127.0.0.1:35829 2024-12-02T01:12:58,494 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:12:58,495 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data3/current/BP-365843615-172.17.0.2-1733101887286 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:12:58,495 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data4/current/BP-365843615-172.17.0.2-1733101887286 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:12:58,495 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:12:58,497 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4aa14da8{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:12:58,497 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@55682800{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:12:58,497 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:58,497 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:12:58,497 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1df47ba6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:12:58,497 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@68604cef{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir/,STOPPED} 2024-12-02T01:12:58,498 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:12:58,498 WARN [BP-365843615-172.17.0.2-1733101887286 heartbeating to localhost/127.0.0.1:35829 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:12:58,499 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:12:58,499 WARN [BP-365843615-172.17.0.2-1733101887286 heartbeating to localhost/127.0.0.1:35829 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-365843615-172.17.0.2-1733101887286 (Datanode Uuid 1451500d-ad1d-4e1c-8a7f-5a80ce5c779d) service to localhost/127.0.0.1:35829 2024-12-02T01:12:58,499 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data1/current/BP-365843615-172.17.0.2-1733101887286 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:12:58,499 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/cluster_500621b8-eb94-f8d0-3524-f8a028510f33/dfs/data/data2/current/BP-365843615-172.17.0.2-1733101887286 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:12:58,499 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:12:58,504 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3a6af3f1{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:12:58,505 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4525abcd{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:12:58,505 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:12:58,505 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6780485a{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:12:58,505 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3741aa5c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir/,STOPPED} 2024-12-02T01:12:58,511 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-02T01:12:58,539 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-02T01:12:58,545 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRolling Thread=126 (was 112) - Thread LEAK? -, OpenFileDescriptor=490 (was 466) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=158 (was 212), ProcessCount=11 (was 11), AvailableMemoryMB=4781 (was 5320) 2024-12-02T01:12:58,550 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnNothingWritten Thread=127, OpenFileDescriptor=490, MaxFileDescriptor=1048576, SystemLoadAverage=158, ProcessCount=11, AvailableMemoryMB=4780 2024-12-02T01:12:58,550 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.log.dir so I do NOT create it in target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/2c55a281-c0c4-d40e-0985-c4998b08bbcd/hadoop.tmp.dir so I do NOT create it in target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577, deleteOnExit=true 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/test.cache.data in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.tmp.dir in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-02T01:12:58,551 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-02T01:12:58,551 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/nfs.dump.dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/java.io.tmpdir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-02T01:12:58,552 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-02T01:12:58,563 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:12:58,801 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:12:58,805 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:12:58,806 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:12:58,806 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:12:58,806 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-02T01:12:58,806 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:12:58,807 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7fcf5e50{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:12:58,807 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@692b0555{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:12:58,896 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6787fcf{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/java.io.tmpdir/jetty-localhost-42317-hadoop-hdfs-3_4_1-tests_jar-_-any-14322852644274068777/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:12:58,897 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7c46377a{HTTP/1.1, (http/1.1)}{localhost:42317} 2024-12-02T01:12:58,897 INFO [Time-limited test {}] server.Server(415): Started @386025ms 2024-12-02T01:12:58,907 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-02T01:12:59,157 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:12:59,160 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:12:59,163 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:12:59,163 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:12:59,163 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:12:59,164 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24616503{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:12:59,164 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5ebf0c9b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:12:59,253 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6c6a7c75{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/java.io.tmpdir/jetty-localhost-40227-hadoop-hdfs-3_4_1-tests_jar-_-any-13318519995959116302/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:12:59,254 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@29776a3e{HTTP/1.1, (http/1.1)}{localhost:40227} 2024-12-02T01:12:59,254 INFO [Time-limited test {}] server.Server(415): Started @386382ms 2024-12-02T01:12:59,255 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:12:59,278 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-02T01:12:59,281 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-02T01:12:59,281 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-02T01:12:59,281 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-02T01:12:59,281 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-02T01:12:59,282 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7284e1dc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir/,AVAILABLE} 2024-12-02T01:12:59,282 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@707d70b1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-02T01:12:59,371 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@1706478b{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/java.io.tmpdir/jetty-localhost-46005-hadoop-hdfs-3_4_1-tests_jar-_-any-17375163322546340514/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:12:59,371 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@2c2c59ef{HTTP/1.1, (http/1.1)}{localhost:46005} 2024-12-02T01:12:59,371 INFO [Time-limited test {}] server.Server(415): Started @386500ms 2024-12-02T01:12:59,372 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-02T01:12:59,498 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:12:59,548 INFO [regionserver/a82f277fe18c:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:12:59,952 WARN [Thread-2246 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data1/current/BP-2124532887-172.17.0.2-1733101978566/current, will proceed with Du for space computation calculation, 2024-12-02T01:12:59,953 WARN [Thread-2247 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data2/current/BP-2124532887-172.17.0.2-1733101978566/current, will proceed with Du for space computation calculation, 2024-12-02T01:12:59,972 WARN [Thread-2210 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:12:59,973 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd30fb70e237b94a4 with lease ID 0x3f10e4ca359c645f: Processing first storage report for DS-17220729-369d-4bb7-9d6c-9d34172022c2 from datanode DatanodeRegistration(127.0.0.1:34733, datanodeUuid=1ec20cb5-fddc-49a2-b15e-7d6679039a2f, infoPort=40737, infoSecurePort=0, ipcPort=34203, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566) 2024-12-02T01:12:59,973 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd30fb70e237b94a4 with lease ID 0x3f10e4ca359c645f: from storage DS-17220729-369d-4bb7-9d6c-9d34172022c2 node DatanodeRegistration(127.0.0.1:34733, datanodeUuid=1ec20cb5-fddc-49a2-b15e-7d6679039a2f, infoPort=40737, infoSecurePort=0, ipcPort=34203, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:12:59,974 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xd30fb70e237b94a4 with lease ID 0x3f10e4ca359c645f: Processing first storage report for DS-b2e0f49b-e12c-4cce-832c-cfe123bc418a from datanode DatanodeRegistration(127.0.0.1:34733, datanodeUuid=1ec20cb5-fddc-49a2-b15e-7d6679039a2f, infoPort=40737, infoSecurePort=0, ipcPort=34203, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566) 2024-12-02T01:12:59,974 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xd30fb70e237b94a4 with lease ID 0x3f10e4ca359c645f: from storage DS-b2e0f49b-e12c-4cce-832c-cfe123bc418a node DatanodeRegistration(127.0.0.1:34733, datanodeUuid=1ec20cb5-fddc-49a2-b15e-7d6679039a2f, infoPort=40737, infoSecurePort=0, ipcPort=34203, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:13:00,043 WARN [Thread-2257 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data3/current/BP-2124532887-172.17.0.2-1733101978566/current, will proceed with Du for space computation calculation, 2024-12-02T01:13:00,043 WARN [Thread-2258 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data4/current/BP-2124532887-172.17.0.2-1733101978566/current, will proceed with Du for space computation calculation, 2024-12-02T01:13:00,064 WARN [Thread-2233 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-02T01:13:00,066 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa54c4b5acdb9bae6 with lease ID 0x3f10e4ca359c6460: Processing first storage report for DS-f357490f-f864-4e28-856b-613b00ce8e9e from datanode DatanodeRegistration(127.0.0.1:35771, datanodeUuid=b34e62b6-8253-4a3d-863a-c22c8b20c76d, infoPort=40333, infoSecurePort=0, ipcPort=40845, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566) 2024-12-02T01:13:00,066 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa54c4b5acdb9bae6 with lease ID 0x3f10e4ca359c6460: from storage DS-f357490f-f864-4e28-856b-613b00ce8e9e node DatanodeRegistration(127.0.0.1:35771, datanodeUuid=b34e62b6-8253-4a3d-863a-c22c8b20c76d, infoPort=40333, infoSecurePort=0, ipcPort=40845, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-02T01:13:00,066 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xa54c4b5acdb9bae6 with lease ID 0x3f10e4ca359c6460: Processing first storage report for DS-2fd3ba40-02a7-4e5c-adce-c99d70447605 from datanode DatanodeRegistration(127.0.0.1:35771, datanodeUuid=b34e62b6-8253-4a3d-863a-c22c8b20c76d, infoPort=40333, infoSecurePort=0, ipcPort=40845, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566) 2024-12-02T01:13:00,067 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa54c4b5acdb9bae6 with lease ID 0x3f10e4ca359c6460: from storage DS-2fd3ba40-02a7-4e5c-adce-c99d70447605 node DatanodeRegistration(127.0.0.1:35771, datanodeUuid=b34e62b6-8253-4a3d-863a-c22c8b20c76d, infoPort=40333, infoSecurePort=0, ipcPort=40845, storageInfo=lv=-57;cid=testClusterID;nsid=806342048;c=1733101978566), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-02T01:13:00,095 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db 2024-12-02T01:13:00,100 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/zookeeper_0, clientPort=63646, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-02T01:13:00,101 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=63646 2024-12-02T01:13:00,101 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,103 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,114 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:13:00,114 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741825_1001 (size=7) 2024-12-02T01:13:00,115 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6 with version=8 2024-12-02T01:13:00,115 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:41885/user/jenkins/test-data/867791d2-cb74-54e4-ffbf-3262951c8345/hbase-staging 2024-12-02T01:13:00,117 INFO [Time-limited test {}] client.ConnectionUtils(129): master/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:13:00,117 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:13:00,118 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:40313 2024-12-02T01:13:00,118 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,119 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,121 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:40313 connecting to ZooKeeper ensemble=127.0.0.1:63646 2024-12-02T01:13:00,173 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:403130x0, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:13:00,174 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:40313-0x10194e32c360000 connected 2024-12-02T01:13:00,243 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:13:00,244 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:13:00,244 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:13:00,245 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=40313 2024-12-02T01:13:00,245 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=40313 2024-12-02T01:13:00,245 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=40313 2024-12-02T01:13:00,246 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=40313 2024-12-02T01:13:00,246 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=40313 2024-12-02T01:13:00,246 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6, hbase.cluster.distributed=false 2024-12-02T01:13:00,266 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/a82f277fe18c:0 server-side Connection retries=45 2024-12-02T01:13:00,266 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:13:00,267 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-02T01:13:00,267 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-02T01:13:00,267 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-02T01:13:00,267 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-02T01:13:00,267 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-02T01:13:00,267 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-02T01:13:00,268 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:39381 2024-12-02T01:13:00,268 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-02T01:13:00,268 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-02T01:13:00,269 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,271 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,274 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:39381 connecting to ZooKeeper ensemble=127.0.0.1:63646 2024-12-02T01:13:00,285 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:393810x0, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-02T01:13:00,285 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39381-0x10194e32c360001 connected 2024-12-02T01:13:00,285 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:13:00,286 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:13:00,286 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-02T01:13:00,287 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39381 2024-12-02T01:13:00,287 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39381 2024-12-02T01:13:00,287 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39381 2024-12-02T01:13:00,287 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39381 2024-12-02T01:13:00,288 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39381 2024-12-02T01:13:00,288 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,293 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:13:00,293 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:13:00,293 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,298 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;a82f277fe18c:40313 2024-12-02T01:13:00,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:13:00,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-02T01:13:00,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,302 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:13:00,302 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/a82f277fe18c,40313,1733101980116 from backup master directory 2024-12-02T01:13:00,310 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,310 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:13:00,310 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:13:00,310 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-02T01:13:00,310 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,310 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-02T01:13:00,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:13:00,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741826_1002 (size=42) 2024-12-02T01:13:00,319 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/hbase.id with ID: 8f35db8b-eeba-4386-9005-fd5005dbf318 2024-12-02T01:13:00,327 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:00,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,334 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:13:00,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741827_1003 (size=196) 2024-12-02T01:13:00,341 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-02T01:13:00,342 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-02T01:13:00,342 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:13:00,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:13:00,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741828_1004 (size=1189) 2024-12-02T01:13:00,352 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store 2024-12-02T01:13:00,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:13:00,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741829_1005 (size=34) 2024-12-02T01:13:00,358 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:13:00,359 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:13:00,359 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:00,359 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:00,359 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:13:00,359 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:00,359 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:00,359 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:13:00,360 WARN [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/.initializing 2024-12-02T01:13:00,360 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/WALs/a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,362 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C40313%2C1733101980116, suffix=, logDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/WALs/a82f277fe18c,40313,1733101980116, archiveDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/oldWALs, maxLogs=10 2024-12-02T01:13:00,362 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C40313%2C1733101980116.1733101980362 2024-12-02T01:13:00,366 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/WALs/a82f277fe18c,40313,1733101980116/a82f277fe18c%2C40313%2C1733101980116.1733101980362 2024-12-02T01:13:00,367 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40333:40333),(127.0.0.1/127.0.0.1:40737:40737)] 2024-12-02T01:13:00,367 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:13:00,367 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:13:00,367 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,367 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,369 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,370 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-02T01:13:00,370 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,370 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:00,371 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,372 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-02T01:13:00,372 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,372 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:13:00,372 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,373 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-02T01:13:00,373 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,373 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:13:00,374 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,374 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-02T01:13:00,374 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,375 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:13:00,375 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,376 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,377 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-02T01:13:00,378 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-02T01:13:00,380 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:13:00,380 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=762532, jitterRate=-0.030390769243240356}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-02T01:13:00,381 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:13:00,381 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-02T01:13:00,384 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5fa5cf92, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:13:00,385 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-02T01:13:00,385 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-02T01:13:00,385 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-02T01:13:00,385 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-02T01:13:00,386 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-02T01:13:00,386 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-02T01:13:00,386 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-02T01:13:00,388 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-02T01:13:00,388 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-02T01:13:00,393 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-02T01:13:00,393 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-02T01:13:00,393 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-02T01:13:00,401 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-02T01:13:00,401 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-02T01:13:00,402 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-02T01:13:00,409 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-02T01:13:00,410 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-02T01:13:00,422 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-02T01:13:00,424 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-02T01:13:00,434 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-02T01:13:00,443 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:13:00,443 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-02T01:13:00,443 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,443 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,443 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=a82f277fe18c,40313,1733101980116, sessionid=0x10194e32c360000, setting cluster-up flag (Was=false) 2024-12-02T01:13:00,460 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,460 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,485 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-02T01:13:00,487 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,499 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:13:00,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,501 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,526 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-02T01:13:00,529 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=a82f277fe18c,40313,1733101980116 2024-12-02T01:13:00,534 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-02T01:13:00,534 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-02T01:13:00,534 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-02T01:13:00,535 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: a82f277fe18c,40313,1733101980116 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-02T01:13:00,535 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:13:00,535 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:13:00,535 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:13:00,535 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/a82f277fe18c:0, corePoolSize=5, maxPoolSize=5 2024-12-02T01:13:00,535 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/a82f277fe18c:0, corePoolSize=10, maxPoolSize=10 2024-12-02T01:13:00,536 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,536 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:13:00,536 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733102010537 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-02T01:13:00,537 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:13:00,537 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-02T01:13:00,537 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,538 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-02T01:13:00,538 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-02T01:13:00,538 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-02T01:13:00,538 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-02T01:13:00,538 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-02T01:13:00,538 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,538 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101980538,5,FailOnTimeoutGroup] 2024-12-02T01:13:00,538 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101980538,5,FailOnTimeoutGroup] 2024-12-02T01:13:00,538 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:13:00,538 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,539 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-02T01:13:00,539 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,539 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:13:00,544 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741831_1007 (size=1039) 2024-12-02T01:13:00,545 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-02T01:13:00,545 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6 2024-12-02T01:13:00,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:13:00,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741832_1008 (size=32) 2024-12-02T01:13:00,553 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:13:00,554 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:13:00,555 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:13:00,555 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,556 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:00,556 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:13:00,557 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:13:00,557 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,557 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:00,557 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:13:00,558 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:13:00,559 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:00,559 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:00,560 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740 2024-12-02T01:13:00,560 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740 2024-12-02T01:13:00,562 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:13:00,563 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:13:00,565 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:13:00,566 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=873918, jitterRate=0.11124420166015625}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:13:00,566 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:13:00,566 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:13:00,566 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:13:00,566 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:13:00,566 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:13:00,566 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:13:00,567 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:13:00,567 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:13:00,568 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-02T01:13:00,568 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-02T01:13:00,568 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-02T01:13:00,569 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-02T01:13:00,570 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-02T01:13:00,603 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;a82f277fe18c:39381 2024-12-02T01:13:00,604 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1008): ClusterId : 8f35db8b-eeba-4386-9005-fd5005dbf318 2024-12-02T01:13:00,604 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-02T01:13:00,610 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-02T01:13:00,610 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-02T01:13:00,618 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-02T01:13:00,619 DEBUG [RS:0;a82f277fe18c:39381 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6078499b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:13:00,619 DEBUG [RS:0;a82f277fe18c:39381 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@14e4a377, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:13:00,619 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-02T01:13:00,619 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-02T01:13:00,619 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-02T01:13:00,620 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(3073): reportForDuty to master=a82f277fe18c,40313,1733101980116 with isa=a82f277fe18c/172.17.0.2:39381, startcode=1733101980266 2024-12-02T01:13:00,620 DEBUG [RS:0;a82f277fe18c:39381 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-02T01:13:00,622 INFO [RS-EventLoopGroup-14-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43319, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.6 (auth:SIMPLE), service=RegionServerStatusService 2024-12-02T01:13:00,622 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40313 {}] master.ServerManager(332): Checking decommissioned status of RegionServer a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,622 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=40313 {}] master.ServerManager(486): Registering regionserver=a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,624 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6 2024-12-02T01:13:00,624 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:34277 2024-12-02T01:13:00,624 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-02T01:13:00,635 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:13:00,635 DEBUG [RS:0;a82f277fe18c:39381 {}] zookeeper.ZKUtil(111): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,635 WARN [RS:0;a82f277fe18c:39381 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-02T01:13:00,635 INFO [RS:0;a82f277fe18c:39381 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:13:00,635 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,636 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [a82f277fe18c,39381,1733101980266] 2024-12-02T01:13:00,639 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-02T01:13:00,639 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-02T01:13:00,641 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-02T01:13:00,641 INFO [RS:0;a82f277fe18c:39381 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-02T01:13:00,641 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,641 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-02T01:13:00,642 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,642 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,642 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,642 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,642 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,642 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/a82f277fe18c:0, corePoolSize=2, maxPoolSize=2 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/a82f277fe18c:0, corePoolSize=1, maxPoolSize=1 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:13:00,643 DEBUG [RS:0;a82f277fe18c:39381 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/a82f277fe18c:0, corePoolSize=3, maxPoolSize=3 2024-12-02T01:13:00,643 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,643 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,643 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,643 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,643 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39381,1733101980266-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:13:00,661 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-02T01:13:00,661 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,39381,1733101980266-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:00,674 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.Replication(204): a82f277fe18c,39381,1733101980266 started 2024-12-02T01:13:00,674 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1767): Serving as a82f277fe18c,39381,1733101980266, RpcServer on a82f277fe18c/172.17.0.2:39381, sessionid=0x10194e32c360001 2024-12-02T01:13:00,674 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-02T01:13:00,674 DEBUG [RS:0;a82f277fe18c:39381 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,674 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,39381,1733101980266' 2024-12-02T01:13:00,674 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-02T01:13:00,675 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-02T01:13:00,675 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-02T01:13:00,675 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-02T01:13:00,675 DEBUG [RS:0;a82f277fe18c:39381 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,675 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'a82f277fe18c,39381,1733101980266' 2024-12-02T01:13:00,676 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-02T01:13:00,676 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-02T01:13:00,676 DEBUG [RS:0;a82f277fe18c:39381 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-02T01:13:00,676 INFO [RS:0;a82f277fe18c:39381 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-02T01:13:00,676 INFO [RS:0;a82f277fe18c:39381 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-02T01:13:00,720 WARN [a82f277fe18c:40313 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-02T01:13:00,780 INFO [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C39381%2C1733101980266, suffix=, logDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266, archiveDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs, maxLogs=32 2024-12-02T01:13:00,781 INFO [RS:0;a82f277fe18c:39381 {}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C39381%2C1733101980266.1733101980781 2024-12-02T01:13:00,792 INFO [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266/a82f277fe18c%2C39381%2C1733101980266.1733101980781 2024-12-02T01:13:00,792 DEBUG [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40737:40737),(127.0.0.1/127.0.0.1:40333:40333)] 2024-12-02T01:13:00,971 DEBUG [a82f277fe18c:40313 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-02T01:13:00,971 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=a82f277fe18c,39381,1733101980266 2024-12-02T01:13:00,973 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,39381,1733101980266, state=OPENING 2024-12-02T01:13:00,985 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-02T01:13:00,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:00,994 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=a82f277fe18c,39381,1733101980266}] 2024-12-02T01:13:00,994 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:13:00,994 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:13:01,148 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to a82f277fe18c,39381,1733101980266 2024-12-02T01:13:01,149 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-02T01:13:01,151 INFO [RS-EventLoopGroup-15-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51950, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-02T01:13:01,156 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-02T01:13:01,156 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:13:01,159 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=a82f277fe18c%2C39381%2C1733101980266.meta, suffix=.meta, logDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266, archiveDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs, maxLogs=32 2024-12-02T01:13:01,160 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor a82f277fe18c%2C39381%2C1733101980266.meta.1733101981159.meta 2024-12-02T01:13:01,168 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266/a82f277fe18c%2C39381%2C1733101980266.meta.1733101981159.meta 2024-12-02T01:13:01,168 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40737:40737),(127.0.0.1/127.0.0.1:40333:40333)] 2024-12-02T01:13:01,168 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:13:01,169 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-02T01:13:01,169 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-02T01:13:01,169 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-02T01:13:01,169 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-02T01:13:01,169 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:13:01,169 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-02T01:13:01,169 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-02T01:13:01,171 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-02T01:13:01,171 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-02T01:13:01,171 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:01,172 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:01,172 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-02T01:13:01,172 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-02T01:13:01,172 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:01,173 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:01,173 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-02T01:13:01,173 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-02T01:13:01,173 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:01,174 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-02T01:13:01,174 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740 2024-12-02T01:13:01,175 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740 2024-12-02T01:13:01,176 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-02T01:13:01,177 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-02T01:13:01,178 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=695140, jitterRate=-0.11608459055423737}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-02T01:13:01,178 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-02T01:13:01,178 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733101981148 2024-12-02T01:13:01,180 DEBUG [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-02T01:13:01,180 INFO [RS_OPEN_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-02T01:13:01,180 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,39381,1733101980266 2024-12-02T01:13:01,181 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as a82f277fe18c,39381,1733101980266, state=OPEN 2024-12-02T01:13:01,240 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:13:01,240 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-02T01:13:01,241 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:13:01,241 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-02T01:13:01,243 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-02T01:13:01,243 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=a82f277fe18c,39381,1733101980266 in 246 msec 2024-12-02T01:13:01,246 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-02T01:13:01,246 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 675 msec 2024-12-02T01:13:01,248 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 715 msec 2024-12-02T01:13:01,249 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733101981248, completionTime=-1 2024-12-02T01:13:01,249 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-02T01:13:01,249 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-02T01:13:01,250 DEBUG [hconnection-0x59664240-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:13:01,251 INFO [RS-EventLoopGroup-15-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51952, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:13:01,252 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-02T01:13:01,252 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733102041252 2024-12-02T01:13:01,252 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733102101252 2024-12-02T01:13:01,252 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 2 msec 2024-12-02T01:13:01,276 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40313,1733101980116-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:01,276 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40313,1733101980116-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:01,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40313,1733101980116-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:01,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-a82f277fe18c:40313, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:01,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:01,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-02T01:13:01,277 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-02T01:13:01,278 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-02T01:13:01,278 DEBUG [master/a82f277fe18c:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-02T01:13:01,279 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-02T01:13:01,279 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:01,280 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-02T01:13:01,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:13:01,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741835_1011 (size=358) 2024-12-02T01:13:01,292 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 66584dc4c952477e12dd85d81277ebd9, NAME => 'hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6 2024-12-02T01:13:01,299 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:13:01,299 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741836_1012 (size=42) 2024-12-02T01:13:01,300 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:13:01,300 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 66584dc4c952477e12dd85d81277ebd9, disabling compactions & flushes 2024-12-02T01:13:01,300 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,300 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,300 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. after waiting 0 ms 2024-12-02T01:13:01,300 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,300 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,300 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 66584dc4c952477e12dd85d81277ebd9: 2024-12-02T01:13:01,301 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-02T01:13:01,301 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733101981301"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733101981301"}]},"ts":"1733101981301"} 2024-12-02T01:13:01,303 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-02T01:13:01,304 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-02T01:13:01,304 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101981304"}]},"ts":"1733101981304"} 2024-12-02T01:13:01,305 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-02T01:13:01,318 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=66584dc4c952477e12dd85d81277ebd9, ASSIGN}] 2024-12-02T01:13:01,319 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=66584dc4c952477e12dd85d81277ebd9, ASSIGN 2024-12-02T01:13:01,320 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=66584dc4c952477e12dd85d81277ebd9, ASSIGN; state=OFFLINE, location=a82f277fe18c,39381,1733101980266; forceNewPlan=false, retain=false 2024-12-02T01:13:01,471 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=66584dc4c952477e12dd85d81277ebd9, regionState=OPENING, regionLocation=a82f277fe18c,39381,1733101980266 2024-12-02T01:13:01,474 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 66584dc4c952477e12dd85d81277ebd9, server=a82f277fe18c,39381,1733101980266}] 2024-12-02T01:13:01,500 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:13:01,627 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to a82f277fe18c,39381,1733101980266 2024-12-02T01:13:01,633 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,633 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 66584dc4c952477e12dd85d81277ebd9, NAME => 'hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9.', STARTKEY => '', ENDKEY => ''} 2024-12-02T01:13:01,633 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,633 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-02T01:13:01,634 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,634 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,636 INFO [StoreOpener-66584dc4c952477e12dd85d81277ebd9-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,639 INFO [StoreOpener-66584dc4c952477e12dd85d81277ebd9-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 66584dc4c952477e12dd85d81277ebd9 columnFamilyName info 2024-12-02T01:13:01,639 DEBUG [StoreOpener-66584dc4c952477e12dd85d81277ebd9-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-02T01:13:01,640 INFO [StoreOpener-66584dc4c952477e12dd85d81277ebd9-1 {}] regionserver.HStore(327): Store=66584dc4c952477e12dd85d81277ebd9/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-02T01:13:01,641 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,642 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,645 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,648 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-02T01:13:01,649 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 66584dc4c952477e12dd85d81277ebd9; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=830447, jitterRate=0.05596804618835449}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-02T01:13:01,649 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 66584dc4c952477e12dd85d81277ebd9: 2024-12-02T01:13:01,650 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9., pid=6, masterSystemTime=1733101981627 2024-12-02T01:13:01,651 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,652 INFO [RS_OPEN_PRIORITY_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,652 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=66584dc4c952477e12dd85d81277ebd9, regionState=OPEN, openSeqNum=2, regionLocation=a82f277fe18c,39381,1733101980266 2024-12-02T01:13:01,655 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-02T01:13:01,655 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 66584dc4c952477e12dd85d81277ebd9, server=a82f277fe18c,39381,1733101980266 in 180 msec 2024-12-02T01:13:01,656 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-02T01:13:01,656 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=66584dc4c952477e12dd85d81277ebd9, ASSIGN in 337 msec 2024-12-02T01:13:01,657 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-02T01:13:01,657 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733101981657"}]},"ts":"1733101981657"} 2024-12-02T01:13:01,658 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-02T01:13:01,694 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-02T01:13:01,694 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-02T01:13:01,696 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 417 msec 2024-12-02T01:13:01,701 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:01,701 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:13:01,701 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:01,707 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-02T01:13:01,718 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:13:01,729 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 21 msec 2024-12-02T01:13:01,741 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-02T01:13:01,759 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-02T01:13:01,771 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 30 msec 2024-12-02T01:13:01,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-02T01:13:01,818 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.508sec 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40313,1733101980116-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-02T01:13:01,818 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40313,1733101980116-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-02T01:13:01,821 DEBUG [master/a82f277fe18c:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-02T01:13:01,821 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-02T01:13:01,821 INFO [master/a82f277fe18c:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=a82f277fe18c,40313,1733101980116-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-02T01:13:01,891 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x09f59a7f to 127.0.0.1:63646 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@4d2167c8 2024-12-02T01:13:01,903 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@31e8c90e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-02T01:13:01,906 DEBUG [hconnection-0x6375b289-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-02T01:13:01,909 INFO [RS-EventLoopGroup-15-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:51964, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-02T01:13:01,912 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=a82f277fe18c,40313,1733101980116 2024-12-02T01:13:01,913 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-02T01:13:01,916 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-02T01:13:01,916 INFO [Time-limited test {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-02T01:13:01,920 INFO [Time-limited test {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=test.com%2C8080%2C1, suffix=, logDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1, archiveDir=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs, maxLogs=32 2024-12-02T01:13:01,921 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor test.com%2C8080%2C1.1733101981920 2024-12-02T01:13:01,928 INFO [Time-limited test {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1/test.com%2C8080%2C1.1733101981920 2024-12-02T01:13:01,928 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40737:40737),(127.0.0.1/127.0.0.1:40333:40333)] 2024-12-02T01:13:01,928 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor test.com%2C8080%2C1.1733101981928 2024-12-02T01:13:01,934 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1/test.com%2C8080%2C1.1733101981920 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1/test.com%2C8080%2C1.1733101981928 2024-12-02T01:13:01,934 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40737:40737),(127.0.0.1/127.0.0.1:40333:40333)] 2024-12-02T01:13:01,934 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1/test.com%2C8080%2C1.1733101981920 is not closed yet, will try archiving it next time 2024-12-02T01:13:01,935 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1 2024-12-02T01:13:01,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741837_1013 (size=93) 2024-12-02T01:13:01,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741837_1013 (size=93) 2024-12-02T01:13:01,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741838_1014 (size=93) 2024-12-02T01:13:01,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741838_1014 (size=93) 2024-12-02T01:13:01,937 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/test.com,8080,1/test.com%2C8080%2C1.1733101981920 to hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs/test.com%2C8080%2C1.1733101981920 2024-12-02T01:13:01,939 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs 2024-12-02T01:13:01,939 INFO [Time-limited test {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog test.com%2C8080%2C1:(num 1733101981928) 2024-12-02T01:13:01,939 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-02T01:13:01,939 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x09f59a7f to 127.0.0.1:63646 2024-12-02T01:13:01,939 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:13:01,939 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-02T01:13:01,939 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1398723059, stopped=false 2024-12-02T01:13:01,939 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=a82f277fe18c,40313,1733101980116 2024-12-02T01:13:01,951 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:13:01,951 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-02T01:13:01,951 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:01,951 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:01,951 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-02T01:13:01,952 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:13:01,952 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-02T01:13:01,952 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:13:01,952 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,39381,1733101980266' ***** 2024-12-02T01:13:01,952 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-02T01:13:01,952 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(3579): Received CLOSE for 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,39381,1733101980266 2024-12-02T01:13:01,952 DEBUG [RS:0;a82f277fe18c:39381 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-02T01:13:01,952 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 66584dc4c952477e12dd85d81277ebd9, disabling compactions & flushes 2024-12-02T01:13:01,952 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-02T01:13:01,953 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. after waiting 0 ms 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,953 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 66584dc4c952477e12dd85d81277ebd9 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-02T01:13:01,953 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1599): Waiting on 2 regions to close 2024-12-02T01:13:01,953 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1603): Online Regions={66584dc4c952477e12dd85d81277ebd9=hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9., 1588230740=hbase:meta,,1.1588230740} 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-02T01:13:01,953 DEBUG [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 66584dc4c952477e12dd85d81277ebd9 2024-12-02T01:13:01,953 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-02T01:13:01,953 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-02T01:13:01,953 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=1.23 KB heapSize=2.87 KB 2024-12-02T01:13:01,965 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/.tmp/info/7a56a7d9df754467af1533bf41420ba1 is 45, key is default/info:d/1733101981713/Put/seqid=0 2024-12-02T01:13:01,965 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/.tmp/info/504a763fd252406597a7c18d572a8e84 is 143, key is hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9./info:regioninfo/1733101981652/Put/seqid=0 2024-12-02T01:13:01,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741839_1015 (size=6595) 2024-12-02T01:13:01,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741840_1016 (size=5037) 2024-12-02T01:13:01,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741839_1015 (size=6595) 2024-12-02T01:13:01,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741840_1016 (size=5037) 2024-12-02T01:13:01,970 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.14 KB at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/.tmp/info/504a763fd252406597a7c18d572a8e84 2024-12-02T01:13:01,971 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/.tmp/info/7a56a7d9df754467af1533bf41420ba1 2024-12-02T01:13:01,976 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/.tmp/info/7a56a7d9df754467af1533bf41420ba1 as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/info/7a56a7d9df754467af1533bf41420ba1 2024-12-02T01:13:01,980 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/info/7a56a7d9df754467af1533bf41420ba1, entries=2, sequenceid=6, filesize=4.9 K 2024-12-02T01:13:01,981 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 66584dc4c952477e12dd85d81277ebd9 in 28ms, sequenceid=6, compaction requested=false 2024-12-02T01:13:01,981 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-02T01:13:01,984 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/namespace/66584dc4c952477e12dd85d81277ebd9/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-02T01:13:01,985 INFO [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,985 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 66584dc4c952477e12dd85d81277ebd9: 2024-12-02T01:13:01,985 DEBUG [RS_CLOSE_REGION-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733101981277.66584dc4c952477e12dd85d81277ebd9. 2024-12-02T01:13:01,988 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/.tmp/table/e7c9c2de100c495bba350bc797ba75b7 is 51, key is hbase:namespace/table:state/1733101981657/Put/seqid=0 2024-12-02T01:13:01,992 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741841_1017 (size=5242) 2024-12-02T01:13:01,992 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741841_1017 (size=5242) 2024-12-02T01:13:01,993 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=94 B at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/.tmp/table/e7c9c2de100c495bba350bc797ba75b7 2024-12-02T01:13:01,997 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/.tmp/info/504a763fd252406597a7c18d572a8e84 as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/info/504a763fd252406597a7c18d572a8e84 2024-12-02T01:13:02,001 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/info/504a763fd252406597a7c18d572a8e84, entries=10, sequenceid=9, filesize=6.4 K 2024-12-02T01:13:02,002 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/.tmp/table/e7c9c2de100c495bba350bc797ba75b7 as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/table/e7c9c2de100c495bba350bc797ba75b7 2024-12-02T01:13:02,007 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/table/e7c9c2de100c495bba350bc797ba75b7, entries=2, sequenceid=9, filesize=5.1 K 2024-12-02T01:13:02,007 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~1.23 KB/1264, heapSize ~2.59 KB/2648, currentSize=0 B/0 for 1588230740 in 54ms, sequenceid=9, compaction requested=false 2024-12-02T01:13:02,007 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-02T01:13:02,011 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/data/hbase/meta/1588230740/recovered.edits/12.seqid, newMaxSeqId=12, maxSeqId=1 2024-12-02T01:13:02,012 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-02T01:13:02,012 INFO [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-02T01:13:02,012 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-02T01:13:02,012 DEBUG [RS_CLOSE_META-regionserver/a82f277fe18c:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-02T01:13:02,153 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,39381,1733101980266; all regions closed. 2024-12-02T01:13:02,154 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266 2024-12-02T01:13:02,156 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741834_1010 (size=2484) 2024-12-02T01:13:02,157 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741834_1010 (size=2484) 2024-12-02T01:13:02,161 DEBUG [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs 2024-12-02T01:13:02,161 INFO [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C39381%2C1733101980266.meta:.meta(num 1733101981159) 2024-12-02T01:13:02,162 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/WALs/a82f277fe18c,39381,1733101980266 2024-12-02T01:13:02,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741833_1009 (size=1414) 2024-12-02T01:13:02,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741833_1009 (size=1414) 2024-12-02T01:13:02,166 DEBUG [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/oldWALs 2024-12-02T01:13:02,166 INFO [RS:0;a82f277fe18c:39381 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog a82f277fe18c%2C39381%2C1733101980266:(num 1733101980781) 2024-12-02T01:13:02,166 DEBUG [RS:0;a82f277fe18c:39381 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:13:02,167 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.LeaseManager(133): Closed leases 2024-12-02T01:13:02,167 INFO [RS:0;a82f277fe18c:39381 {}] hbase.ChoreService(370): Chore service for: regionserver/a82f277fe18c:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-02T01:13:02,167 INFO [regionserver/a82f277fe18c:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:13:02,167 INFO [RS:0;a82f277fe18c:39381 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:39381 2024-12-02T01:13:02,176 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/a82f277fe18c,39381,1733101980266 2024-12-02T01:13:02,176 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-02T01:13:02,184 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [a82f277fe18c,39381,1733101980266] 2024-12-02T01:13:02,184 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing a82f277fe18c,39381,1733101980266; numProcessing=1 2024-12-02T01:13:02,193 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/a82f277fe18c,39381,1733101980266 already deleted, retry=false 2024-12-02T01:13:02,193 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; a82f277fe18c,39381,1733101980266 expired; onlineServers=0 2024-12-02T01:13:02,193 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'a82f277fe18c,40313,1733101980116' ***** 2024-12-02T01:13:02,193 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-02T01:13:02,193 DEBUG [M:0;a82f277fe18c:40313 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@8eb1218, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=a82f277fe18c/172.17.0.2:0 2024-12-02T01:13:02,193 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegionServer(1224): stopping server a82f277fe18c,40313,1733101980116 2024-12-02T01:13:02,193 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegionServer(1250): stopping server a82f277fe18c,40313,1733101980116; all regions closed. 2024-12-02T01:13:02,193 DEBUG [M:0;a82f277fe18c:40313 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-02T01:13:02,193 DEBUG [M:0;a82f277fe18c:40313 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-02T01:13:02,193 DEBUG [M:0;a82f277fe18c:40313 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-02T01:13:02,193 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-02T01:13:02,193 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101980538 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.small.0-1733101980538,5,FailOnTimeoutGroup] 2024-12-02T01:13:02,193 DEBUG [master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101980538 {}] cleaner.HFileCleaner(306): Exit Thread[master/a82f277fe18c:0:becomeActiveMaster-HFileCleaner.large.0-1733101980538,5,FailOnTimeoutGroup] 2024-12-02T01:13:02,194 INFO [M:0;a82f277fe18c:40313 {}] hbase.ChoreService(370): Chore service for: master/a82f277fe18c:0 had [] on shutdown 2024-12-02T01:13:02,194 DEBUG [M:0;a82f277fe18c:40313 {}] master.HMaster(1733): Stopping service threads 2024-12-02T01:13:02,194 INFO [M:0;a82f277fe18c:40313 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-02T01:13:02,194 INFO [M:0;a82f277fe18c:40313 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-02T01:13:02,194 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-02T01:13:02,201 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-02T01:13:02,201 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-02T01:13:02,201 DEBUG [M:0;a82f277fe18c:40313 {}] zookeeper.ZKUtil(347): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-02T01:13:02,201 WARN [M:0;a82f277fe18c:40313 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-02T01:13:02,201 INFO [M:0;a82f277fe18c:40313 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-02T01:13:02,202 INFO [M:0;a82f277fe18c:40313 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-02T01:13:02,202 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-02T01:13:02,202 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-02T01:13:02,202 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:02,202 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:02,202 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-02T01:13:02,202 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:02,202 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=25.32 KB heapSize=32.31 KB 2024-12-02T01:13:02,222 DEBUG [M:0;a82f277fe18c:40313 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a7271fedb4204ed388fc2990a25b011e is 82, key is hbase:meta,,1/info:regioninfo/1733101981180/Put/seqid=0 2024-12-02T01:13:02,226 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741842_1018 (size=5672) 2024-12-02T01:13:02,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741842_1018 (size=5672) 2024-12-02T01:13:02,227 INFO [M:0;a82f277fe18c:40313 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a7271fedb4204ed388fc2990a25b011e 2024-12-02T01:13:02,242 DEBUG [M:0;a82f277fe18c:40313 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6db7823b1537416395141f5c836fa5e1 is 696, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733101981695/Put/seqid=0 2024-12-02T01:13:02,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741843_1019 (size=6626) 2024-12-02T01:13:02,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741843_1019 (size=6626) 2024-12-02T01:13:02,247 INFO [M:0;a82f277fe18c:40313 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.72 KB at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6db7823b1537416395141f5c836fa5e1 2024-12-02T01:13:02,263 DEBUG [M:0;a82f277fe18c:40313 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c1f6e8abb83e4e6cb3d875062a19ff82 is 69, key is a82f277fe18c,39381,1733101980266/rs:state/1733101980622/Put/seqid=0 2024-12-02T01:13:02,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741844_1020 (size=5156) 2024-12-02T01:13:02,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741844_1020 (size=5156) 2024-12-02T01:13:02,267 INFO [M:0;a82f277fe18c:40313 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c1f6e8abb83e4e6cb3d875062a19ff82 2024-12-02T01:13:02,284 DEBUG [M:0;a82f277fe18c:40313 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/8580170c89d04b858b150b13efc3e9b2 is 52, key is load_balancer_on/state:d/1733101981915/Put/seqid=0 2024-12-02T01:13:02,285 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:13:02,285 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39381-0x10194e32c360001, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:13:02,285 INFO [RS:0;a82f277fe18c:39381 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,39381,1733101980266; zookeeper connection closed. 2024-12-02T01:13:02,285 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@2cdd8bf6 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@2cdd8bf6 2024-12-02T01:13:02,285 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-02T01:13:02,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741845_1021 (size=5056) 2024-12-02T01:13:02,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741845_1021 (size=5056) 2024-12-02T01:13:02,289 INFO [M:0;a82f277fe18c:40313 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/8580170c89d04b858b150b13efc3e9b2 2024-12-02T01:13:02,293 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/a7271fedb4204ed388fc2990a25b011e as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a7271fedb4204ed388fc2990a25b011e 2024-12-02T01:13:02,298 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/a7271fedb4204ed388fc2990a25b011e, entries=8, sequenceid=70, filesize=5.5 K 2024-12-02T01:13:02,299 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/6db7823b1537416395141f5c836fa5e1 as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/6db7823b1537416395141f5c836fa5e1 2024-12-02T01:13:02,304 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/6db7823b1537416395141f5c836fa5e1, entries=8, sequenceid=70, filesize=6.5 K 2024-12-02T01:13:02,305 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/c1f6e8abb83e4e6cb3d875062a19ff82 as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/c1f6e8abb83e4e6cb3d875062a19ff82 2024-12-02T01:13:02,309 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/c1f6e8abb83e4e6cb3d875062a19ff82, entries=1, sequenceid=70, filesize=5.0 K 2024-12-02T01:13:02,310 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/8580170c89d04b858b150b13efc3e9b2 as hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/8580170c89d04b858b150b13efc3e9b2 2024-12-02T01:13:02,314 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34277/user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/8580170c89d04b858b150b13efc3e9b2, entries=1, sequenceid=70, filesize=4.9 K 2024-12-02T01:13:02,315 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(3040): Finished flush of dataSize ~25.32 KB/25929, heapSize ~32.25 KB/33024, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 113ms, sequenceid=70, compaction requested=false 2024-12-02T01:13:02,316 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-02T01:13:02,316 DEBUG [M:0;a82f277fe18c:40313 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-02T01:13:02,317 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9a268951-df6a-be52-d58c-380e418959d6/MasterData/WALs/a82f277fe18c,40313,1733101980116 2024-12-02T01:13:02,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:34733 is added to blk_1073741830_1006 (size=31030) 2024-12-02T01:13:02,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35771 is added to blk_1073741830_1006 (size=31030) 2024-12-02T01:13:02,319 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-02T01:13:02,319 INFO [M:0;a82f277fe18c:40313 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-02T01:13:02,319 INFO [M:0;a82f277fe18c:40313 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:40313 2024-12-02T01:13:02,326 DEBUG [M:0;a82f277fe18c:40313 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/a82f277fe18c,40313,1733101980116 already deleted, retry=false 2024-12-02T01:13:02,435 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:13:02,435 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:40313-0x10194e32c360000, quorum=127.0.0.1:63646, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-02T01:13:02,435 INFO [M:0;a82f277fe18c:40313 {}] regionserver.HRegionServer(1307): Exiting; stopping=a82f277fe18c,40313,1733101980116; zookeeper connection closed. 2024-12-02T01:13:02,475 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@1706478b{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:13:02,476 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@2c2c59ef{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:13:02,477 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:13:02,477 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@707d70b1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:13:02,477 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7284e1dc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir/,STOPPED} 2024-12-02T01:13:02,479 WARN [BP-2124532887-172.17.0.2-1733101978566 heartbeating to localhost/127.0.0.1:34277 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:13:02,479 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:13:02,479 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:13:02,479 WARN [BP-2124532887-172.17.0.2-1733101978566 heartbeating to localhost/127.0.0.1:34277 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2124532887-172.17.0.2-1733101978566 (Datanode Uuid b34e62b6-8253-4a3d-863a-c22c8b20c76d) service to localhost/127.0.0.1:34277 2024-12-02T01:13:02,480 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data3/current/BP-2124532887-172.17.0.2-1733101978566 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:13:02,480 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data4/current/BP-2124532887-172.17.0.2-1733101978566 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:13:02,480 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:13:02,483 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6c6a7c75{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-02T01:13:02,483 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@29776a3e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:13:02,483 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:13:02,483 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5ebf0c9b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:13:02,483 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24616503{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir/,STOPPED} 2024-12-02T01:13:02,484 WARN [BP-2124532887-172.17.0.2-1733101978566 heartbeating to localhost/127.0.0.1:34277 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-02T01:13:02,484 WARN [BP-2124532887-172.17.0.2-1733101978566 heartbeating to localhost/127.0.0.1:34277 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-2124532887-172.17.0.2-1733101978566 (Datanode Uuid 1ec20cb5-fddc-49a2-b15e-7d6679039a2f) service to localhost/127.0.0.1:34277 2024-12-02T01:13:02,484 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-02T01:13:02,484 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-02T01:13:02,484 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data1/current/BP-2124532887-172.17.0.2-1733101978566 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:13:02,485 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/cluster_ce6aedbc-55da-5413-373e-2218342b1577/dfs/data/data2/current/BP-2124532887-172.17.0.2-1733101978566 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-02T01:13:02,485 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-02T01:13:02,489 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6787fcf{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-02T01:13:02,489 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7c46377a{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-02T01:13:02,489 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-02T01:13:02,489 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@692b0555{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-02T01:13:02,489 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7fcf5e50{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/3418232a-f584-ddf5-f75e-6a3179a1b6db/hadoop.log.dir/,STOPPED} 2024-12-02T01:13:02,494 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-02T01:13:02,500 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:45673/user/jenkins/test-data/15dde93c-de62-1da8-45e4-dad8b1a01aa4/WALs/a82f277fe18c,43621,1733101712797/a82f277fe18c%2C43621%2C1733101712797.meta.1733101713779.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor118.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-02T01:13:02,508 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-02T01:13:02,514 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnNothingWritten Thread=146 (was 127) - Thread LEAK? -, OpenFileDescriptor=519 (was 490) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=161 (was 158) - SystemLoadAverage LEAK? -, ProcessCount=11 (was 11), AvailableMemoryMB=4774 (was 4780)